WorldWideScience

Sample records for general ensemble biogeochemical

  1. Protein folding simulations by generalized-ensemble algorithms.

    Science.gov (United States)

    Yoda, Takao; Sugita, Yuji; Okamoto, Yuko

    2014-01-01

    In the protein folding problem, conventional simulations in physical statistical mechanical ensembles, such as the canonical ensemble with fixed temperature, face a great difficulty. This is because there exist a huge number of local-minimum-energy states in the system and the conventional simulations tend to get trapped in these states, giving wrong results. Generalized-ensemble algorithms are based on artificial unphysical ensembles and overcome the above difficulty by performing random walks in potential energy, volume, and other physical quantities or their corresponding conjugate parameters such as temperature, pressure, etc. The advantage of generalized-ensemble simulations lies in the fact that they not only avoid getting trapped in states of energy local minima but also allows the calculations of physical quantities as functions of temperature or other parameters from a single simulation run. In this article we review the generalized-ensemble algorithms. Four examples, multicanonical algorithm, replica-exchange method, replica-exchange multicanonical algorithm, and multicanonical replica-exchange method, are described in detail. Examples of their applications to the protein folding problem are presented.

  2. Decimated Input Ensembles for Improved Generalization

    Science.gov (United States)

    Tumer, Kagan; Oza, Nikunj C.; Norvig, Peter (Technical Monitor)

    1999-01-01

    Recently, many researchers have demonstrated that using classifier ensembles (e.g., averaging the outputs of multiple classifiers before reaching a classification decision) leads to improved performance for many difficult generalization problems. However, in many domains there are serious impediments to such "turnkey" classification accuracy improvements. Most notable among these is the deleterious effect of highly correlated classifiers on the ensemble performance. One particular solution to this problem is generating "new" training sets by sampling the original one. However, with finite number of patterns, this causes a reduction in the training patterns each classifier sees, often resulting in considerably worsened generalization performance (particularly for high dimensional data domains) for each individual classifier. Generally, this drop in the accuracy of the individual classifier performance more than offsets any potential gains due to combining, unless diversity among classifiers is actively promoted. In this work, we introduce a method that: (1) reduces the correlation among the classifiers; (2) reduces the dimensionality of the data, thus lessening the impact of the 'curse of dimensionality'; and (3) improves the classification performance of the ensemble.

  3. Time-dependent generalized Gibbs ensembles in open quantum systems

    Science.gov (United States)

    Lange, Florian; Lenarčič, Zala; Rosch, Achim

    2018-04-01

    Generalized Gibbs ensembles have been used as powerful tools to describe the steady state of integrable many-particle quantum systems after a sudden change of the Hamiltonian. Here, we demonstrate numerically that they can be used for a much broader class of problems. We consider integrable systems in the presence of weak perturbations which break both integrability and drive the system to a state far from equilibrium. Under these conditions, we show that the steady state and the time evolution on long timescales can be accurately described by a (truncated) generalized Gibbs ensemble with time-dependent Lagrange parameters, determined from simple rate equations. We compare the numerically exact time evolutions of density matrices for small systems with a theory based on block-diagonal density matrices (diagonal ensemble) and a time-dependent generalized Gibbs ensemble containing only a small number of approximately conserved quantities, using the one-dimensional Heisenberg model with perturbations described by Lindblad operators as an example.

  4. Generalized ensemble theory with non-extensive statistics

    Science.gov (United States)

    Shen, Ke-Ming; Zhang, Ben-Wei; Wang, En-Ke

    2017-12-01

    The non-extensive canonical ensemble theory is reconsidered with the method of Lagrange multipliers by maximizing Tsallis entropy, with the constraint that the normalized term of Tsallis' q -average of physical quantities, the sum ∑ pjq, is independent of the probability pi for Tsallis parameter q. The self-referential problem in the deduced probability and thermal quantities in non-extensive statistics is thus avoided, and thermodynamical relationships are obtained in a consistent and natural way. We also extend the study to the non-extensive grand canonical ensemble theory and obtain the q-deformed Bose-Einstein distribution as well as the q-deformed Fermi-Dirac distribution. The theory is further applied to the generalized Planck law to demonstrate the distinct behaviors of the various generalized q-distribution functions discussed in literature.

  5. Dynamical predictive power of the generalized Gibbs ensemble revealed in a second quench.

    Science.gov (United States)

    Zhang, J M; Cui, F C; Hu, Jiangping

    2012-04-01

    We show that a quenched and relaxed completely integrable system is hardly distinguishable from the corresponding generalized Gibbs ensemble in a dynamical sense. To be specific, the response of the quenched and relaxed system to a second quench can be accurately reproduced by using the generalized Gibbs ensemble as a substitute. Remarkably, as demonstrated with the transverse Ising model and the hard-core bosons in one dimension, not only the steady values but even the transient, relaxation dynamics of the physical variables can be accurately reproduced by using the generalized Gibbs ensemble as a pseudoinitial state. This result is an important complement to the previously established result that a quenched and relaxed system is hardly distinguishable from the generalized Gibbs ensemble in a static sense. The relevance of the generalized Gibbs ensemble in the nonequilibrium dynamics of completely integrable systems is then greatly strengthened.

  6. Curve Boxplot: Generalization of Boxplot for Ensembles of Curves.

    Science.gov (United States)

    Mirzargar, Mahsa; Whitaker, Ross T; Kirby, Robert M

    2014-12-01

    In simulation science, computational scientists often study the behavior of their simulations by repeated solutions with variations in parameters and/or boundary values or initial conditions. Through such simulation ensembles, one can try to understand or quantify the variability or uncertainty in a solution as a function of the various inputs or model assumptions. In response to a growing interest in simulation ensembles, the visualization community has developed a suite of methods for allowing users to observe and understand the properties of these ensembles in an efficient and effective manner. An important aspect of visualizing simulations is the analysis of derived features, often represented as points, surfaces, or curves. In this paper, we present a novel, nonparametric method for summarizing ensembles of 2D and 3D curves. We propose an extension of a method from descriptive statistics, data depth, to curves. We also demonstrate a set of rendering and visualization strategies for showing rank statistics of an ensemble of curves, which is a generalization of traditional whisker plots or boxplots to multidimensional curves. Results are presented for applications in neuroimaging, hurricane forecasting and fluid dynamics.

  7. Microcanonical ensemble and algebra of conserved generators for generalized quantum dynamics

    International Nuclear Information System (INIS)

    Adler, S.L.; Horwitz, L.P.

    1996-01-01

    It has recently been shown, by application of statistical mechanical methods to determine the canonical ensemble governing the equilibrium distribution of operator initial values, that complex quantum field theory can emerge as a statistical approximation to an underlying generalized quantum dynamics. This result was obtained by an argument based on a Ward identity analogous to the equipartition theorem of classical statistical mechanics. We construct here a microcanonical ensemble which forms the basis of this canonical ensemble. This construction enables us to define the microcanonical entropy and free energy of the field configuration of the equilibrium distribution and to study the stability of the canonical ensemble. We also study the algebraic structure of the conserved generators from which the microcanonical and canonical ensembles are constructed, and the flows they induce on the phase space. copyright 1996 American Institute of Physics

  8. Grand Canonical Ensembles in General Relativity

    International Nuclear Information System (INIS)

    Klein, David; Yang, Wei-Shih

    2012-01-01

    We develop a formalism for general relativistic, grand canonical ensembles in space-times with timelike Killing fields. Using that, we derive ideal gas laws, and show how they depend on the geometry of the particular space-times. A systematic method for calculating Newtonian limits is given for a class of these space-times, which is illustrated for Kerr space-time. In addition, we prove uniqueness of the infinite volume Gibbs measure, and absence of phase transitions for a class of interaction potentials in anti-de Sitter space.

  9. Korean Percussion Ensemble ("Samulnori") in the General Music Classroom

    Science.gov (United States)

    Kang, Sangmi; Yoo, Hyesoo

    2016-01-01

    This article introduces "samulnori" (Korean percussion ensemble), its cultural background, and instructional methods as parts of a classroom approach to teaching upper-level general music. We introduce five of eight sections from "youngnam nong-ak" (a style of samulnori) as a repertoire for teaching Korean percussion music to…

  10. Loss of Ensemble Segregation in Dentate Gyrus, but Not in Somatosensory Cortex, during Contextual Fear Memory Generalization

    Directory of Open Access Journals (Sweden)

    Marie Yokoyama

    2016-11-01

    Full Text Available The details of contextual or episodic memories are lost and generalized with the passage of time. Proper generalization may underlie the formation and assimilation of semantic memories and enable animals to adapt to ever-changing environments, whereas overgeneralization of fear memory evokes maladaptive fear responses to harmless stimuli, which is a symptom of anxiety disorders such as post-traumatic stress disorder (PTSD. To understand the neural basis of fear memory generalization, we investigated the patterns of neuronal ensemble reactivation during memory retrieval when contextual fear memory expression is generalized using transgenic mice that allowed us to visualize specific neuronal ensembles activated during memory encoding and retrieval. We found preferential reactivations of neuronal ensembles in the primary somatosensory cortex, when mice were returned to the conditioned context to retrieve their memory 1 day after conditioning. In the hippocampal dentate gyrus (DG, exclusively separated ensemble reactivation was observed when mice were exposed to a novel context. These results suggest that the DG as well as the somatosensory cortex were likely to distinguish the two different contexts at the ensemble activity level when memory is not generalized at the behavioral level. However, 9 days after conditioning when animals exhibited generalized fear, the unique reactivation pattern in the DG, but not in the somatosensory cortex, was lost. Our results suggest that the alternations in the ensemble representation within the DG, or in upstream structures that link the sensory cortex to the hippocampus, may underlie generalized contextual fear memory expression.

  11. A brief history of the introduction of generalized ensembles to Markov chain Monte Carlo simulations

    Science.gov (United States)

    Berg, Bernd A.

    2017-03-01

    The most efficient weights for Markov chain Monte Carlo calculations of physical observables are not necessarily those of the canonical ensemble. Generalized ensembles, which do not exist in nature but can be simulated on computers, lead often to a much faster convergence. In particular, they have been used for simulations of first order phase transitions and for simulations of complex systems in which conflicting constraints lead to a rugged free energy landscape. Starting off with the Metropolis algorithm and Hastings' extension, I present a minireview which focuses on the explosive use of generalized ensembles in the early 1990s. Illustrations are given, which range from spin models to peptides.

  12. Generalized ensemble method applied to study systems with strong first order transitions

    Science.gov (United States)

    Małolepsza, E.; Kim, J.; Keyes, T.

    2015-09-01

    At strong first-order phase transitions, the entropy versus energy or, at constant pressure, enthalpy, exhibits convex behavior, and the statistical temperature curve correspondingly exhibits an S-loop or back-bending. In the canonical and isothermal-isobaric ensembles, with temperature as the control variable, the probability density functions become bimodal with peaks localized outside of the S-loop region. Inside, states are unstable, and as a result simulation of equilibrium phase coexistence becomes impossible. To overcome this problem, a method was proposed by Kim, Keyes and Straub [1], where optimally designed generalized ensemble sampling was combined with replica exchange, and denoted generalized replica exchange method (gREM). This new technique uses parametrized effective sampling weights that lead to a unimodal energy distribution, transforming unstable states into stable ones. In the present study, the gREM, originally developed as a Monte Carlo algorithm, was implemented to work with molecular dynamics in an isobaric ensemble and coded into LAMMPS, a highly optimized open source molecular simulation package. The method is illustrated in a study of the very strong solid/liquid transition in water.

  13. S-AMP: Approximate Message Passing for General Matrix Ensembles

    DEFF Research Database (Denmark)

    Cakmak, Burak; Winther, Ole; Fleury, Bernard H.

    2014-01-01

    the approximate message-passing (AMP) algorithm to general matrix ensembles with a well-defined large system size limit. The generalization is based on the S-transform (in free probability) of the spectrum of the measurement matrix. Furthermore, we show that the optimality of S-AMP follows directly from its......We propose a novel iterative estimation algorithm for linear observation models called S-AMP. The fixed points of S-AMP are the stationary points of the exact Gibbs free energy under a set of (first- and second-) moment consistency constraints in the large system limit. S-AMP extends...

  14. Multidimensional generalized-ensemble algorithms for complex systems.

    Science.gov (United States)

    Mitsutake, Ayori; Okamoto, Yuko

    2009-06-07

    We give general formulations of the multidimensional multicanonical algorithm, simulated tempering, and replica-exchange method. We generalize the original potential energy function E(0) by adding any physical quantity V of interest as a new energy term. These multidimensional generalized-ensemble algorithms then perform a random walk not only in E(0) space but also in V space. Among the three algorithms, the replica-exchange method is the easiest to perform because the weight factor is just a product of regular Boltzmann-like factors, while the weight factors for the multicanonical algorithm and simulated tempering are not a priori known. We give a simple procedure for obtaining the weight factors for these two latter algorithms, which uses a short replica-exchange simulation and the multiple-histogram reweighting techniques. As an example of applications of these algorithms, we have performed a two-dimensional replica-exchange simulation and a two-dimensional simulated-tempering simulation using an alpha-helical peptide system. From these simulations, we study the helix-coil transitions of the peptide in gas phase and in aqueous solution.

  15. Generalized rate-code model for neuron ensembles with finite populations

    International Nuclear Information System (INIS)

    Hasegawa, Hideo

    2007-01-01

    We have proposed a generalized Langevin-type rate-code model subjected to multiplicative noise, in order to study stationary and dynamical properties of an ensemble containing a finite number N of neurons. Calculations using the Fokker-Planck equation have shown that, owing to the multiplicative noise, our rate model yields various kinds of stationary non-Gaussian distributions such as Γ, inverse-Gaussian-like, and log-normal-like distributions, which have been experimentally observed. The dynamical properties of the rate model have been studied with the use of the augmented moment method (AMM), which was previously proposed by the author from a macroscopic point of view for finite-unit stochastic systems. In the AMM, the original N-dimensional stochastic differential equations (DEs) are transformed into three-dimensional deterministic DEs for the means and fluctuations of local and global variables. The dynamical responses of the neuron ensemble to pulse and sinusoidal inputs calculated by the AMM are in good agreement with those obtained by direct simulation. The synchronization in the neuronal ensemble is discussed. The variabilities of the firing rate and of the interspike interval are shown to increase with increasing magnitude of multiplicative noise, which may be a conceivable origin of the observed large variability in cortical neurons

  16. Exploring and Listening to Chinese Classical Ensembles in General Music

    Science.gov (United States)

    Zhang, Wenzhuo

    2017-01-01

    Music diversity is valued in theory, but the extent to which it is efficiently presented in music class remains limited. Within this article, I aim to bridge this gap by introducing four genres of Chinese classical ensembles--Qin and Xiao duets, Jiang Nan bamboo and silk ensembles, Cantonese ensembles, and contemporary Chinese orchestras--into the…

  17. Entropy of network ensembles

    Science.gov (United States)

    Bianconi, Ginestra

    2009-03-01

    In this paper we generalize the concept of random networks to describe network ensembles with nontrivial features by a statistical mechanics approach. This framework is able to describe undirected and directed network ensembles as well as weighted network ensembles. These networks might have nontrivial community structure or, in the case of networks embedded in a given space, they might have a link probability with a nontrivial dependence on the distance between the nodes. These ensembles are characterized by their entropy, which evaluates the cardinality of networks in the ensemble. In particular, in this paper we define and evaluate the structural entropy, i.e., the entropy of the ensembles of undirected uncorrelated simple networks with given degree sequence. We stress the apparent paradox that scale-free degree distributions are characterized by having small structural entropy while they are so widely encountered in natural, social, and technological complex systems. We propose a solution to the paradox by proving that scale-free degree distributions are the most likely degree distribution with the corresponding value of the structural entropy. Finally, the general framework we present in this paper is able to describe microcanonical ensembles of networks as well as canonical or hidden-variable network ensembles with significant implications for the formulation of network-constructing algorithms.

  18. Wang-Landau Reaction Ensemble Method: Simulation of Weak Polyelectrolytes and General Acid-Base Reactions.

    Science.gov (United States)

    Landsgesell, Jonas; Holm, Christian; Smiatek, Jens

    2017-02-14

    We present a novel method for the study of weak polyelectrolytes and general acid-base reactions in molecular dynamics and Monte Carlo simulations. The approach combines the advantages of the reaction ensemble and the Wang-Landau sampling method. Deprotonation and protonation reactions are simulated explicitly with the help of the reaction ensemble method, while the accurate sampling of the corresponding phase space is achieved by the Wang-Landau approach. The combination of both techniques provides a sufficient statistical accuracy such that meaningful estimates for the density of states and the partition sum can be obtained. With regard to these estimates, several thermodynamic observables like the heat capacity or reaction free energies can be calculated. We demonstrate that the computation times for the calculation of titration curves with a high statistical accuracy can be significantly decreased when compared to the original reaction ensemble method. The applicability of our approach is validated by the study of weak polyelectrolytes and their thermodynamic properties.

  19. Biogeochemical Controls on Technetium Mobility in Biogeochemical Controls on Technetium Mobility in FRC Sediments

    International Nuclear Information System (INIS)

    Lloyd, J.R.; McBeth, J.M.; Livens, F.R.; Bryan, N.D.; Ellis, B.; Sharma, H.; Burke, I.T.; Morris, K.

    2004-01-01

    Technetium-99 is a priority pollutant at numerous DOE sites, due to its long half-life (2.1 x 10 5 years), high mobility as Tc(VII) in oxic waters, and bioavailability as a sulfate analog. 99 Tc is far less mobile under anaerobic conditions, forming insoluble Tc(IV) precipitates. As anaerobic microorganisms can reduce soluble Tc(VII) to insoluble Tc(IV), microbial metabolism may have the potential to treat sediments and waters contaminated with Tc. Baseline studies of fundamental mechanisms of Tc(VII) bioreduction and precipitation (reviewed by Lloyd et al, 2002) have generally used pure cultures of metal-reducing bacteria, in order to develop conceptual models for the biogeochemical cycling of Tc. There is, however, comparatively little known about interactions of metal-reducing bacteria with environmentally relevant trace concentrations of Tc, against a more complex biogeochemical background provided by mixed microbial communities in the subsurface. The objective of this new NABIR project is to probe the site specific biogeochemical conditions that control the mobility of Tc at the FRC (Oak Ridge, TN). This information is required for the rational design of in situ bioremediation strategies for technetium-contaminated subsurface environments. We will use a combination of geochemical, mineralogical, microbiological and spectroscopic techniques to determine the solubility and phase associations of Tc in FRC sediments, and characterize the underpinning biogeochemical controls. A key strength of this project is that many of the techniques we are using have already been optimized by our research team, who are also studying the biogeochemical controls on Tc mobility in marine and freshwater sediments in the UK in a NERC funded companion study.

  20. A generalized polynomial chaos based ensemble Kalman filter with high accuracy

    International Nuclear Information System (INIS)

    Li Jia; Xiu Dongbin

    2009-01-01

    As one of the most adopted sequential data assimilation methods in many areas, especially those involving complex nonlinear dynamics, the ensemble Kalman filter (EnKF) has been under extensive investigation regarding its properties and efficiency. Compared to other variants of the Kalman filter (KF), EnKF is straightforward to implement, as it employs random ensembles to represent solution states. This, however, introduces sampling errors that affect the accuracy of EnKF in a negative manner. Though sampling errors can be easily reduced by using a large number of samples, in practice this is undesirable as each ensemble member is a solution of the system of state equations and can be time consuming to compute for large-scale problems. In this paper we present an efficient EnKF implementation via generalized polynomial chaos (gPC) expansion. The key ingredients of the proposed approach involve (1) solving the system of stochastic state equations via the gPC methodology to gain efficiency; and (2) sampling the gPC approximation of the stochastic solution with an arbitrarily large number of samples, at virtually no additional computational cost, to drastically reduce the sampling errors. The resulting algorithm thus achieves a high accuracy at reduced computational cost, compared to the classical implementations of EnKF. Numerical examples are provided to verify the convergence property and accuracy improvement of the new algorithm. We also prove that for linear systems with Gaussian noise, the first-order gPC Kalman filter method is equivalent to the exact Kalman filter.

  1. Simultaneous escaping of explicit and hidden free energy barriers: application of the orthogonal space random walk strategy in generalized ensemble based conformational sampling.

    Science.gov (United States)

    Zheng, Lianqing; Chen, Mengen; Yang, Wei

    2009-06-21

    To overcome the pseudoergodicity problem, conformational sampling can be accelerated via generalized ensemble methods, e.g., through the realization of random walks along prechosen collective variables, such as spatial order parameters, energy scaling parameters, or even system temperatures or pressures, etc. As usually observed, in generalized ensemble simulations, hidden barriers are likely to exist in the space perpendicular to the collective variable direction and these residual free energy barriers could greatly abolish the sampling efficiency. This sampling issue is particularly severe when the collective variable is defined in a low-dimension subset of the target system; then the "Hamiltonian lagging" problem, which reveals the fact that necessary structural relaxation falls behind the move of the collective variable, may be likely to occur. To overcome this problem in equilibrium conformational sampling, we adopted the orthogonal space random walk (OSRW) strategy, which was originally developed in the context of free energy simulation [L. Zheng, M. Chen, and W. Yang, Proc. Natl. Acad. Sci. U.S.A. 105, 20227 (2008)]. Thereby, generalized ensemble simulations can simultaneously escape both the explicit barriers along the collective variable direction and the hidden barriers that are strongly coupled with the collective variable move. As demonstrated in our model studies, the present OSRW based generalized ensemble treatments show improved sampling capability over the corresponding classical generalized ensemble treatments.

  2. Effect of land model ensemble versus coupled model ensemble on the simulation of precipitation climatology and variability

    Science.gov (United States)

    Wei, Jiangfeng; Dirmeyer, Paul A.; Yang, Zong-Liang; Chen, Haishan

    2017-10-01

    Through a series of model simulations with an atmospheric general circulation model coupled to three different land surface models, this study investigates the impacts of land model ensembles and coupled model ensemble on precipitation simulation. It is found that coupling an ensemble of land models to an atmospheric model has a very minor impact on the improvement of precipitation climatology and variability, but a simple ensemble average of the precipitation from three individually coupled land-atmosphere models produces better results, especially for precipitation variability. The generally weak impact of land processes on precipitation should be the main reason that the land model ensembles do not improve precipitation simulation. However, if there are big biases in the land surface model or land surface data set, correcting them could improve the simulated climate, especially for well-constrained regional climate simulations.

  3. Ensemble Methods

    Science.gov (United States)

    Re, Matteo; Valentini, Giorgio

    2012-03-01

    proposed to explain the characteristics and the successful application of ensembles to different application domains. For instance, Allwein, Schapire, and Singer interpreted the improved generalization capabilities of ensembles of learning machines in the framework of large margin classifiers [4,177], Kleinberg in the context of stochastic discrimination theory [112], and Breiman and Friedman in the light of the bias-variance analysis borrowed from classical statistics [21,70]. Empirical studies showed that both in classification and regression problems, ensembles improve on single learning machines, and moreover large experimental studies compared the effectiveness of different ensemble methods on benchmark data sets [10,11,49,188]. The interest in this research area is motivated also by the availability of very fast computers and networks of workstations at a relatively low cost that allow the implementation and the experimentation of complex ensemble methods using off-the-shelf computer platforms. However, as explained in Section 26.2 there are deeper reasons to use ensembles of learning machines, motivated by the intrinsic characteristics of the ensemble methods. The main aim of this chapter is to introduce ensemble methods and to provide an overview and a bibliography of the main areas of research, without pretending to be exhaustive or to explain the detailed characteristics of each ensemble method. The paper is organized as follows. In the next section, the main theoretical and practical reasons for combining multiple learners are introduced. Section 26.3 depicts the main taxonomies on ensemble methods proposed in the literature. In Section 26.4 and 26.5, we present an overview of the main supervised ensemble methods reported in the literature, adopting a simple taxonomy, originally proposed in Ref. [201]. Applications of ensemble methods are only marginally considered, but a specific section on some relevant applications of ensemble methods in astronomy and

  4. A class of energy-based ensembles in Tsallis statistics

    International Nuclear Information System (INIS)

    Chandrashekar, R; Naina Mohammed, S S

    2011-01-01

    A comprehensive investigation is carried out on the class of energy-based ensembles. The eight ensembles are divided into two main classes. In the isothermal class of ensembles the individual members are at the same temperature. A unified framework is evolved to describe the four isothermal ensembles using the currently accepted third constraint formalism. The isothermal–isobaric, grand canonical and generalized ensembles are illustrated through a study of the classical nonrelativistic and extreme relativistic ideal gas models. An exact calculation is possible only in the case of the isothermal–isobaric ensemble. The study of the ideal gas models in the grand canonical and the generalized ensembles has been carried out using a perturbative procedure with the nonextensivity parameter (1 − q) as the expansion parameter. Though all the thermodynamic quantities have been computed up to a particular order in (1 − q) the procedure can be extended up to any arbitrary order in the expansion parameter. In the adiabatic class of ensembles the individual members of the ensemble have the same value of the heat function and a unified formulation to described all four ensembles is given. The nonrelativistic and the extreme relativistic ideal gases are studied in the isoenthalpic–isobaric ensemble, the adiabatic ensemble with number fluctuations and the adiabatic ensemble with number and particle fluctuations

  5. Ensemble data assimilation in the Red Sea: sensitivity to ensemble selection and atmospheric forcing

    KAUST Repository

    Toye, Habib

    2017-05-26

    We present our efforts to build an ensemble data assimilation and forecasting system for the Red Sea. The system consists of the high-resolution Massachusetts Institute of Technology general circulation model (MITgcm) to simulate ocean circulation and of the Data Research Testbed (DART) for ensemble data assimilation. DART has been configured to integrate all members of an ensemble adjustment Kalman filter (EAKF) in parallel, based on which we adapted the ensemble operations in DART to use an invariant ensemble, i.e., an ensemble Optimal Interpolation (EnOI) algorithm. This approach requires only single forward model integration in the forecast step and therefore saves substantial computational cost. To deal with the strong seasonal variability of the Red Sea, the EnOI ensemble is then seasonally selected from a climatology of long-term model outputs. Observations of remote sensing sea surface height (SSH) and sea surface temperature (SST) are assimilated every 3 days. Real-time atmospheric fields from the National Center for Environmental Prediction (NCEP) and the European Center for Medium-Range Weather Forecasts (ECMWF) are used as forcing in different assimilation experiments. We investigate the behaviors of the EAKF and (seasonal-) EnOI and compare their performances for assimilating and forecasting the circulation of the Red Sea. We further assess the sensitivity of the assimilation system to various filtering parameters (ensemble size, inflation) and atmospheric forcing.

  6. Linking 1D coastal ocean modelling to environmental management: an ensemble approach

    Science.gov (United States)

    Mussap, Giulia; Zavatarelli, Marco; Pinardi, Nadia

    2017-12-01

    The use of a one-dimensional interdisciplinary numerical model of the coastal ocean as a tool contributing to the formulation of ecosystem-based management (EBM) is explored. The focus is on the definition of an experimental design based on ensemble simulations, integrating variability linked to scenarios (characterised by changes in the system forcing) and to the concurrent variation of selected, and poorly constrained, model parameters. The modelling system used was previously specifically designed for the use in "data-rich" areas, so that horizontal dynamics can be resolved by a diagnostic approach and external inputs can be parameterised by nudging schemes properly calibrated. Ensembles determined by changes in the simulated environmental (physical and biogeochemical) dynamics, under joint forcing and parameterisation variations, highlight the uncertainties associated to the application of specific scenarios that are relevant to EBM, providing an assessment of the reliability of the predicted changes. The work has been carried out by implementing the coupled modelling system BFM-POM1D in an area of Gulf of Trieste (northern Adriatic Sea), considered homogeneous from the point of view of hydrological properties, and forcing it by changing climatic (warming) and anthropogenic (reduction of the land-based nutrient input) pressure. Model parameters affected by considerable uncertainties (due to the lack of relevant observations) were varied jointly with the scenarios of change. The resulting large set of ensemble simulations provided a general estimation of the model uncertainties related to the joint variation of pressures and model parameters. The information of the model result variability aimed at conveying efficiently and comprehensibly the information on the uncertainties/reliability of the model results to non-technical EBM planners and stakeholders, in order to have the model-based information effectively contributing to EBM.

  7. The classicality and quantumness of a quantum ensemble

    International Nuclear Information System (INIS)

    Zhu Xuanmin; Pang Shengshi; Wu Shengjun; Liu Quanhui

    2011-01-01

    In this Letter, we investigate the classicality and quantumness of a quantum ensemble. We define a quantity called ensemble classicality based on classical cloning strategy (ECCC) to characterize how classical a quantum ensemble is. An ensemble of commuting states has a unit ECCC, while a general ensemble can have a ECCC less than 1. We also study how quantum an ensemble is by defining a related quantity called quantumness. We find that the classicality of an ensemble is closely related to how perfectly the ensemble can be cloned, and that the quantumness of the ensemble used in a quantum key distribution (QKD) protocol is exactly the attainable lower bound of the error rate in the sifted key. - Highlights: → A quantity is defined to characterize how classical a quantum ensemble is. → The classicality of an ensemble is closely related to the cloning performance. → Another quantity is also defined to investigate how quantum an ensemble is. → This quantity gives the lower bound of the error rate in a QKD protocol.

  8. Characterization of mixing errors in a coupled physical biogeochemical model of the North Atlantic: implications for nonlinear estimation using Gaussian anamorphosis

    Directory of Open Access Journals (Sweden)

    D. Béal

    2010-02-01

    Full Text Available In biogeochemical models coupled to ocean circulation models, vertical mixing is an important physical process which governs the nutrient supply and the plankton residence in the euphotic layer. However, vertical mixing is often poorly represented in numerical simulations because of approximate parameterizations of sub-grid scale turbulence, wind forcing errors and other mis-represented processes such as restratification by mesoscale eddies. Getting a sufficient knowledge of the nature and structure of these errors is necessary to implement appropriate data assimilation methods and to evaluate if they can be controlled by a given observation system.

    In this paper, Monte Carlo simulations are conducted to study mixing errors induced by approximate wind forcings in a three-dimensional coupled physical-biogeochemical model of the North Atlantic with a 1/4° horizontal resolution. An ensemble forecast involving 200 members is performed during the 1998 spring bloom, by prescribing perturbations of the wind forcing to generate mixing errors. The biogeochemical response is shown to be rather complex because of nonlinearities and threshold effects in the coupled model. The response of the surface phytoplankton depends on the region of interest and is particularly sensitive to the local stratification. In addition, the statistical relationships computed between the various physical and biogeochemical variables reflect the signature of the non-Gaussian behaviour of the system. It is shown that significant information on the ecosystem can be retrieved from observations of chlorophyll concentration or sea surface temperature if a simple nonlinear change of variables (anamorphosis is performed by mapping separately and locally the ensemble percentiles of the distributions of each state variable on the Gaussian percentiles. The results of idealized observational updates (performed with perfect observations and neglecting horizontal correlations

  9. Ensemble modelling of nitrogen fluxes: data fusion for a Swedish meso-scale catchment

    Directory of Open Access Journals (Sweden)

    J.-F. Exbrayat

    2010-12-01

    Full Text Available Model predictions of biogeochemical fluxes at the landscape scale are highly uncertain, both with respect to stochastic (parameter and structural uncertainty. In this study 5 different models (LASCAM, LASCAM-S, a self-developed tool, SWAT and HBV-N-D designed to simulate hydrological fluxes as well as mobilisation and transport of one or several nitrogen species were applied to the mesoscale River Fyris catchment in mid-eastern Sweden.

    Hydrological calibration against 5 years of recorded daily discharge at two stations gave highly variable results with Nash-Sutcliffe Efficiency (NSE ranging between 0.48 and 0.83. Using the calibrated hydrological parameter sets, the parameter uncertainty linked to the nitrogen parameters was explored in order to cover the range of possible predictions of exported loads for 3 nitrogen species: nitrate (NO3, ammonium (NH4 and total nitrogen (Tot-N. For each model and each nitrogen species, predictions were ranked in two different ways according to the performance indicated by two different goodness-of-fit measures: the coefficient of determination R2 and the root mean square error RMSE. A total of 2160 deterministic Single Model Ensembles (SME was generated using an increasing number of members (from the 2 best to the 10 best single predictions. Finally the best SME for each model, nitrogen species and discharge station were selected and merged into 330 different Multi-Model Ensembles (MME. The evolution of changes in R2 and RMSE was used as a performance descriptor of the ensemble procedure.

    In each studied case, numerous ensemble merging schemes were identified which outperformed any of their members. Improvement rates were generally higher when worse members were introduced. The highest improvements were achieved for the nitrogen SMEs compiled with multiple linear regression models with R2 selected members, which

  10. Impacts of calibration strategies and ensemble methods on ensemble flood forecasting over Lanjiang basin, Southeast China

    Science.gov (United States)

    Liu, Li; Xu, Yue-Ping

    2017-04-01

    Ensemble flood forecasting driven by numerical weather prediction products is becoming more commonly used in operational flood forecasting applications.In this study, a hydrological ensemble flood forecasting system based on Variable Infiltration Capacity (VIC) model and quantitative precipitation forecasts from TIGGE dataset is constructed for Lanjiang Basin, Southeast China. The impacts of calibration strategies and ensemble methods on the performance of the system are then evaluated.The hydrological model is optimized by parallel programmed ɛ-NSGAII multi-objective algorithm and two respectively parameterized models are determined to simulate daily flows and peak flows coupled with a modular approach.The results indicatethat the ɛ-NSGAII algorithm permits more efficient optimization and rational determination on parameter setting.It is demonstrated that the multimodel ensemble streamflow mean have better skills than the best singlemodel ensemble mean (ECMWF) and the multimodel ensembles weighted on members and skill scores outperform other multimodel ensembles. For typical flood event, it is proved that the flood can be predicted 3-4 days in advance, but the flows in rising limb can be captured with only 1-2 days ahead due to the flash feature. With respect to peak flows selected by Peaks Over Threshold approach, the ensemble means from either singlemodel or multimodels are generally underestimated as the extreme values are smoothed out by ensemble process.

  11. The Use of Artificial-Intelligence-Based Ensembles for Intrusion Detection: A Review

    Directory of Open Access Journals (Sweden)

    Gulshan Kumar

    2012-01-01

    Full Text Available In supervised learning-based classification, ensembles have been successfully employed to different application domains. In the literature, many researchers have proposed different ensembles by considering different combination methods, training datasets, base classifiers, and many other factors. Artificial-intelligence-(AI- based techniques play prominent role in development of ensemble for intrusion detection (ID and have many benefits over other techniques. However, there is no comprehensive review of ensembles in general and AI-based ensembles for ID to examine and understand their current research status to solve the ID problem. Here, an updated review of ensembles and their taxonomies has been presented in general. The paper also presents the updated review of various AI-based ensembles for ID (in particular during last decade. The related studies of AI-based ensembles are compared by set of evaluation metrics driven from (1 architecture & approach followed; (2 different methods utilized in different phases of ensemble learning; (3 other measures used to evaluate classification performance of the ensembles. The paper also provides the future directions of the research in this area. The paper will help the better understanding of different directions in which research of ensembles has been done in general and specifically: field of intrusion detection systems (IDSs.

  12. Kinetic theory of nonequilibrium ensembles, irreversible thermodynamics, and generalized hydrodynamics

    CERN Document Server

    Eu, Byung Chan

    2016-01-01

    This book presents the fundamentals of irreversible thermodynamics for nonlinear transport processes in gases and liquids, as well as for generalized hydrodynamics extending the classical hydrodynamics of Navier, Stokes, Fourier, and Fick. Together with its companion volume on relativistic theories, it provides a comprehensive picture of the kinetic theory formulated from the viewpoint of nonequilibrium ensembles in both nonrelativistic and, in Vol. 2, relativistic contexts. Theories of macroscopic irreversible processes must strictly conform to the thermodynamic laws at every step and in all approximations that enter their derivation from the mechanical principles. Upholding this as the inviolable tenet, the author develops theories of irreversible transport processes in fluids (gases or liquids) on the basis of irreversible kinetic equations satisfying the H theorem. They apply regardless of whether the processes are near to or far removed from equilibrium, or whether they are linear or nonlinear with respe...

  13. Neural Network Ensembles

    DEFF Research Database (Denmark)

    Hansen, Lars Kai; Salamon, Peter

    1990-01-01

    We propose several means for improving the performance an training of neural networks for classification. We use crossvalidation as a tool for optimizing network parameters and architecture. We show further that the remaining generalization error can be reduced by invoking ensembles of similar...... networks....

  14. The Ensembl REST API: Ensembl Data for Any Language.

    Science.gov (United States)

    Yates, Andrew; Beal, Kathryn; Keenan, Stephen; McLaren, William; Pignatelli, Miguel; Ritchie, Graham R S; Ruffier, Magali; Taylor, Kieron; Vullo, Alessandro; Flicek, Paul

    2015-01-01

    We present a Web service to access Ensembl data using Representational State Transfer (REST). The Ensembl REST server enables the easy retrieval of a wide range of Ensembl data by most programming languages, using standard formats such as JSON and FASTA while minimizing client work. We also introduce bindings to the popular Ensembl Variant Effect Predictor tool permitting large-scale programmatic variant analysis independent of any specific programming language. The Ensembl REST API can be accessed at http://rest.ensembl.org and source code is freely available under an Apache 2.0 license from http://github.com/Ensembl/ensembl-rest. © The Author 2014. Published by Oxford University Press.

  15. A hybrid ensemble-OI Kalman filter for efficient data assimilation into a 3-D biogeochemical model of the Mediterranean

    KAUST Repository

    Tsiaras, Kostas P.

    2017-04-20

    A hybrid ensemble data assimilation scheme (HYBRID), combining a flow-dependent with a static background covariance, was developed and implemented for assimilating satellite (SeaWiFS) Chl-a data into a marine ecosystem model of the Mediterranean. The performance of HYBRID was assessed against a model free-run, the ensemble-based singular evolutive interpolated Kalman (SEIK) and its variant with static covariance (SFEK), with regard to the assimilated variable (Chl-a) and non-assimilated variables (dissolved inorganic nutrients). HYBRID was found more efficient than both SEIK and SFEK, reducing the Chl-a error by more than 40% in most areas, as compared to the free-run. Data assimilation had a positive overall impact on nutrients, except for a deterioration of nitrates simulation by SEIK in the most productive area (Adriatic). This was related to SEIK pronounced update in this area and the phytoplankton limitation on phosphate that lead to a built up of excess nitrates. SEIK was found more efficient in productive and variable areas, where its ensemble exhibited important spread. SFEK had an effect mostly on Chl-a, performing better than SEIK in less dynamic areas, adequately described by the dominant modes of its static covariance. HYBRID performed well in all areas, due to its “blended” covariance. Its flow-dependent component appears to track changes in the system dynamics, while its static covariance helps maintaining sufficient spread in the forecast. HYBRID sensitivity experiments showed that an increased contribution from the flow-dependent covariance results in a deterioration of nitrates, similar to SEIK, while the improvement of HYBRID with increasing flow-dependent ensemble size quickly levels off.

  16. A hybrid ensemble-OI Kalman filter for efficient data assimilation into a 3-D biogeochemical model of the Mediterranean

    Science.gov (United States)

    Tsiaras, Kostas P.; Hoteit, Ibrahim; Kalaroni, Sofia; Petihakis, George; Triantafyllou, George

    2017-06-01

    A hybrid ensemble data assimilation scheme (HYBRID), combining a flow-dependent with a static background covariance, was developed and implemented for assimilating satellite (SeaWiFS) Chl-a data into a marine ecosystem model of the Mediterranean. The performance of HYBRID was assessed against a model free-run, the ensemble-based singular evolutive interpolated Kalman (SEIK) and its variant with static covariance (SFEK), with regard to the assimilated variable (Chl-a) and non-assimilated variables (dissolved inorganic nutrients). HYBRID was found more efficient than both SEIK and SFEK, reducing the Chl-a error by more than 40% in most areas, as compared to the free-run. Data assimilation had a positive overall impact on nutrients, except for a deterioration of nitrates simulation by SEIK in the most productive area (Adriatic). This was related to SEIK pronounced update in this area and the phytoplankton limitation on phosphate that lead to a built up of excess nitrates. SEIK was found more efficient in productive and variable areas, where its ensemble exhibited important spread. SFEK had an effect mostly on Chl-a, performing better than SEIK in less dynamic areas, adequately described by the dominant modes of its static covariance. HYBRID performed well in all areas, due to its "blended" covariance. Its flow-dependent component appears to track changes in the system dynamics, while its static covariance helps maintaining sufficient spread in the forecast. HYBRID sensitivity experiments showed that an increased contribution from the flow-dependent covariance results in a deterioration of nitrates, similar to SEIK, while the improvement of HYBRID with increasing flow-dependent ensemble size quickly levels off.

  17. A hybrid ensemble-OI Kalman filter for efficient data assimilation into a 3-D biogeochemical model of the Mediterranean

    KAUST Repository

    Tsiaras, Kostas P.; Hoteit, Ibrahim; Kalaroni, Sofia; Petihakis, George; Triantafyllou, George

    2017-01-01

    A hybrid ensemble data assimilation scheme (HYBRID), combining a flow-dependent with a static background covariance, was developed and implemented for assimilating satellite (SeaWiFS) Chl-a data into a marine ecosystem model of the Mediterranean. The performance of HYBRID was assessed against a model free-run, the ensemble-based singular evolutive interpolated Kalman (SEIK) and its variant with static covariance (SFEK), with regard to the assimilated variable (Chl-a) and non-assimilated variables (dissolved inorganic nutrients). HYBRID was found more efficient than both SEIK and SFEK, reducing the Chl-a error by more than 40% in most areas, as compared to the free-run. Data assimilation had a positive overall impact on nutrients, except for a deterioration of nitrates simulation by SEIK in the most productive area (Adriatic). This was related to SEIK pronounced update in this area and the phytoplankton limitation on phosphate that lead to a built up of excess nitrates. SEIK was found more efficient in productive and variable areas, where its ensemble exhibited important spread. SFEK had an effect mostly on Chl-a, performing better than SEIK in less dynamic areas, adequately described by the dominant modes of its static covariance. HYBRID performed well in all areas, due to its “blended” covariance. Its flow-dependent component appears to track changes in the system dynamics, while its static covariance helps maintaining sufficient spread in the forecast. HYBRID sensitivity experiments showed that an increased contribution from the flow-dependent covariance results in a deterioration of nitrates, similar to SEIK, while the improvement of HYBRID with increasing flow-dependent ensemble size quickly levels off.

  18. SSAGES: Software Suite for Advanced General Ensemble Simulations

    Energy Technology Data Exchange (ETDEWEB)

    Sidky, Hythem [Department of Chemical and Biomolecular Engineering, University of Notre Dame, Notre Dame, Indiana 46556, USA; Colón, Yamil J. [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Institute for Molecular Engineering and Materials Science Division, Argonne National Laboratory, Lemont, Illinois 60439, USA; Helfferich, Julian [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Steinbuch Center for Computing, Karlsruhe Institute of Technology, Hermann-von-Helmholtz-Platz 1, 76344 Eggenstein-Leopoldshafen, Germany; Sikora, Benjamin J. [Department of Chemical and Biomolecular Engineering, University of Notre Dame, Notre Dame, Indiana 46556, USA; Bezik, Cody [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Chu, Weiwei [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Giberti, Federico [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Guo, Ashley Z. [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Jiang, Xikai [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Lequieu, Joshua [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Li, Jiyuan [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Moller, Joshua [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Quevillon, Michael J. [Department of Chemical and Biomolecular Engineering, University of Notre Dame, Notre Dame, Indiana 46556, USA; Rahimi, Mohammad [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Ramezani-Dakhel, Hadi [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Department of Biochemistry and Molecular Biology, University of Chicago, Chicago, Illinois 60637, USA; Rathee, Vikramjit S. [Department of Chemical and Biomolecular Engineering, University of Notre Dame, Notre Dame, Indiana 46556, USA; Reid, Daniel R. [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Sevgen, Emre [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Thapar, Vikram [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Webb, Michael A. [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Institute for Molecular Engineering and Materials Science Division, Argonne National Laboratory, Lemont, Illinois 60439, USA; Whitmer, Jonathan K. [Department of Chemical and Biomolecular Engineering, University of Notre Dame, Notre Dame, Indiana 46556, USA; de Pablo, Juan J. [Institute for Molecular Engineering, University of Chicago, Chicago, Illinois 60637, USA; Institute for Molecular Engineering and Materials Science Division, Argonne National Laboratory, Lemont, Illinois 60439, USA

    2018-01-28

    Molecular simulation has emerged as an essential tool for modern-day research, but obtaining proper results and making reliable conclusions from simulations requires adequate sampling of the system under consideration. To this end, a variety of methods exist in the literature that can enhance sampling considerably, and increasingly sophisticated, effective algorithms continue to be developed at a rapid pace. Implementation of these techniques, however, can be challenging for experts and non-experts alike. There is a clear need for software that provides rapid, reliable, and easy access to a wide range of advanced sampling methods, and that facilitates implementation of new techniques as they emerge. Here we present SSAGES, a publicly available Software Suite for Advanced General Ensemble Simulations designed to interface with multiple widely used molecular dynamics simulations packages. SSAGES allows facile application of a variety of enhanced sampling techniques—including adaptive biasing force, string methods, and forward flux sampling—that extract meaningful free energy and transition path data from all-atom and coarse grained simulations. A noteworthy feature of SSAGES is a user-friendly framework that facilitates further development and implementation of new methods and collective variables. In this work, the use of SSAGES is illustrated in the context of simple representative applications involving distinct methods and different collective variables that are available in the current release of the suite.

  19. Short ensembles: an efficient method for discerning climate-relevant sensitivities in atmospheric general circulation models

    Directory of Open Access Journals (Sweden)

    H. Wan

    2014-09-01

    Full Text Available This paper explores the feasibility of an experimentation strategy for investigating sensitivities in fast components of atmospheric general circulation models. The basic idea is to replace the traditional serial-in-time long-term climate integrations by representative ensembles of shorter simulations. The key advantage of the proposed method lies in its efficiency: since fewer days of simulation are needed, the computational cost is less, and because individual realizations are independent and can be integrated simultaneously, the new dimension of parallelism can dramatically reduce the turnaround time in benchmark tests, sensitivities studies, and model tuning exercises. The strategy is not appropriate for exploring sensitivity of all model features, but it is very effective in many situations. Two examples are presented using the Community Atmosphere Model, version 5. In the first example, the method is used to characterize sensitivities of the simulated clouds to time-step length. Results show that 3-day ensembles of 20 to 50 members are sufficient to reproduce the main signals revealed by traditional 5-year simulations. A nudging technique is applied to an additional set of simulations to help understand the contribution of physics–dynamics interaction to the detected time-step sensitivity. In the second example, multiple empirical parameters related to cloud microphysics and aerosol life cycle are perturbed simultaneously in order to find out which parameters have the largest impact on the simulated global mean top-of-atmosphere radiation balance. It turns out that 12-member ensembles of 10-day simulations are able to reveal the same sensitivities as seen in 4-year simulations performed in a previous study. In both cases, the ensemble method reduces the total computational time by a factor of about 15, and the turnaround time by a factor of several hundred. The efficiency of the method makes it particularly useful for the development of

  20. Crossover ensembles of random matrices and skew-orthogonal polynomials

    International Nuclear Information System (INIS)

    Kumar, Santosh; Pandey, Akhilesh

    2011-01-01

    Highlights: → We study crossover ensembles of Jacobi family of random matrices. → We consider correlations for orthogonal-unitary and symplectic-unitary crossovers. → We use the method of skew-orthogonal polynomials and quaternion determinants. → We prove universality of spectral correlations in crossover ensembles. → We discuss applications to quantum conductance and communication theory problems. - Abstract: In a recent paper (S. Kumar, A. Pandey, Phys. Rev. E, 79, 2009, p. 026211) we considered Jacobi family (including Laguerre and Gaussian cases) of random matrix ensembles and reported exact solutions of crossover problems involving time-reversal symmetry breaking. In the present paper we give details of the work. We start with Dyson's Brownian motion description of random matrix ensembles and obtain universal hierarchic relations among the unfolded correlation functions. For arbitrary dimensions we derive the joint probability density (jpd) of eigenvalues for all transitions leading to unitary ensembles as equilibrium ensembles. We focus on the orthogonal-unitary and symplectic-unitary crossovers and give generic expressions for jpd of eigenvalues, two-point kernels and n-level correlation functions. This involves generalization of the theory of skew-orthogonal polynomials to crossover ensembles. We also consider crossovers in the circular ensembles to show the generality of our method. In the large dimensionality limit, correlations in spectra with arbitrary initial density are shown to be universal when expressed in terms of a rescaled symmetry breaking parameter. Applications of our crossover results to communication theory and quantum conductance problems are also briefly discussed.

  1. Ensemble data assimilation in the Red Sea: sensitivity to ensemble selection and atmospheric forcing

    KAUST Repository

    Toye, Habib; Zhan, Peng; Gopalakrishnan, Ganesh; Kartadikaria, Aditya R.; Huang, Huang; Knio, Omar; Hoteit, Ibrahim

    2017-01-01

    We present our efforts to build an ensemble data assimilation and forecasting system for the Red Sea. The system consists of the high-resolution Massachusetts Institute of Technology general circulation model (MITgcm) to simulate ocean circulation

  2. Robust Ensemble Filtering and Its Relation to Covariance Inflation in the Ensemble Kalman Filter

    KAUST Repository

    Luo, Xiaodong

    2011-12-01

    A robust ensemble filtering scheme based on the H∞ filtering theory is proposed. The optimal H∞ filter is derived by minimizing the supremum (or maximum) of a predefined cost function, a criterion different from the minimum variance used in the Kalman filter. By design, the H∞ filter is more robust than the Kalman filter, in the sense that the estimation error in the H∞ filter in general has a finite growth rate with respect to the uncertainties in assimilation, except for a special case that corresponds to the Kalman filter. The original form of the H∞ filter contains global constraints in time, which may be inconvenient for sequential data assimilation problems. Therefore a variant is introduced that solves some time-local constraints instead, and hence it is called the time-local H∞ filter (TLHF). By analogy to the ensemble Kalman filter (EnKF), the concept of ensemble time-local H∞ filter (EnTLHF) is also proposed. The general form of the EnTLHF is outlined, and some of its special cases are discussed. In particular, it is shown that an EnKF with certain covariance inflation is essentially an EnTLHF. In this sense, the EnTLHF provides a general framework for conducting covariance inflation in the EnKF-based methods. Some numerical examples are used to assess the relative robustness of the TLHF–EnTLHF in comparison with the corresponding KF–EnKF method.

  3. SSAGES: Software Suite for Advanced General Ensemble Simulations

    Science.gov (United States)

    Sidky, Hythem; Colón, Yamil J.; Helfferich, Julian; Sikora, Benjamin J.; Bezik, Cody; Chu, Weiwei; Giberti, Federico; Guo, Ashley Z.; Jiang, Xikai; Lequieu, Joshua; Li, Jiyuan; Moller, Joshua; Quevillon, Michael J.; Rahimi, Mohammad; Ramezani-Dakhel, Hadi; Rathee, Vikramjit S.; Reid, Daniel R.; Sevgen, Emre; Thapar, Vikram; Webb, Michael A.; Whitmer, Jonathan K.; de Pablo, Juan J.

    2018-01-01

    Molecular simulation has emerged as an essential tool for modern-day research, but obtaining proper results and making reliable conclusions from simulations requires adequate sampling of the system under consideration. To this end, a variety of methods exist in the literature that can enhance sampling considerably, and increasingly sophisticated, effective algorithms continue to be developed at a rapid pace. Implementation of these techniques, however, can be challenging for experts and non-experts alike. There is a clear need for software that provides rapid, reliable, and easy access to a wide range of advanced sampling methods and that facilitates implementation of new techniques as they emerge. Here we present SSAGES, a publicly available Software Suite for Advanced General Ensemble Simulations designed to interface with multiple widely used molecular dynamics simulations packages. SSAGES allows facile application of a variety of enhanced sampling techniques—including adaptive biasing force, string methods, and forward flux sampling—that extract meaningful free energy and transition path data from all-atom and coarse-grained simulations. A noteworthy feature of SSAGES is a user-friendly framework that facilitates further development and implementation of new methods and collective variables. In this work, the use of SSAGES is illustrated in the context of simple representative applications involving distinct methods and different collective variables that are available in the current release of the suite. The code may be found at: https://github.com/MICCoM/SSAGES-public.

  4. Short ensembles: An Efficient Method for Discerning Climate-relevant Sensitivities in Atmospheric General Circulation Models

    Energy Technology Data Exchange (ETDEWEB)

    Wan, Hui; Rasch, Philip J.; Zhang, Kai; Qian, Yun; Yan, Huiping; Zhao, Chun

    2014-09-08

    This paper explores the feasibility of an experimentation strategy for investigating sensitivities in fast components of atmospheric general circulation models. The basic idea is to replace the traditional serial-in-time long-term climate integrations by representative ensembles of shorter simulations. The key advantage of the proposed method lies in its efficiency: since fewer days of simulation are needed, the computational cost is less, and because individual realizations are independent and can be integrated simultaneously, the new dimension of parallelism can dramatically reduce the turnaround time in benchmark tests, sensitivities studies, and model tuning exercises. The strategy is not appropriate for exploring sensitivity of all model features, but it is very effective in many situations. Two examples are presented using the Community Atmosphere Model version 5. The first example demonstrates that the method is capable of characterizing the model cloud and precipitation sensitivity to time step length. A nudging technique is also applied to an additional set of simulations to help understand the contribution of physics-dynamics interaction to the detected time step sensitivity. In the second example, multiple empirical parameters related to cloud microphysics and aerosol lifecycle are perturbed simultaneously in order to explore which parameters have the largest impact on the simulated global mean top-of-atmosphere radiation balance. Results show that in both examples, short ensembles are able to correctly reproduce the main signals of model sensitivities revealed by traditional long-term climate simulations for fast processes in the climate system. The efficiency of the ensemble method makes it particularly useful for the development of high-resolution, costly and complex climate models.

  5. Ensemble inequivalence: Landau theory and the ABC model

    International Nuclear Information System (INIS)

    Cohen, O; Mukamel, D

    2012-01-01

    It is well known that systems with long-range interactions may exhibit different phase diagrams when studied within two different ensembles. In many of the previously studied examples of ensemble inequivalence, the phase diagrams differ only when the transition in one of the ensembles is first order. By contrast, in a recent study of a generalized ABC model, the canonical and grand-canonical ensembles of the model were shown to differ even when they both exhibit a continuous transition. Here we show that the order of the transition where ensemble inequivalence may occur is related to the symmetry properties of the order parameter associated with the transition. This is done by analyzing the Landau expansion of a generic model with long-range interactions. The conclusions drawn from the generic analysis are demonstrated for the ABC model by explicit calculation of its Landau expansion. (paper)

  6. Non-Hermitian Extensions of Wishart Random Matrix Ensembles

    International Nuclear Information System (INIS)

    Akemann, G.

    2011-01-01

    We briefly review the solution of three ensembles of non-Hermitian random matrices generalizing the Wishart-Laguerre (also called chiral) ensembles. These generalizations are realized as Gaussian two-matrix models, where the complex eigenvalues of the product of the two independent rectangular matrices are sought, with the matrix elements of both matrices being either real, complex or quaternion real. We also present the more general case depending on a non-Hermiticity parameter, that allows us to interpolate between the corresponding three Hermitian Wishart ensembles with real eigenvalues and the maximally non-Hermitian case. All three symmetry classes are explicitly solved for finite matrix size N x M for all complex eigenvalue correlations functions (and real or mixed correlations for real matrix elements). These are given in terms of the corresponding kernels built from orthogonal or skew-orthogonal Laguerre polynomials in the complex plane. We then present the corresponding three Bessel kernels in the complex plane in the microscopic large-N scaling limit at the origin, both at weak and strong non-Hermiticity with M - N ≥ 0 fixed. (author)

  7. Analysis of ensemble learning using simple perceptrons based on online learning theory

    Science.gov (United States)

    Miyoshi, Seiji; Hara, Kazuyuki; Okada, Masato

    2005-03-01

    Ensemble learning of K nonlinear perceptrons, which determine their outputs by sign functions, is discussed within the framework of online learning and statistical mechanics. One purpose of statistical learning theory is to theoretically obtain the generalization error. This paper shows that ensemble generalization error can be calculated by using two order parameters, that is, the similarity between a teacher and a student, and the similarity among students. The differential equations that describe the dynamical behaviors of these order parameters are derived in the case of general learning rules. The concrete forms of these differential equations are derived analytically in the cases of three well-known rules: Hebbian learning, perceptron learning, and AdaTron (adaptive perceptron) learning. Ensemble generalization errors of these three rules are calculated by using the results determined by solving their differential equations. As a result, these three rules show different characteristics in their affinity for ensemble learning, that is “maintaining variety among students.” Results show that AdaTron learning is superior to the other two rules with respect to that affinity.

  8. On-line Learning of Unlearnable True Teacher through Mobile Ensemble Teachers

    Science.gov (United States)

    Hirama, Takeshi; Hukushima, Koji

    2008-09-01

    The on-line learning of a hierarchical learning model is studied by a method based on statistical mechanics. In our model, a student of a simple perceptron learns from not a true teacher directly, but ensemble teachers who learn from a true teacher with a perceptron learning rule. Since the true teacher and ensemble teachers are expressed as nonmonotonic and simple perceptrons, respectively, the ensemble teachers go around the unlearnable true teacher with the distance between them fixed in an asymptotic steady state. The generalization performance of the student is shown to exceed that of the ensemble teachers in a transient state, as was shown in similar ensemble-teachers models. Furthermore, it is found that moving the ensemble teachers even in the steady state, in contrast to the fixed ensemble teachers, is efficient for the performance of the student.

  9. Ensemble Deep Learning for Biomedical Time Series Classification

    Directory of Open Access Journals (Sweden)

    Lin-peng Jin

    2016-01-01

    Full Text Available Ensemble learning has been proved to improve the generalization ability effectively in both theory and practice. In this paper, we briefly outline the current status of research on it first. Then, a new deep neural network-based ensemble method that integrates filtering views, local views, distorted views, explicit training, implicit training, subview prediction, and Simple Average is proposed for biomedical time series classification. Finally, we validate its effectiveness on the Chinese Cardiovascular Disease Database containing a large number of electrocardiogram recordings. The experimental results show that the proposed method has certain advantages compared to some well-known ensemble methods, such as Bagging and AdaBoost.

  10. Evaluation of medium-range ensemble flood forecasting based on calibration strategies and ensemble methods in Lanjiang Basin, Southeast China

    Science.gov (United States)

    Liu, Li; Gao, Chao; Xuan, Weidong; Xu, Yue-Ping

    2017-11-01

    Ensemble flood forecasts by hydrological models using numerical weather prediction products as forcing data are becoming more commonly used in operational flood forecasting applications. In this study, a hydrological ensemble flood forecasting system comprised of an automatically calibrated Variable Infiltration Capacity model and quantitative precipitation forecasts from TIGGE dataset is constructed for Lanjiang Basin, Southeast China. The impacts of calibration strategies and ensemble methods on the performance of the system are then evaluated. The hydrological model is optimized by the parallel programmed ε-NSGA II multi-objective algorithm. According to the solutions by ε-NSGA II, two differently parameterized models are determined to simulate daily flows and peak flows at each of the three hydrological stations. Then a simple yet effective modular approach is proposed to combine these daily and peak flows at the same station into one composite series. Five ensemble methods and various evaluation metrics are adopted. The results show that ε-NSGA II can provide an objective determination on parameter estimation, and the parallel program permits a more efficient simulation. It is also demonstrated that the forecasts from ECMWF have more favorable skill scores than other Ensemble Prediction Systems. The multimodel ensembles have advantages over all the single model ensembles and the multimodel methods weighted on members and skill scores outperform other methods. Furthermore, the overall performance at three stations can be satisfactory up to ten days, however the hydrological errors can degrade the skill score by approximately 2 days, and the influence persists until a lead time of 10 days with a weakening trend. With respect to peak flows selected by the Peaks Over Threshold approach, the ensemble means from single models or multimodels are generally underestimated, indicating that the ensemble mean can bring overall improvement in forecasting of flows. For

  11. On Ensemble Nonlinear Kalman Filtering with Symmetric Analysis Ensembles

    KAUST Repository

    Luo, Xiaodong; Hoteit, Ibrahim; Moroz, Irene M.

    2010-01-01

    However, by adopting the Monte Carlo method, the EnSRF also incurs certain sampling errors. One way to alleviate this problem is to introduce certain symmetry to the ensembles, which can reduce the sampling errors and spurious modes in evaluation of the means and covariances of the ensembles [7]. In this contribution, we present two methods to produce symmetric ensembles. One is based on the unscented transform [8, 9], which leads to the unscented Kalman filter (UKF) [8, 9] and its variant, the ensemble unscented Kalman filter (EnUKF) [7]. The other is based on Stirling’s interpolation formula (SIF), which results in the divided difference filter (DDF) [10]. Here we propose a simplified divided difference filter (sDDF) in the context of ensemble filtering. The similarity and difference between the sDDF and the EnUKF will be discussed. Numerical experiments will also be conducted to investigate the performance of the sDDF and the EnUKF, and compare them to a well‐established EnSRF, the ensemble transform Kalman filter (ETKF) [2].

  12. Decadal climate predictions improved by ocean ensemble dispersion filtering

    Science.gov (United States)

    Kadow, C.; Illing, S.; Kröner, I.; Ulbrich, U.; Cubasch, U.

    2017-06-01

    Decadal predictions by Earth system models aim to capture the state and phase of the climate several years in advance. Atmosphere-ocean interaction plays an important role for such climate forecasts. While short-term weather forecasts represent an initial value problem and long-term climate projections represent a boundary condition problem, the decadal climate prediction falls in-between these two time scales. In recent years, more precise initialization techniques of coupled Earth system models and increased ensemble sizes have improved decadal predictions. However, climate models in general start losing the initialized signal and its predictive skill from one forecast year to the next. Here we show that the climate prediction skill of an Earth system model can be improved by a shift of the ocean state toward the ensemble mean of its individual members at seasonal intervals. We found that this procedure, called ensemble dispersion filter, results in more accurate results than the standard decadal prediction. Global mean and regional temperature, precipitation, and winter cyclone predictions show an increased skill up to 5 years ahead. Furthermore, the novel technique outperforms predictions with larger ensembles and higher resolution. Our results demonstrate how decadal climate predictions benefit from ocean ensemble dispersion filtering toward the ensemble mean.Plain Language SummaryDecadal predictions aim to predict the climate several years in advance. Atmosphere-ocean interaction plays an important role for such climate forecasts. The ocean memory due to its heat capacity holds big potential skill. In recent years, more precise initialization techniques of coupled Earth system models (incl. atmosphere and ocean) have improved decadal predictions. Ensembles are another important aspect. Applying slightly perturbed predictions to trigger the famous butterfly effect results in an ensemble. Instead of evaluating one prediction, but the whole ensemble with its

  13. Assessing uncertainties in crop and pasture ensemble model simulations of productivity and N2 O emissions.

    Science.gov (United States)

    Ehrhardt, Fiona; Soussana, Jean-François; Bellocchi, Gianni; Grace, Peter; McAuliffe, Russel; Recous, Sylvie; Sándor, Renáta; Smith, Pete; Snow, Val; de Antoni Migliorati, Massimiliano; Basso, Bruno; Bhatia, Arti; Brilli, Lorenzo; Doltra, Jordi; Dorich, Christopher D; Doro, Luca; Fitton, Nuala; Giacomini, Sandro J; Grant, Brian; Harrison, Matthew T; Jones, Stephanie K; Kirschbaum, Miko U F; Klumpp, Katja; Laville, Patricia; Léonard, Joël; Liebig, Mark; Lieffering, Mark; Martin, Raphaël; Massad, Raia S; Meier, Elizabeth; Merbold, Lutz; Moore, Andrew D; Myrgiotis, Vasileios; Newton, Paul; Pattey, Elizabeth; Rolinski, Susanne; Sharp, Joanna; Smith, Ward N; Wu, Lianhai; Zhang, Qing

    2018-02-01

    Simulation models are extensively used to predict agricultural productivity and greenhouse gas emissions. However, the uncertainties of (reduced) model ensemble simulations have not been assessed systematically for variables affecting food security and climate change mitigation, within multi-species agricultural contexts. We report an international model comparison and benchmarking exercise, showing the potential of multi-model ensembles to predict productivity and nitrous oxide (N 2 O) emissions for wheat, maize, rice and temperate grasslands. Using a multi-stage modelling protocol, from blind simulations (stage 1) to partial (stages 2-4) and full calibration (stage 5), 24 process-based biogeochemical models were assessed individually or as an ensemble against long-term experimental data from four temperate grassland and five arable crop rotation sites spanning four continents. Comparisons were performed by reference to the experimental uncertainties of observed yields and N 2 O emissions. Results showed that across sites and crop/grassland types, 23%-40% of the uncalibrated individual models were within two standard deviations (SD) of observed yields, while 42 (rice) to 96% (grasslands) of the models were within 1 SD of observed N 2 O emissions. At stage 1, ensembles formed by the three lowest prediction model errors predicted both yields and N 2 O emissions within experimental uncertainties for 44% and 33% of the crop and grassland growth cycles, respectively. Partial model calibration (stages 2-4) markedly reduced prediction errors of the full model ensemble E-median for crop grain yields (from 36% at stage 1 down to 4% on average) and grassland productivity (from 44% to 27%) and to a lesser and more variable extent for N 2 O emissions. Yield-scaled N 2 O emissions (N 2 O emissions divided by crop yields) were ranked accurately by three-model ensembles across crop species and field sites. The potential of using process-based model ensembles to predict jointly

  14. On characteristic polynomials for a generalized chiral random matrix ensemble with a source

    Science.gov (United States)

    Fyodorov, Yan V.; Grela, Jacek; Strahov, Eugene

    2018-04-01

    We evaluate averages involving characteristic polynomials, inverse characteristic polynomials and ratios of characteristic polynomials for a N× N random matrix taken from a L-deformed chiral Gaussian Unitary Ensemble with an external source Ω. Relation to a recently studied statistics of bi-orthogonal eigenvectors in the complex Ginibre ensemble, see Fyodorov (2017 arXiv:1710.04699), is briefly discussed as a motivation to study asymptotics of these objects in the case of external source proportional to the identity matrix. In particular, for an associated complex bulk/chiral edge scaling regime we retrieve the kernel related to Bessel/Macdonald functions.

  15. Power law deformation of Wishart–Laguerre ensembles of random matrices

    International Nuclear Information System (INIS)

    Akemann, Gernot; Vivo, Pierpaolo

    2008-01-01

    We introduce a one-parameter deformation of the Wishart–Laguerre or chiral ensembles of positive definite random matrices with Dyson index β = 1,2 and 4. Our generalized model has a fat-tailed distribution while preserving the invariance under orthogonal, unitary or symplectic transformations. The spectral properties are derived analytically for finite matrix size N × M for all three values of β, in terms of the orthogonal polynomials of the standard Wishart–Laguerre ensembles. For large N in a certain double-scaling limit we obtain a generalized Marčenko–Pastur distribution on the macroscopic scale, and a generalized Bessel law at the hard edge which is shown to be universal. Both macroscopic and microscopic correlations exhibit power law tails, where the microscopic limit depends on β and the difference M−N. In the limit where our parameter governing the power law goes to infinity we recover the correlations of the Wishart–Laguerre ensembles. To illustrate these findings, the generalized Marčenko–Pastur distribution is shown to be in very good agreement with empirical data from financial covariance matrices

  16. Generalized-ensemble molecular dynamics and Monte Carlo algorithms beyond the limit of the multicanonical algorithm

    International Nuclear Information System (INIS)

    Okumura, Hisashi

    2010-01-01

    I review two new generalized-ensemble algorithms for molecular dynamics and Monte Carlo simulations of biomolecules, that is, the multibaric–multithermal algorithm and the partial multicanonical algorithm. In the multibaric–multithermal algorithm, two-dimensional random walks not only in the potential-energy space but also in the volume space are realized. One can discuss the temperature dependence and pressure dependence of biomolecules with this algorithm. The partial multicanonical simulation samples a wide range of only an important part of potential energy, so that one can concentrate the effort to determine a multicanonical weight factor only on the important energy terms. This algorithm has higher sampling efficiency than the multicanonical and canonical algorithms. (review)

  17. Integrated cumulus ensemble and turbulence (ICET): An integrated parameterization system for general circulation models (GCMs)

    Energy Technology Data Exchange (ETDEWEB)

    Evans, J.L.; Frank, W.M.; Young, G.S. [Pennsylvania State Univ., University Park, PA (United States)

    1996-04-01

    Successful simulations of the global circulation and climate require accurate representation of the properties of shallow and deep convective clouds, stable-layer clouds, and the interactions between various cloud types, the boundary layer, and the radiative fluxes. Each of these phenomena play an important role in the global energy balance, and each must be parameterized in a global climate model. These processes are highly interactive. One major problem limiting the accuracy of parameterizations of clouds and other processes in general circulation models (GCMs) is that most of the parameterization packages are not linked with a common physical basis. Further, these schemes have not, in general, been rigorously verified against observations adequate to the task of resolving subgrid-scale effects. To address these problems, we are designing a new Integrated Cumulus Ensemble and Turbulence (ICET) parameterization scheme, installing it in a climate model (CCM2), and evaluating the performance of the new scheme using data from Atmospheric Radiation Measurement (ARM) Program Cloud and Radiation Testbed (CART) sites.

  18. On Ensemble Nonlinear Kalman Filtering with Symmetric Analysis Ensembles

    KAUST Repository

    Luo, Xiaodong

    2010-09-19

    The ensemble square root filter (EnSRF) [1, 2, 3, 4] is a popular method for data assimilation in high dimensional systems (e.g., geophysics models). Essentially the EnSRF is a Monte Carlo implementation of the conventional Kalman filter (KF) [5, 6]. It is mainly different from the KF at the prediction steps, where it is some ensembles, rather then the means and covariance matrices, of the system state that are propagated forward. In doing this, the EnSRF is computationally more efficient than the KF, since propagating a covariance matrix forward in high dimensional systems is prohibitively expensive. In addition, the EnSRF is also very convenient in implementation. By propagating the ensembles of the system state, the EnSRF can be directly applied to nonlinear systems without any change in comparison to the assimilation procedures in linear systems. However, by adopting the Monte Carlo method, the EnSRF also incurs certain sampling errors. One way to alleviate this problem is to introduce certain symmetry to the ensembles, which can reduce the sampling errors and spurious modes in evaluation of the means and covariances of the ensembles [7]. In this contribution, we present two methods to produce symmetric ensembles. One is based on the unscented transform [8, 9], which leads to the unscented Kalman filter (UKF) [8, 9] and its variant, the ensemble unscented Kalman filter (EnUKF) [7]. The other is based on Stirling’s interpolation formula (SIF), which results in the divided difference filter (DDF) [10]. Here we propose a simplified divided difference filter (sDDF) in the context of ensemble filtering. The similarity and difference between the sDDF and the EnUKF will be discussed. Numerical experiments will also be conducted to investigate the performance of the sDDF and the EnUKF, and compare them to a well‐established EnSRF, the ensemble transform Kalman filter (ETKF) [2].

  19. A Link-Based Cluster Ensemble Approach For Improved Gene Expression Data Analysis

    Directory of Open Access Journals (Sweden)

    P.Balaji

    2015-01-01

    Full Text Available Abstract It is difficult from possibilities to select a most suitable effective way of clustering algorithm and its dataset for a defined set of gene expression data because we have a huge number of ways and huge number of gene expressions. At present many researchers are preferring to use hierarchical clustering in different forms this is no more totally optimal. Cluster ensemble research can solve this type of problem by automatically merging multiple data partitions from a wide range of different clusterings of any dimensions to improve both the quality and robustness of the clustering result. But we have many existing ensemble approaches using an association matrix to condense sample-cluster and co-occurrence statistics and relations within the ensemble are encapsulated only at raw level while the existing among clusters are totally discriminated. Finding these missing associations can greatly expand the capability of those ensemble methodologies for microarray data clustering. We propose general K-means cluster ensemble approach for the clustering of general categorical data into required number of partitions.

  20. MSEBAG: a dynamic classifier ensemble generation based on `minimum-sufficient ensemble' and bagging

    Science.gov (United States)

    Chen, Lei; Kamel, Mohamed S.

    2016-01-01

    In this paper, we propose a dynamic classifier system, MSEBAG, which is characterised by searching for the 'minimum-sufficient ensemble' and bagging at the ensemble level. It adopts an 'over-generation and selection' strategy and aims to achieve a good bias-variance trade-off. In the training phase, MSEBAG first searches for the 'minimum-sufficient ensemble', which maximises the in-sample fitness with the minimal number of base classifiers. Then, starting from the 'minimum-sufficient ensemble', a backward stepwise algorithm is employed to generate a collection of ensembles. The objective is to create a collection of ensembles with a descending fitness on the data, as well as a descending complexity in the structure. MSEBAG dynamically selects the ensembles from the collection for the decision aggregation. The extended adaptive aggregation (EAA) approach, a bagging-style algorithm performed at the ensemble level, is employed for this task. EAA searches for the competent ensembles using a score function, which takes into consideration both the in-sample fitness and the confidence of the statistical inference, and averages the decisions of the selected ensembles to label the test pattern. The experimental results show that the proposed MSEBAG outperforms the benchmarks on average.

  1. Deep-Sea Microbes: Linking Biogeochemical Rates to -Omics Approaches

    Science.gov (United States)

    Herndl, G. J.; Sintes, E.; Bayer, B.; Bergauer, K.; Amano, C.; Hansman, R.; Garcia, J.; Reinthaler, T.

    2016-02-01

    Over the past decade substantial progress has been made in determining deep ocean microbial activity and resolving some of the enigmas in understanding the deep ocean carbon flux. Also, metagenomics approaches have shed light onto the dark ocean's microbes but linking -omics approaches to biogeochemical rate measurements are generally rare in microbial oceanography and even more so for the deep ocean. In this presentation, we will show by combining metagenomics, -proteomics and biogeochemical rate measurements on the bulk and single-cell level that deep-sea microbes exhibit characteristics of generalists with a large genome repertoire, versatile in utilizing substrate as revealed by metaproteomics. This is in striking contrast with the apparently rather uniform dissolved organic matter pool in the deep ocean. Combining the different -omics approaches with metabolic rate measurements, we will highlight some major inconsistencies and enigmas in our understanding of the carbon cycling and microbial food web structure in the dark ocean.

  2. ABCD of Beta Ensembles and Topological Strings

    CERN Document Server

    Krefl, Daniel

    2012-01-01

    We study beta-ensembles with Bn, Cn, and Dn eigenvalue measure and their relation with refined topological strings. Our results generalize the familiar connections between local topological strings and matrix models leading to An measure, and illustrate that all those classical eigenvalue ensembles, and their topological string counterparts, are related one to another via various deformations and specializations, quantum shifts and discrete quotients. We review the solution of the Gaussian models via Macdonald identities, and interpret them as conifold theories. The interpolation between the various models is plainly apparent in this case. For general polynomial potential, we calculate the partition function in the multi-cut phase in a perturbative fashion, beyond tree-level in the large-N limit. The relation to refined topological string orientifolds on the corresponding local geometry is discussed along the way.

  3. Development of interactive graphic user interfaces for modeling reaction-based biogeochemical processes in batch systems with BIOGEOCHEM

    Science.gov (United States)

    Chang, C.; Li, M.; Yeh, G.

    2010-12-01

    The BIOGEOCHEM numerical model (Yeh and Fang, 2002; Fang et al., 2003) was developed with FORTRAN for simulating reaction-based geochemical and biochemical processes with mixed equilibrium and kinetic reactions in batch systems. A complete suite of reactions including aqueous complexation, adsorption/desorption, ion-exchange, redox, precipitation/dissolution, acid-base reactions, and microbial mediated reactions were embodied in this unique modeling tool. Any reaction can be treated as fast/equilibrium or slow/kinetic reaction. An equilibrium reaction is modeled with an implicit finite rate governed by a mass action equilibrium equation or by a user-specified algebraic equation. A kinetic reaction is modeled with an explicit finite rate with an elementary rate, microbial mediated enzymatic kinetics, or a user-specified rate equation. None of the existing models has encompassed this wide array of scopes. To ease the input/output learning curve using the unique feature of BIOGEOCHEM, an interactive graphic user interface was developed with the Microsoft Visual Studio and .Net tools. Several user-friendly features, such as pop-up help windows, typo warning messages, and on-screen input hints, were implemented, which are robust. All input data can be real-time viewed and automated to conform with the input file format of BIOGEOCHEM. A post-processor for graphic visualizations of simulated results was also embedded for immediate demonstrations. By following data input windows step by step, errorless BIOGEOCHEM input files can be created even if users have little prior experiences in FORTRAN. With this user-friendly interface, the time effort to conduct simulations with BIOGEOCHEM can be greatly reduced.

  4. Ensembl 2004.

    Science.gov (United States)

    Birney, E; Andrews, D; Bevan, P; Caccamo, M; Cameron, G; Chen, Y; Clarke, L; Coates, G; Cox, T; Cuff, J; Curwen, V; Cutts, T; Down, T; Durbin, R; Eyras, E; Fernandez-Suarez, X M; Gane, P; Gibbins, B; Gilbert, J; Hammond, M; Hotz, H; Iyer, V; Kahari, A; Jekosch, K; Kasprzyk, A; Keefe, D; Keenan, S; Lehvaslaiho, H; McVicker, G; Melsopp, C; Meidl, P; Mongin, E; Pettett, R; Potter, S; Proctor, G; Rae, M; Searle, S; Slater, G; Smedley, D; Smith, J; Spooner, W; Stabenau, A; Stalker, J; Storey, R; Ureta-Vidal, A; Woodwark, C; Clamp, M; Hubbard, T

    2004-01-01

    The Ensembl (http://www.ensembl.org/) database project provides a bioinformatics framework to organize biology around the sequences of large genomes. It is a comprehensive and integrated source of annotation of large genome sequences, available via interactive website, web services or flat files. As well as being one of the leading sources of genome annotation, Ensembl is an open source software engineering project to develop a portable system able to handle very large genomes and associated requirements. The facilities of the system range from sequence analysis to data storage and visualization and installations exist around the world both in companies and at academic sites. With a total of nine genome sequences available from Ensembl and more genomes to follow, recent developments have focused mainly on closer integration between genomes and external data.

  5. Towards a GME ensemble forecasting system: Ensemble initialization using the breeding technique

    Directory of Open Access Journals (Sweden)

    Jan D. Keller

    2008-12-01

    Full Text Available The quantitative forecast of precipitation requires a probabilistic background particularly with regard to forecast lead times of more than 3 days. As only ensemble simulations can provide useful information of the underlying probability density function, we built a new ensemble forecasting system (GME-EFS based on the GME model of the German Meteorological Service (DWD. For the generation of appropriate initial ensemble perturbations we chose the breeding technique developed by Toth and Kalnay (1993, 1997, which develops perturbations by estimating the regions of largest model error induced uncertainty. This method is applied and tested in the framework of quasi-operational forecasts for a three month period in 2007. The performance of the resulting ensemble forecasts are compared to the operational ensemble prediction systems ECMWF EPS and NCEP GFS by means of ensemble spread of free atmosphere parameters (geopotential and temperature and ensemble skill of precipitation forecasting. This comparison indicates that the GME ensemble forecasting system (GME-EFS provides reasonable forecasts with spread skill score comparable to that of the NCEP GFS. An analysis with the continuous ranked probability score exhibits a lack of resolution for the GME forecasts compared to the operational ensembles. However, with significant enhancements during the 3 month test period, the first results of our work with the GME-EFS indicate possibilities for further development as well as the potential for later operational usage.

  6. Shallow cumuli ensemble statistics for development of a stochastic parameterization

    Science.gov (United States)

    Sakradzija, Mirjana; Seifert, Axel; Heus, Thijs

    2014-05-01

    According to a conventional deterministic approach to the parameterization of moist convection in numerical atmospheric models, a given large scale forcing produces an unique response from the unresolved convective processes. This representation leaves out the small-scale variability of convection, as it is known from the empirical studies of deep and shallow convective cloud ensembles, there is a whole distribution of sub-grid states corresponding to the given large scale forcing. Moreover, this distribution gets broader with the increasing model resolution. This behavior is also consistent with our theoretical understanding of a coarse-grained nonlinear system. We propose an approach to represent the variability of the unresolved shallow-convective states, including the dependence of the sub-grid states distribution spread and shape on the model horizontal resolution. Starting from the Gibbs canonical ensemble theory, Craig and Cohen (2006) developed a theory for the fluctuations in a deep convective ensemble. The micro-states of a deep convective cloud ensemble are characterized by the cloud-base mass flux, which, according to the theory, is exponentially distributed (Boltzmann distribution). Following their work, we study the shallow cumulus ensemble statistics and the distribution of the cloud-base mass flux. We employ a Large-Eddy Simulation model (LES) and a cloud tracking algorithm, followed by a conditional sampling of clouds at the cloud base level, to retrieve the information about the individual cloud life cycles and the cloud ensemble as a whole. In the case of shallow cumulus cloud ensemble, the distribution of micro-states is a generalized exponential distribution. Based on the empirical and theoretical findings, a stochastic model has been developed to simulate the shallow convective cloud ensemble and to test the convective ensemble theory. Stochastic model simulates a compound random process, with the number of convective elements drawn from a

  7. Tailored Random Graph Ensembles

    International Nuclear Information System (INIS)

    Roberts, E S; Annibale, A; Coolen, A C C

    2013-01-01

    Tailored graph ensembles are a developing bridge between biological networks and statistical mechanics. The aim is to use this concept to generate a suite of rigorous tools that can be used to quantify and compare the topology of cellular signalling networks, such as protein-protein interaction networks and gene regulation networks. We calculate exact and explicit formulae for the leading orders in the system size of the Shannon entropies of random graph ensembles constrained with degree distribution and degree-degree correlation. We also construct an ergodic detailed balance Markov chain with non-trivial acceptance probabilities which converges to a strictly uniform measure and is based on edge swaps that conserve all degrees. The acceptance probabilities can be generalized to define Markov chains that target any alternative desired measure on the space of directed or undirected graphs, in order to generate graphs with more sophisticated topological features.

  8. NIMEFI: gene regulatory network inference using multiple ensemble feature importance algorithms.

    Directory of Open Access Journals (Sweden)

    Joeri Ruyssinck

    Full Text Available One of the long-standing open challenges in computational systems biology is the topology inference of gene regulatory networks from high-throughput omics data. Recently, two community-wide efforts, DREAM4 and DREAM5, have been established to benchmark network inference techniques using gene expression measurements. In these challenges the overall top performer was the GENIE3 algorithm. This method decomposes the network inference task into separate regression problems for each gene in the network in which the expression values of a particular target gene are predicted using all other genes as possible predictors. Next, using tree-based ensemble methods, an importance measure for each predictor gene is calculated with respect to the target gene and a high feature importance is considered as putative evidence of a regulatory link existing between both genes. The contribution of this work is twofold. First, we generalize the regression decomposition strategy of GENIE3 to other feature importance methods. We compare the performance of support vector regression, the elastic net, random forest regression, symbolic regression and their ensemble variants in this setting to the original GENIE3 algorithm. To create the ensemble variants, we propose a subsampling approach which allows us to cast any feature selection algorithm that produces a feature ranking into an ensemble feature importance algorithm. We demonstrate that the ensemble setting is key to the network inference task, as only ensemble variants achieve top performance. As second contribution, we explore the effect of using rankwise averaged predictions of multiple ensemble algorithms as opposed to only one. We name this approach NIMEFI (Network Inference using Multiple Ensemble Feature Importance algorithms and show that this approach outperforms all individual methods in general, although on a specific network a single method can perform better. An implementation of NIMEFI has been made

  9. Dynamics of heterogeneous oscillator ensembles in terms of collective variables

    Science.gov (United States)

    Pikovsky, Arkady; Rosenblum, Michael

    2011-04-01

    We consider general heterogeneous ensembles of phase oscillators, sine coupled to arbitrary external fields. Starting with the infinitely large ensembles, we extend the Watanabe-Strogatz theory, valid for identical oscillators, to cover the case of an arbitrary parameter distribution. The obtained equations yield the description of the ensemble dynamics in terms of collective variables and constants of motion. As a particular case of the general setup we consider hierarchically organized ensembles, consisting of a finite number of subpopulations, whereas the number of elements in a subpopulation can be both finite or infinite. Next, we link the Watanabe-Strogatz and Ott-Antonsen theories and demonstrate that the latter one corresponds to a particular choice of constants of motion. The approach is applied to the standard Kuramoto-Sakaguchi model, to its extension for the case of nonlinear coupling, and to the description of two interacting subpopulations, exhibiting a chimera state. With these examples we illustrate that, although the asymptotic dynamics can be found within the framework of the Ott-Antonsen theory, the transients depend on the constants of motion. The most dramatic effect is the dependence of the basins of attraction of different synchronous regimes on the initial configuration of phases.

  10. Kohn-Sham Theory for Ground-State Ensembles

    International Nuclear Information System (INIS)

    Ullrich, C. A.; Kohn, W.

    2001-01-01

    An electron density distribution n(r) which can be represented by that of a single-determinant ground state of noninteracting electrons in an external potential v(r) is called pure-state v -representable (P-VR). Most physical electronic systems are P-VR. Systems which require a weighted sum of several such determinants to represent their density are called ensemble v -representable (E-VR). This paper develops formal Kohn-Sham equations for E-VR physical systems, using the appropriate coupling constant integration. It also derives local density- and generalized gradient approximations, and conditions and corrections specific to ensembles

  11. Stream biogeochemical resilience in the age of Anthropocene

    Science.gov (United States)

    Dong, H.; Creed, I. F.

    2017-12-01

    Recent evidence indicates that biogeochemical cycles are being pushed beyond the tolerance limits of the earth system in the age of the Anthropocene placing terrestrial and aquatic ecosystems at risk. Here, we explored the question: Is there empirical evidence of global atmospheric changes driving losses in stream biogeochemical resilience towards a new normal? Stream biogeochemical resilience is the process of returning to equilibrium conditions after a disturbance and can be measured using three metrics: reactivity (the highest initial response after a disturbance), return rate (the rate of return to equilibrium condition after reactive changes), and variance of the stationary distribution (the signal to noise ratio). Multivariate autoregressive models were used to derive the three metrics for streams along a disturbance gradient - from natural systems where global drivers would dominate, to relatively managed or modified systems where global and local drivers would interact. We observed a loss of biogeochemical resilience in all streams. The key biogeochemical constituent(s) that may be driving loss of biogeochemical resilience were identified from the time series of the stream biogeochemical constituents. Non-stationary trends (detected by Mann-Kendall analysis) and stationary cycles (revealed through Morlet wavelet analysis) were removed, and the standard deviation (SD) of the remaining residuals were analyzed to determine if there was an increase in SD over time that would indicate a pending shift towards a new normal. We observed that nitrate-N and total phosphorus showed behaviours indicative of a pending shift in natural and managed forest systems, but not in agricultural systems. This study provides empirical support that stream ecosystems are showing signs of exceeding planetary boundary tolerance levels and shifting towards a "new normal" in response to global changes, which can be exacerbated by local management activities. Future work will consider

  12. Universal critical wrapping probabilities in the canonical ensemble

    Directory of Open Access Journals (Sweden)

    Hao Hu

    2015-09-01

    Full Text Available Universal dimensionless quantities, such as Binder ratios and wrapping probabilities, play an important role in the study of critical phenomena. We study the finite-size scaling behavior of the wrapping probability for the Potts model in the random-cluster representation, under the constraint that the total number of occupied bonds is fixed, so that the canonical ensemble applies. We derive that, in the limit L→∞, the critical values of the wrapping probability are different from those of the unconstrained model, i.e. the model in the grand-canonical ensemble, but still universal, for systems with 2yt−d>0 where yt=1/ν is the thermal renormalization exponent and d is the spatial dimension. Similar modifications apply to other dimensionless quantities, such as Binder ratios. For systems with 2yt−d≤0, these quantities share same critical universal values in the two ensembles. It is also derived that new finite-size corrections are induced. These findings apply more generally to systems in the canonical ensemble, e.g. the dilute Potts model with a fixed total number of vacancies. Finally, we formulate an efficient cluster-type algorithm for the canonical ensemble, and confirm these predictions by extensive simulations.

  13. Statistical properties of many-particle spectra. IV. New ensembles by Stieltjes transform methods

    International Nuclear Information System (INIS)

    Pandey, A.

    1981-01-01

    New Gaussian matrix ensembles, with arbitrary centroids and variances for the matrix elements, are defined as modifications of the three standard ones: GOE, GUE and GSE. The average density and two-point correlation function are given in the general case in terms of the corresponding Stieltjes transforms, first used by Pastur for the density. It is shown for the centroid-modified ensemble K+αH that when the operator K preserves the underlying symmetries of the standard ensemble H, then, as the magnitude of α grows, the transition of the fluctuations to those of H is very rapid and discontinuous in the limit of asymptotic dimensionality. Corresponding results are found for other ensembles. A similar Dyson result for the effects of the breaking of a model symmetry on the fluctuations is generalized to any model symmetry, as well as to the fundamental symmetries such as time-reversed invariance

  14. An automated approach to network features of protein structure ensembles

    Science.gov (United States)

    Bhattacharyya, Moitrayee; Bhat, Chanda R; Vishveshwara, Saraswathi

    2013-01-01

    Network theory applied to protein structures provides insights into numerous problems of biological relevance. The explosion in structural data available from PDB and simulations establishes a need to introduce a standalone-efficient program that assembles network concepts/parameters under one hood in an automated manner. Herein, we discuss the development/application of an exhaustive, user-friendly, standalone program package named PSN-Ensemble, which can handle structural ensembles generated through molecular dynamics (MD) simulation/NMR studies or from multiple X-ray structures. The novelty in network construction lies in the explicit consideration of side-chain interactions among amino acids. The program evaluates network parameters dealing with topological organization and long-range allosteric communication. The introduction of a flexible weighing scheme in terms of residue pairwise cross-correlation/interaction energy in PSN-Ensemble brings in dynamical/chemical knowledge into the network representation. Also, the results are mapped on a graphical display of the structure, allowing an easy access of network analysis to a general biological community. The potential of PSN-Ensemble toward examining structural ensemble is exemplified using MD trajectories of an ubiquitin-conjugating enzyme (UbcH5b). Furthermore, insights derived from network parameters evaluated using PSN-Ensemble for single-static structures of active/inactive states of β2-adrenergic receptor and the ternary tRNA complexes of tyrosyl tRNA synthetases (from organisms across kingdoms) are discussed. PSN-Ensemble is freely available from http://vishgraph.mbu.iisc.ernet.in/PSN-Ensemble/psn_index.html. PMID:23934896

  15. Embedded random matrix ensembles in quantum physics

    CERN Document Server

    Kota, V K B

    2014-01-01

    Although used with increasing frequency in many branches of physics, random matrix ensembles are not always sufficiently specific to account for important features of the physical system at hand. One refinement which retains the basic stochastic approach but allows for such features consists in the use of embedded ensembles.  The present text is an exhaustive introduction to and survey of this important field. Starting with an easy-to-read introduction to general random matrix theory, the text then develops the necessary concepts from the beginning, accompanying the reader to the frontiers of present-day research. With some notable exceptions, to date these ensembles have primarily been applied in nuclear spectroscopy. A characteristic example is the use of a random two-body interaction in the framework of the nuclear shell model. Yet, topics in atomic physics, mesoscopic physics, quantum information science and statistical mechanics of isolated finite quantum systems can also be addressed using these ensemb...

  16. Ensemble-free configurational temperature for spin systems

    Science.gov (United States)

    Palma, G.; Gutiérrez, G.; Davis, S.

    2016-12-01

    An estimator for the dynamical temperature in an arbitrary ensemble is derived in the framework of the conjugate variables theorem. We prove directly that its average indeed gives the inverse temperature and that it is independent of the ensemble. We test this estimator numerically by a simulation of the two-dimensional X Y model in the canonical ensemble. As this model is critical in the whole region of temperatures below the Berezinski-Kosterlitz-Thouless critical temperature TBKT, we use a generalization of Wolff's unicluster algorithm. The numerical results allow us to confirm the robustness of the analytical expression for the microscopic estimator of the temperature. This microscopic estimator has also the advantage that it gives a direct measure of the thermalization process and can be used to compute absolute errors associated with statistical fluctuations. In consequence, this estimator allows for a direct, absolute, and stringent test of the ergodicity of the underlying Markov process, which encodes the algorithm used in a numerical simulation.

  17. Ensembl variation resources

    Directory of Open Access Journals (Sweden)

    Marin-Garcia Pablo

    2010-05-01

    Full Text Available Abstract Background The maturing field of genomics is rapidly increasing the number of sequenced genomes and producing more information from those previously sequenced. Much of this additional information is variation data derived from sampling multiple individuals of a given species with the goal of discovering new variants and characterising the population frequencies of the variants that are already known. These data have immense value for many studies, including those designed to understand evolution and connect genotype to phenotype. Maximising the utility of the data requires that it be stored in an accessible manner that facilitates the integration of variation data with other genome resources such as gene annotation and comparative genomics. Description The Ensembl project provides comprehensive and integrated variation resources for a wide variety of chordate genomes. This paper provides a detailed description of the sources of data and the methods for creating the Ensembl variation databases. It also explores the utility of the information by explaining the range of query options available, from using interactive web displays, to online data mining tools and connecting directly to the data servers programmatically. It gives a good overview of the variation resources and future plans for expanding the variation data within Ensembl. Conclusions Variation data is an important key to understanding the functional and phenotypic differences between individuals. The development of new sequencing and genotyping technologies is greatly increasing the amount of variation data known for almost all genomes. The Ensembl variation resources are integrated into the Ensembl genome browser and provide a comprehensive way to access this data in the context of a widely used genome bioinformatics system. All Ensembl data is freely available at http://www.ensembl.org and from the public MySQL database server at ensembldb.ensembl.org.

  18. NYYD Ensemble

    Index Scriptorium Estoniae

    2002-01-01

    NYYD Ensemble'i duost Traksmann - Lukk E.-S. Tüüri teosega "Symbiosis", mis on salvestatud ka hiljuti ilmunud NYYD Ensemble'i CDle. 2. märtsil Rakvere Teatri väikeses saalis ja 3. märtsil Rotermanni Soolalaos, kavas Tüür, Kaumann, Berio, Reich, Yun, Hauta-aho, Buckinx

  19. A framework to assess biogeochemical response to ecosystem disturbance using nutrient partitioning ratios

    Science.gov (United States)

    Kranabetter, J. Marty; McLauchlan, Kendra K.; Enders, Sara K.; Fraterrigo, Jennifer M.; Higuera, Philip E.; Morris, Jesse L.; Rastetter, Edward B.; Barnes, Rebecca; Buma, Brian; Gavin, Daniel G.; Gerhart, Laci M.; Gillson, Lindsey; Hietz, Peter; Mack, Michelle C.; McNeil, Brenden; Perakis, Steven

    2016-01-01

    Disturbances affect almost all terrestrial ecosystems, but it has been difficult to identify general principles regarding these influences. To improve our understanding of the long-term consequences of disturbance on terrestrial ecosystems, we present a conceptual framework that analyzes disturbances by their biogeochemical impacts. We posit that the ratio of soil and plant nutrient stocks in mature ecosystems represents a characteristic site property. Focusing on nitrogen (N), we hypothesize that this partitioning ratio (soil N: plant N) will undergo a predictable trajectory after disturbance. We investigate the nature of this partitioning ratio with three approaches: (1) nutrient stock data from forested ecosystems in North America, (2) a process-based ecosystem model, and (3) conceptual shifts in site nutrient availability with altered disturbance frequency. Partitioning ratios could be applied to a variety of ecosystems and successional states, allowing for improved temporal scaling of disturbance events. The generally short-term empirical evidence for recovery trajectories of nutrient stocks and partitioning ratios suggests two areas for future research. First, we need to recognize and quantify how disturbance effects can be accreting or depleting, depending on whether their net effect is to increase or decrease ecosystem nutrient stocks. Second, we need to test how altered disturbance frequencies from the present state may be constructive or destructive in their effects on biogeochemical cycling and nutrient availability. Long-term studies, with repeated sampling of soils and vegetation, will be essential in further developing this framework of biogeochemical response to disturbance.

  20. Reliability of multi-model and structurally different single-model ensembles

    Energy Technology Data Exchange (ETDEWEB)

    Yokohata, Tokuta [National Institute for Environmental Studies, Center for Global Environmental Research, Tsukuba, Ibaraki (Japan); Annan, James D.; Hargreaves, Julia C. [Japan Agency for Marine-Earth Science and Technology, Research Institute for Global Change, Yokohama, Kanagawa (Japan); Collins, Matthew [University of Exeter, College of Engineering, Mathematics and Physical Sciences, Exeter (United Kingdom); Jackson, Charles S.; Tobis, Michael [The University of Texas at Austin, Institute of Geophysics, 10100 Burnet Rd., ROC-196, Mail Code R2200, Austin, TX (United States); Webb, Mark J. [Met Office Hadley Centre, Exeter (United Kingdom)

    2012-08-15

    The performance of several state-of-the-art climate model ensembles, including two multi-model ensembles (MMEs) and four structurally different (perturbed parameter) single model ensembles (SMEs), are investigated for the first time using the rank histogram approach. In this method, the reliability of a model ensemble is evaluated from the point of view of whether the observations can be regarded as being sampled from the ensemble. Our analysis reveals that, in the MMEs, the climate variables we investigated are broadly reliable on the global scale, with a tendency towards overdispersion. On the other hand, in the SMEs, the reliability differs depending on the ensemble and variable field considered. In general, the mean state and historical trend of surface air temperature, and mean state of precipitation are reliable in the SMEs. However, variables such as sea level pressure or top-of-atmosphere clear-sky shortwave radiation do not cover a sufficiently wide range in some. It is not possible to assess whether this is a fundamental feature of SMEs generated with particular model, or a consequence of the algorithm used to select and perturb the values of the parameters. As under-dispersion is a potentially more serious issue when using ensembles to make projections, we recommend the application of rank histograms to assess reliability when designing and running perturbed physics SMEs. (orig.)

  1. Improving Climate Projections Using "Intelligent" Ensembles

    Science.gov (United States)

    Baker, Noel C.; Taylor, Patrick C.

    2015-01-01

    Recent changes in the climate system have led to growing concern, especially in communities which are highly vulnerable to resource shortages and weather extremes. There is an urgent need for better climate information to develop solutions and strategies for adapting to a changing climate. Climate models provide excellent tools for studying the current state of climate and making future projections. However, these models are subject to biases created by structural uncertainties. Performance metrics-or the systematic determination of model biases-succinctly quantify aspects of climate model behavior. Efforts to standardize climate model experiments and collect simulation data-such as the Coupled Model Intercomparison Project (CMIP)-provide the means to directly compare and assess model performance. Performance metrics have been used to show that some models reproduce present-day climate better than others. Simulation data from multiple models are often used to add value to projections by creating a consensus projection from the model ensemble, in which each model is given an equal weight. It has been shown that the ensemble mean generally outperforms any single model. It is possible to use unequal weights to produce ensemble means, in which models are weighted based on performance (called "intelligent" ensembles). Can performance metrics be used to improve climate projections? Previous work introduced a framework for comparing the utility of model performance metrics, showing that the best metrics are related to the variance of top-of-atmosphere outgoing longwave radiation. These metrics improve present-day climate simulations of Earth's energy budget using the "intelligent" ensemble method. The current project identifies several approaches for testing whether performance metrics can be applied to future simulations to create "intelligent" ensemble-mean climate projections. It is shown that certain performance metrics test key climate processes in the models, and

  2. Problems of a Statistical Ensemble Theory for Systems Far from Equilibrium

    Science.gov (United States)

    Ebeling, Werner

    The development of a general statistical physics of nonequilibrium systems was one of the main unfinished tasks of statistical physics of the 20th century. The aim of this work is the study of a special class of nonequilibrium systems where the formulation of an ensemble theory of some generality is possible. These are the so-called canonical-dissipative systems, where the driving terms are determined by invariants of motion. We construct canonical-dissipative systems which are ergodic on certain surfaces on the phase plane. These systems may be described by a non-equilibrium microcanocical ensemble, corresponding to an equal distribution on the target surface. Next we construct and solve Fokker-Planck equations; this leads to a kind of canonical-dissipative ensemble. In the last part we discuss the thoretical problem how to define bifurcations in the framework of nonequilibrium statistics and several possible applications.

  3. 'Lazy' quantum ensembles

    International Nuclear Information System (INIS)

    Parfionov, George; Zapatrin, Roman

    2006-01-01

    We compare different strategies aimed to prepare an ensemble with a given density matrix ρ. Preparing the ensemble of eigenstates of ρ with appropriate probabilities can be treated as 'generous' strategy: it provides maximal accessible information about the state. Another extremity is the so-called 'Scrooge' ensemble, which is mostly stingy in sharing the information. We introduce 'lazy' ensembles which require minimal effort to prepare the density matrix by selecting pure states with respect to completely random choice. We consider two parties, Alice and Bob, playing a kind of game. Bob wishes to guess which pure state is prepared by Alice. His null hypothesis, based on the lack of any information about Alice's intention, is that Alice prepares any pure state with equal probability. Then, the average quantum state measured by Bob turns out to be ρ, and he has to make a new hypothesis about Alice's intention solely based on the information that the observed density matrix is ρ. The arising 'lazy' ensemble is shown to be the alternative hypothesis which minimizes type I error

  4. Ecotoxicological, ecophysiological and biogeochemical fundamentals of risk assessment

    International Nuclear Information System (INIS)

    Bashkin, V.; Evstafjeva, E.

    1995-01-01

    A quantitative risk assessment (RA) for complex influence of different factors in heavy polluted regions is possible to carry out only on a basis of determination of various links of biogeochemical trophical chains and analysis of the whole biogeochemical structure of the region under study. As an integrative assessment, the human adaptability should be chosen because the majority of trophical chains are closed by man. The given integrative criteria includes biogeochemical, ecophysiological and ecotoxicological assessment of risk factors. Consequently, ecological-biogeochemical regionalization, ecophysiological and ecotoxicological monitoring of human population health are the important approaches to RA. These criteria should be conjugated with LCA of various industrial and agricultural products. At the ultimate degree, the given approaches are needed for areas where traditional pollutants (heavy metals, POPS, pesticides, fertilizers) are enforced sharply by radioactive pollution. Due to the complex influence of pollutants, it is impossible to use individual guidelines. For RA of these complex pollutants, the methods of human adaptability assessment to a polluted environment have to be carried out. These methods include biogeochemical, ecotoxicological and ecophysiological analysis of risk factors as well as quantitative uncertainty analysis. Furthermore, the modern statistical methods such as correlative graphs etc., have to be used for quantitative assessment of human adaptability to complex influence of pollutants. The results obtained in the Chernobyl region have shown the acceptability of suggested methods

  5. Disturbance decouples biogeochemical cycles across forests of the southeastern US

    Science.gov (United States)

    Ashley D. Keiser; Jennifer D. Knoepp; Mark A. Bradford

    2016-01-01

    Biogeochemical cycles are inherently linked through the stoichiometric demands of the organisms that cycle the elements. Landscape disturbance can alter element availability and thus the rates of biogeochemical cycling. Nitrification is a fundamental biogeochemical process positively related to plant productivity and nitrogen loss from soils to aquatic systems, and the...

  6. Imprinting and recalling cortical ensembles.

    Science.gov (United States)

    Carrillo-Reid, Luis; Yang, Weijian; Bando, Yuki; Peterka, Darcy S; Yuste, Rafael

    2016-08-12

    Neuronal ensembles are coactive groups of neurons that may represent building blocks of cortical circuits. These ensembles could be formed by Hebbian plasticity, whereby synapses between coactive neurons are strengthened. Here we report that repetitive activation with two-photon optogenetics of neuronal populations from ensembles in the visual cortex of awake mice builds neuronal ensembles that recur spontaneously after being imprinted and do not disrupt preexisting ones. Moreover, imprinted ensembles can be recalled by single- cell stimulation and remain coactive on consecutive days. Our results demonstrate the persistent reconfiguration of cortical circuits by two-photon optogenetics into neuronal ensembles that can perform pattern completion. Copyright © 2016, American Association for the Advancement of Science.

  7. World Music Ensemble: Kulintang

    Science.gov (United States)

    Beegle, Amy C.

    2012-01-01

    As instrumental world music ensembles such as steel pan, mariachi, gamelan and West African drums are becoming more the norm than the exception in North American school music programs, there are other world music ensembles just starting to gain popularity in particular parts of the United States. The kulintang ensemble, a drum and gong ensemble…

  8. Regional interdependency of precipitation indices across Denmark in two ensembles of high-resolution RCMs

    DEFF Research Database (Denmark)

    Sunyer Pinya, Maria Antonia; Madsen, Henrik; Rosbjerg, Dan

    2013-01-01

    all these methods is that the climate models are independent. This study addresses the validity of this assumption for two ensembles of regional climate models (RCMs) from the Ensemble-Based Predictions of Climate Changes and their Impacts (ENSEMBLES) project based on the land cells covering Denmark....... Daily precipitation indices from an ensemble of RCMs driven by the 40-yrECMWFRe-Analysis (ERA-40) and an ensemble of the same RCMs driven by different general circulation models (GCMs) are analyzed. Two different methods are used to estimate the amount of independent information in the ensembles....... These are based on different statistical properties of a measure of climate model error. Additionally, a hierarchical cluster analysis is carried out. Regardless of the method used, the effective number of RCMs is smaller than the total number of RCMs. The estimated effective number of RCMs varies depending...

  9. Mechanistic site-based emulation of a global ocean biogeochemical model (MEDUSA 1.0 for parametric analysis and calibration: an application of the Marine Model Optimization Testbed (MarMOT 1.1

    Directory of Open Access Journals (Sweden)

    J. C. P. Hemmings

    2015-03-01

    Full Text Available Biogeochemical ocean circulation models used to investigate the role of plankton ecosystems in global change rely on adjustable parameters to capture the dominant biogeochemical dynamics of a complex biological system. In principle, optimal parameter values can be estimated by fitting models to observational data, including satellite ocean colour products such as chlorophyll that achieve good spatial and temporal coverage of the surface ocean. However, comprehensive parametric analyses require large ensemble experiments that are computationally infeasible with global 3-D simulations. Site-based simulations provide an efficient alternative but can only be used to make reliable inferences about global model performance if robust quantitative descriptions of their relationships with the corresponding 3-D simulations can be established. The feasibility of establishing such a relationship is investigated for an intermediate complexity biogeochemistry model (MEDUSA coupled with a widely used global ocean model (NEMO. A site-based mechanistic emulator is constructed for surface chlorophyll output from this target model as a function of model parameters. The emulator comprises an array of 1-D simulators and a statistical quantification of the uncertainty in their predictions. The unknown parameter-dependent biogeochemical environment, in terms of initial tracer concentrations and lateral flux information required by the simulators, is a significant source of uncertainty. It is approximated by a mean environment derived from a small ensemble of 3-D simulations representing variability of the target model behaviour over the parameter space of interest. The performance of two alternative uncertainty quantification schemes is examined: a direct method based on comparisons between simulator output and a sample of known target model "truths" and an indirect method that is only partially reliant on knowledge of the target model output. In general, chlorophyll

  10. Green Infrastructure Increases Biogeochemical Responsiveness, Vegetation Growth and Decreases Runoff in a Semi-Arid City, Tucson, AZ, USA

    Science.gov (United States)

    Meixner, T.; Papuga, S. A.; Luketich, A. M.; Rockhill, T.; Gallo, E. L.; Anderson, J.; Salgado, L.; Pope, K.; Gupta, N.; Korgaonkar, Y.; Guertin, D. P.

    2017-12-01

    Green Infrastructure (GI) is often viewed as a mechanism to minimize the effects of urbanization on hydrology, water quality, and other ecosystem services (including the urban heat island). Quantifying the effects of GI requires field measurements of the dimensions of biogeochemical, ecosystem, and hydrologic function that we expect GI to impact. Here we investigated the effect of GI features in Tucson, Arizona which has a low intensity winter precipitation regime and a high intensity summer regime. We focused on understanding the effect of GI on soil hydraulic and biogeochemical properties as well as the effect on vegetation and canopy temperature. Our results demonstrate profound changes in biogeochemical and hydrologic properties and vegetation growth between GI systems and nearby control sites. In terms of hydrologic properties GI soils had increased water holding capacity and hydraulic conductivity. GI soils also have higher total carbon, total nitrogen, and organic matter in general than control soils. Furthermore, we tested the sampled soils (control and GI) for differences in biogeochemical response upon wetting. GI soils had larger respiration responses indicating greater biogeochemical activity overall. Long-term Lidar surveys were used to investigate the differential canopy growth of GI systems versus control sites. The results of this analysis indicate that while a significant amount of time is needed to observe differences in canopy growth GI features due increase tree size and thus likely impact street scale ambient temperatures. Additionally monitoring of transpiration, soil moisture, and canopy temperature demonstrates that GI features increase vegetation growth and transpiration and reduce canopy temperatures. These biogeochemical and ecohydrologic results indicate that GI can increase the biogeochemical processing of soils and increase tree growth and thus reduce urban ambient temperatures.

  11. On the v-representability of ensemble densities of electron systems

    Science.gov (United States)

    Gonis, A.; Däne, M.

    2018-05-01

    Analogously to the case at zero temperature, where the density of the ground state of an interacting many-particle system determines uniquely (within an arbitrary additive constant) the external potential acting on the system, the thermal average of the density over an ensemble defined by the Boltzmann distribution at the minimum of the thermodynamic potential, or the free energy, determines the external potential uniquely (and not just modulo a constant) acting on a system described by this thermodynamic potential or free energy. The paper describes a formal procedure that generates the domain of a constrained search over general ensembles (at zero or elevated temperatures) that lead to a given density, including as a special case a density thermally averaged at a given temperature, and in the case of a v-representable density determines the external potential leading to the ensemble density. As an immediate consequence of the general formalism, the concept of v-representability is extended beyond the hitherto discussed case of ground state densities to encompass excited states as well. Specific application to thermally averaged densities solves the v-representability problem in connection with the Mermin functional in a manner analogous to that in which this problem was recently settled with respect to the Hohenberg and Kohn functional. The main formalism is illustrated with numerical results for ensembles of one-dimensional, non-interacting systems of particles under a harmonic potential.

  12. Microcanonical ensemble extensive thermodynamics of Tsallis statistics

    International Nuclear Information System (INIS)

    Parvan, A.S.

    2005-01-01

    The microscopic foundation of the generalized equilibrium statistical mechanics based on the Tsallis entropy is given by using the Gibbs idea of statistical ensembles of the classical and quantum mechanics.The equilibrium distribution functions are derived by the thermodynamic method based upon the use of the fundamental equation of thermodynamics and the statistical definition of the functions of the state of the system. It is shown that if the entropic index ξ = 1/q - 1 in the microcanonical ensemble is an extensive variable of the state of the system, then in the thermodynamic limit z bar = 1/(q - 1)N = const the principle of additivity and the zero law of thermodynamics are satisfied. In particular, the Tsallis entropy of the system is extensive and the temperature is intensive. Thus, the Tsallis statistics completely satisfies all the postulates of the equilibrium thermodynamics. Moreover, evaluation of the thermodynamic identities in the microcanonical ensemble is provided by the Euler theorem. The principle of additivity and the Euler theorem are explicitly proved by using the illustration of the classical microcanonical ideal gas in the thermodynamic limit

  13. Microcanonical ensemble extensive thermodynamics of Tsallis statistics

    International Nuclear Information System (INIS)

    Parvan, A.S.

    2006-01-01

    The microscopic foundation of the generalized equilibrium statistical mechanics based on the Tsallis entropy is given by using the Gibbs idea of statistical ensembles of the classical and quantum mechanics. The equilibrium distribution functions are derived by the thermodynamic method based upon the use of the fundamental equation of thermodynamics and the statistical definition of the functions of the state of the system. It is shown that if the entropic index ξ=1/(q-1) in the microcanonical ensemble is an extensive variable of the state of the system, then in the thermodynamic limit z-bar =1/(q-1)N=const the principle of additivity and the zero law of thermodynamics are satisfied. In particular, the Tsallis entropy of the system is extensive and the temperature is intensive. Thus, the Tsallis statistics completely satisfies all the postulates of the equilibrium thermodynamics. Moreover, evaluation of the thermodynamic identities in the microcanonical ensemble is provided by the Euler theorem. The principle of additivity and the Euler theorem are explicitly proved by using the illustration of the classical microcanonical ideal gas in the thermodynamic limit

  14. A molecular dynamics algorithm for simulation of field theories in the canonical ensemble

    International Nuclear Information System (INIS)

    Kogut, J.B.; Sinclair, D.K.

    1986-01-01

    We add a single scalar degree of freedom (''demon'') to the microcanonical ensemble which converts its molecular dynamics into a simulation method for the canonical ensemble (euclidean path integral) of the underlying field theory. This generalization of the microcanonical molecular dynamics algorithm simulates the field theory at fixed coupling with a completely deterministic procedure. We discuss the finite size effects of the method, the equipartition theorem and ergodicity. The method is applied to the planar model in two dimensions and SU(3) lattice gauge theory with four species of light, dynamical quarks in four dimensions. The method is much less sensitive to its discrete time step than conventional Langevin equation simulations of the canonical ensemble. The method is a straightforward generalization of a procedure introduced by S. Nose for molecular physics. (orig.)

  15. Breaking of ensembles of linear and nonlinear oscillators

    International Nuclear Information System (INIS)

    Buts, V.A.

    2016-01-01

    Some results concerning the study of the dynamics of ensembles of linear and nonlinear oscillators are stated. It is shown that, in general, a stable ensemble of linear oscillator has a limited number of oscillators. This number has been defined for some simple models. It is shown that the features of the dynamics of linear oscillators can be used for conversion of the low-frequency energy oscillations into high frequency oscillations. The dynamics of coupled nonlinear oscillators in most cases is chaotic. For such a case, it is shown that the statistical characteristics (moments) of chaotic motion can significantly reduce potential barriers that keep the particles in the capture region

  16. Interpolation of property-values between electron numbers is inconsistent with ensemble averaging

    Energy Technology Data Exchange (ETDEWEB)

    Miranda-Quintana, Ramón Alain [Laboratory of Computational and Theoretical Chemistry, Faculty of Chemistry, University of Havana, Havana (Cuba); Department of Chemistry and Chemical Biology, McMaster University, Hamilton, Ontario L8S 4M1 (Canada); Ayers, Paul W. [Department of Chemistry and Chemical Biology, McMaster University, Hamilton, Ontario L8S 4M1 (Canada)

    2016-06-28

    In this work we explore the physical foundations of models that study the variation of the ground state energy with respect to the number of electrons (E vs. N models), in terms of general grand-canonical (GC) ensemble formulations. In particular, we focus on E vs. N models that interpolate the energy between states with integer number of electrons. We show that if the interpolation of the energy corresponds to a GC ensemble, it is not differentiable. Conversely, if the interpolation is smooth, then it cannot be formulated as any GC ensemble. This proves that interpolation of electronic properties between integer electron numbers is inconsistent with any form of ensemble averaging. This emphasizes the role of derivative discontinuities and the critical role of a subsystem’s surroundings in determining its properties.

  17. Localization of atomic ensembles via superfluorescence

    International Nuclear Information System (INIS)

    Macovei, Mihai; Evers, Joerg; Keitel, Christoph H.; Zubairy, M. Suhail

    2007-01-01

    The subwavelength localization of an ensemble of atoms concentrated to a small volume in space is investigated. The localization relies on the interaction of the ensemble with a standing wave laser field. The light scattered in the interaction of the standing wave field and the atom ensemble depends on the position of the ensemble relative to the standing wave nodes. This relation can be described by a fluorescence intensity profile, which depends on the standing wave field parameters and the ensemble properties and which is modified due to collective effects in the ensemble of nearby particles. We demonstrate that the intensity profile can be tailored to suit different localization setups. Finally, we apply these results to two localization schemes. First, we show how to localize an ensemble fixed at a certain position in the standing wave field. Second, we discuss localization of an ensemble passing through the standing wave field

  18. Ensembl 2017

    OpenAIRE

    Aken, Bronwen L.; Achuthan, Premanand; Akanni, Wasiu; Amode, M. Ridwan; Bernsdorff, Friederike; Bhai, Jyothish; Billis, Konstantinos; Carvalho-Silva, Denise; Cummins, Carla; Clapham, Peter; Gil, Laurent; Gir?n, Carlos Garc?a; Gordon, Leo; Hourlier, Thibaut; Hunt, Sarah E.

    2016-01-01

    Ensembl (www.ensembl.org) is a database and genome browser for enabling research on vertebrate genomes. We import, analyse, curate and integrate a diverse collection of large-scale reference data to create a more comprehensive view of genome biology than would be possible from any individual dataset. Our extensive data resources include evidence-based gene and regulatory region annotation, genome variation and gene trees. An accompanying suite of tools, infrastructure and programmatic access ...

  19. Ensemble Sampling

    OpenAIRE

    Lu, Xiuyuan; Van Roy, Benjamin

    2017-01-01

    Thompson sampling has emerged as an effective heuristic for a broad range of online decision problems. In its basic form, the algorithm requires computing and sampling from a posterior distribution over models, which is tractable only for simple special cases. This paper develops ensemble sampling, which aims to approximate Thompson sampling while maintaining tractability even in the face of complex models such as neural networks. Ensemble sampling dramatically expands on the range of applica...

  20. Robustness of the far-field response of nonlocal plasmonic ensembles

    DEFF Research Database (Denmark)

    Tserkezis, Christos; Maack, Johan Rosenkrantz; Liu, Zhaowei

    2016-01-01

    Contrary to classical predictions, the optical response of few-nm plasmonic particles depends on particle size due to effects such as nonlocality and electron spill-out. Ensembles of such nanoparticles are therefore expected to exhibit a nonclassical inhomogeneous spectral broadening due to size...... distribution. For a normal distribution of free-electron nanoparticles, and within the simple nonlocal hydrodynamic Drude model, both the nonlocal blueshift and the plasmon linewidth are shown to be considerably affected by ensemble averaging. Size-variance effects tend however to conceal nonlocality...... to a lesser extent when the homogeneous size-dependent broadening of individual nanoparticles is taken into account, either through a local size-dependent damping model or through the Generalized Nonlocal Optical Response theory. The role of ensemble averaging is further explored in realistic distributions...

  1. EnsembleGASVR: A novel ensemble method for classifying missense single nucleotide polymorphisms

    KAUST Repository

    Rapakoulia, Trisevgeni; Theofilatos, Konstantinos A.; Kleftogiannis, Dimitrios A.; Likothanasis, Spiridon D.; Tsakalidis, Athanasios K.; Mavroudi, Seferina P.

    2014-01-01

    do not support their predictions with confidence scores. Results: To overcome these limitations, a novel ensemble computational methodology is proposed. EnsembleGASVR facilitates a twostep algorithm, which in its first step applies a novel

  2. EnsembleGraph: Interactive Visual Analysis of Spatial-Temporal Behavior for Ensemble Simulation Data

    Energy Technology Data Exchange (ETDEWEB)

    Shu, Qingya; Guo, Hanqi; Che, Limei; Yuan, Xiaoru; Liu, Junfeng; Liang, Jie

    2016-04-19

    We present a novel visualization framework—EnsembleGraph— for analyzing ensemble simulation data, in order to help scientists understand behavior similarities between ensemble members over space and time. A graph-based representation is used to visualize individual spatiotemporal regions with similar behaviors, which are extracted by hierarchical clustering algorithms. A user interface with multiple-linked views is provided, which enables users to explore, locate, and compare regions that have similar behaviors between and then users can investigate and analyze the selected regions in detail. The driving application of this paper is the studies on regional emission influences over tropospheric ozone, which is based on ensemble simulations conducted with different anthropogenic emission absences using the MOZART-4 (model of ozone and related tracers, version 4) model. We demonstrate the effectiveness of our method by visualizing the MOZART-4 ensemble simulation data and evaluating the relative regional emission influences on tropospheric ozone concentrations. Positive feedbacks from domain experts and two case studies prove efficiency of our method.

  3. What sea-ice biogeochemical modellers need from observers

    OpenAIRE

    Steiner, Nadja; Deal, Clara; Lannuzel, Delphine; Lavoie, Diane; Massonnet, François; Miller, Lisa A.; Moreau, Sebastien; Popova, Ekaterina; Stefels, Jacqueline; Tedesco, Letizia

    2016-01-01

    Abstract Numerical models can be a powerful tool helping to understand the role biogeochemical processes play in local and global systems and how this role may be altered in a changing climate. With respect to sea-ice biogeochemical models, our knowledge is severely limited by our poor confidence in numerical model parameterisations representing those processes. Improving model parameterisations requires communication between observers and modellers to guide model development and improve the ...

  4. The spread amongst ENSEMBLES regional scenarios: regional climate models, driving general circulation models and interannual variability

    Energy Technology Data Exchange (ETDEWEB)

    Deque, M.; Somot, S. [Meteo-France, Centre National de Recherches Meteorologiques, CNRS/GAME, Toulouse Cedex 01 (France); Sanchez-Gomez, E. [Cerfacs/CNRS, SUC URA1875, Toulouse Cedex 01 (France); Goodess, C.M. [University of East Anglia, Climatic Research Unit, Norwich (United Kingdom); Jacob, D. [Max Planck Institute for Meteorology, Hamburg (Germany); Lenderink, G. [KNMI, Postbus 201, De Bilt (Netherlands); Christensen, O.B. [Danish Meteorological Institute, Copenhagen Oe (Denmark)

    2012-03-15

    Various combinations of thirteen regional climate models (RCM) and six general circulation models (GCM) were used in FP6-ENSEMBLES. The response to the SRES-A1B greenhouse gas concentration scenario over Europe, calculated as the difference between the 2021-2050 and the 1961-1990 means can be viewed as an expected value about which various uncertainties exist. Uncertainties are measured here by variance explained for temperature and precipitation changes over eight European sub-areas. Three sources of uncertainty can be evaluated from the ENSEMBLES database. Sampling uncertainty is due to the fact that the model climate is estimated as an average over a finite number of years (30) despite a non-negligible interannual variability. Regional model uncertainty is due to the fact that the RCMs use different techniques to discretize the equations and to represent sub-grid effects. Global model uncertainty is due to the fact that the RCMs have been driven by different GCMs. Two methods are presented to fill the many empty cells of the ENSEMBLES RCM x GCM matrix. The first one is based on the same approach as in FP5-PRUDENCE. The second one uses the concept of weather regimes to attempt to separate the contribution of the GCM and the RCM. The variance of the climate response is analyzed with respect to the contribution of the GCM and the RCM. The two filling methods agree that the main contributor to the spread is the choice of the GCM, except for summer precipitation where the choice of the RCM dominates the uncertainty. Of course the implication of the GCM to the spread varies with the region, being maximum in the South-western part of Europe, whereas the continental parts are more sensitive to the choice of the RCM. The third cause of spread is systematically the interannual variability. The total uncertainty about temperature is not large enough to mask the 2021-2050 response which shows a similar pattern to the one obtained for 2071-2100 in PRUDENCE. The uncertainty

  5. The canonical ensemble redefined - 1: Formalism

    International Nuclear Information System (INIS)

    Venkataraman, R.

    1984-12-01

    For studying the thermodynamic properties of systems we propose an ensemble that lies in between the familiar canonical and microcanonical ensembles. We point out the transition from the canonical to microcanonical ensemble and prove from a comparative study that all these ensembles do not yield the same results even in the thermodynamic limit. An investigation of the coupling between two or more systems with these ensembles suggests that the state of thermodynamical equilibrium is a special case of statistical equilibrium. (author)

  6. Geometric integrator for simulations in the canonical ensemble

    Energy Technology Data Exchange (ETDEWEB)

    Tapias, Diego, E-mail: diego.tapias@nucleares.unam.mx [Departamento de Física, Facultad de Ciencias, Universidad Nacional Autónoma de México, Ciudad Universitaria, Ciudad de México 04510 (Mexico); Sanders, David P., E-mail: dpsanders@ciencias.unam.mx [Departamento de Física, Facultad de Ciencias, Universidad Nacional Autónoma de México, Ciudad Universitaria, Ciudad de México 04510 (Mexico); Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, Massachusetts 02139 (United States); Bravetti, Alessandro, E-mail: alessandro.bravetti@iimas.unam.mx [Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas, Universidad Nacional Autónoma de México, Ciudad Universitaria, Ciudad de México 04510 (Mexico)

    2016-08-28

    We introduce a geometric integrator for molecular dynamics simulations of physical systems in the canonical ensemble that preserves the invariant distribution in equations arising from the density dynamics algorithm, with any possible type of thermostat. Our integrator thus constitutes a unified framework that allows the study and comparison of different thermostats and of their influence on the equilibrium and non-equilibrium (thermo-)dynamic properties of a system. To show the validity and the generality of the integrator, we implement it with a second-order, time-reversible method and apply it to the simulation of a Lennard-Jones system with three different thermostats, obtaining good conservation of the geometrical properties and recovering the expected thermodynamic results. Moreover, to show the advantage of our geometric integrator over a non-geometric one, we compare the results with those obtained by using the non-geometric Gear integrator, which is frequently used to perform simulations in the canonical ensemble. The non-geometric integrator induces a drift in the invariant quantity, while our integrator has no such drift, thus ensuring that the system is effectively sampling the correct ensemble.

  7. Geometric integrator for simulations in the canonical ensemble

    International Nuclear Information System (INIS)

    Tapias, Diego; Sanders, David P.; Bravetti, Alessandro

    2016-01-01

    We introduce a geometric integrator for molecular dynamics simulations of physical systems in the canonical ensemble that preserves the invariant distribution in equations arising from the density dynamics algorithm, with any possible type of thermostat. Our integrator thus constitutes a unified framework that allows the study and comparison of different thermostats and of their influence on the equilibrium and non-equilibrium (thermo-)dynamic properties of a system. To show the validity and the generality of the integrator, we implement it with a second-order, time-reversible method and apply it to the simulation of a Lennard-Jones system with three different thermostats, obtaining good conservation of the geometrical properties and recovering the expected thermodynamic results. Moreover, to show the advantage of our geometric integrator over a non-geometric one, we compare the results with those obtained by using the non-geometric Gear integrator, which is frequently used to perform simulations in the canonical ensemble. The non-geometric integrator induces a drift in the invariant quantity, while our integrator has no such drift, thus ensuring that the system is effectively sampling the correct ensemble.

  8. Ensemble methods for handwritten digit recognition

    DEFF Research Database (Denmark)

    Hansen, Lars Kai; Liisberg, Christian; Salamon, P.

    1992-01-01

    Neural network ensembles are applied to handwritten digit recognition. The individual networks of the ensemble are combinations of sparse look-up tables (LUTs) with random receptive fields. It is shown that the consensus of a group of networks outperforms the best individual of the ensemble....... It is further shown that it is possible to estimate the ensemble performance as well as the learning curve on a medium-size database. In addition the authors present preliminary analysis of experiments on a large database and show that state-of-the-art performance can be obtained using the ensemble approach...... by optimizing the receptive fields. It is concluded that it is possible to improve performance significantly by introducing moderate-size ensembles; in particular, a 20-25% improvement has been found. The ensemble random LUTs, when trained on a medium-size database, reach a performance (without rejects) of 94...

  9. Eigenfunction statistics of Wishart Brownian ensembles

    International Nuclear Information System (INIS)

    Shukla, Pragya

    2017-01-01

    We theoretically analyze the eigenfunction fluctuation measures for a Hermitian ensemble which appears as an intermediate state of the perturbation of a stationary ensemble by another stationary ensemble of Wishart (Laguerre) type. Similar to the perturbation by a Gaussian stationary ensemble, the measures undergo a diffusive dynamics in terms of the perturbation parameter but the energy-dependence of the fluctuations is different in the two cases. This may have important consequences for the eigenfunction dynamics as well as phase transition studies in many areas of complexity where Brownian ensembles appear. (paper)

  10. Measuring social interaction in music ensembles.

    Science.gov (United States)

    Volpe, Gualtiero; D'Ausilio, Alessandro; Badino, Leonardo; Camurri, Antonio; Fadiga, Luciano

    2016-05-05

    Music ensembles are an ideal test-bed for quantitative analysis of social interaction. Music is an inherently social activity, and music ensembles offer a broad variety of scenarios which are particularly suitable for investigation. Small ensembles, such as string quartets, are deemed a significant example of self-managed teams, where all musicians contribute equally to a task. In bigger ensembles, such as orchestras, the relationship between a leader (the conductor) and a group of followers (the musicians) clearly emerges. This paper presents an overview of recent research on social interaction in music ensembles with a particular focus on (i) studies from cognitive neuroscience; and (ii) studies adopting a computational approach for carrying out automatic quantitative analysis of ensemble music performances. © 2016 The Author(s).

  11. Ecotoxicological, ecophysiological, and biogeochemical fundamentals of risk assessment

    International Nuclear Information System (INIS)

    Bashkin, V.N.; Kozlov, M.Ya.; Evstafjeva, E.V.

    1993-01-01

    Risk assessment (RA) influenced by different factors in radionuclide polluted regions is carried out by determining the biogeochemical structure of a region. Consequently, ecological-biogeochemical regionalization, ecotoxicological and ecophysiological monitoring of human population health are the important approach to RA. These criteria should conjugate with LCA of various industrial and agricultural products. Given fundamentals and approaches are needed for areas where traditional pollutants (heavy metals, pesticides, fertilizers, POPs etc) are enforced sharply by radioactive pollution. For RA of these complex pollutants, the methods of human adaptability to a polluted environment have been carried out. These techniques include biogeochemical, ecotoxicological, and ecophysiological analyses of risk factors as well as quantitative analysis of uncertainties using expert-modeling systems. Furthermore, the modern statistical methods are used for quantitative assessment of human adaptability to radioactive and nonradioactive pollutants. The results obtained in Chernobyl regions show the acceptability of these methods for risk assessment

  12. Biogeochemical provinces in the global ocean based on phytoplankton growth limitation

    Science.gov (United States)

    Hashioka, T.; Hirata, T.; Aita, M. N.; Chiba, S.

    2016-02-01

    The biogeochemical province is one of the useful concepts for the comprehensive understanding of regional differences of the marine ecosystem. Various biogeochemical provinces for lower-trophic level ecosystem have been proposed using a similarity-based classification of seasonal variations of chl-a concentration typified by Longhurst 1995 and 2006. Such categorizations well capture the regional differences of seasonality as "total phytoplankton". However, background biogeochemical mechanism to characterize the province boundary is not clear. Namely, the dominant phytoplankton group is different among regions and seasons, and their physiological characteristics are significantly different among groups. Recently some pieces of new biogeochemical information are available. One is an estimation of phytoplankton community structure from satellite observation, and it makes clear the key phytoplankton type in each region. Another is an estimation of limitation factors for phytoplankton growth (e.g., nutrients, temperature, light) in each region from modeling studies. In this study, we propose new biogeochemical provinces as a combination between the dominance of phytoplankton (i.e., diatoms, nano-, pico-phytoplankton or coexistence of two/three types) and their growth limitation factors (particularly we focused on nutrient limitation; N, P, Si or Fe). In this combination, we classified the global ocean into 23 biogeochemical provinces. The result suggests that even if the same type of phytoplankton dominates, the background mechanism could be different among regions. On the contrary, even if the regions geographically separate, the background mechanism could be similar among regions. This is important to understand that region/boundary does respond to environmental change. This biogeochemical province is useful for identification of key areas for future observation.

  13. Joys of Community Ensemble Playing: The Case of the Happy Roll Elastic Ensemble in Taiwan

    Science.gov (United States)

    Hsieh, Yuan-Mei; Kao, Kai-Chi

    2012-01-01

    The Happy Roll Elastic Ensemble (HREE) is a community music ensemble supported by Tainan Culture Centre in Taiwan. With enjoyment and friendship as its primary goals, it aims to facilitate the joys of ensemble playing and the spirit of social networking. This article highlights the key aspects of HREE's development in its first two years…

  14. Ensemble Data Mining Methods

    Science.gov (United States)

    Oza, Nikunj C.

    2004-01-01

    Ensemble Data Mining Methods, also known as Committee Methods or Model Combiners, are machine learning methods that leverage the power of multiple models to achieve better prediction accuracy than any of the individual models could on their own. The basic goal when designing an ensemble is the same as when establishing a committee of people: each member of the committee should be as competent as possible, but the members should be complementary to one another. If the members are not complementary, Le., if they always agree, then the committee is unnecessary---any one member is sufficient. If the members are complementary, then when one or a few members make an error, the probability is high that the remaining members can correct this error. Research in ensemble methods has largely revolved around designing ensembles consisting of competent yet complementary models.

  15. Multi-objective optimization for generating a weighted multi-model ensemble

    Science.gov (United States)

    Lee, H.

    2017-12-01

    Many studies have demonstrated that multi-model ensembles generally show better skill than each ensemble member. When generating weighted multi-model ensembles, the first step is measuring the performance of individual model simulations using observations. There is a consensus on the assignment of weighting factors based on a single evaluation metric. When considering only one evaluation metric, the weighting factor for each model is proportional to a performance score or inversely proportional to an error for the model. While this conventional approach can provide appropriate combinations of multiple models, the approach confronts a big challenge when there are multiple metrics under consideration. When considering multiple evaluation metrics, it is obvious that a simple averaging of multiple performance scores or model ranks does not address the trade-off problem between conflicting metrics. So far, there seems to be no best method to generate weighted multi-model ensembles based on multiple performance metrics. The current study applies the multi-objective optimization, a mathematical process that provides a set of optimal trade-off solutions based on a range of evaluation metrics, to combining multiple performance metrics for the global climate models and their dynamically downscaled regional climate simulations over North America and generating a weighted multi-model ensemble. NASA satellite data and the Regional Climate Model Evaluation System (RCMES) software toolkit are used for assessment of the climate simulations. Overall, the performance of each model differs markedly with strong seasonal dependence. Because of the considerable variability across the climate simulations, it is important to evaluate models systematically and make future projections by assigning optimized weighting factors to the models with relatively good performance. Our results indicate that the optimally weighted multi-model ensemble always shows better performance than an arithmetic

  16. Gridded Calibration of Ensemble Wind Vector Forecasts Using Ensemble Model Output Statistics

    Science.gov (United States)

    Lazarus, S. M.; Holman, B. P.; Splitt, M. E.

    2017-12-01

    A computationally efficient method is developed that performs gridded post processing of ensemble wind vector forecasts. An expansive set of idealized WRF model simulations are generated to provide physically consistent high resolution winds over a coastal domain characterized by an intricate land / water mask. Ensemble model output statistics (EMOS) is used to calibrate the ensemble wind vector forecasts at observation locations. The local EMOS predictive parameters (mean and variance) are then spread throughout the grid utilizing flow-dependent statistical relationships extracted from the downscaled WRF winds. Using data withdrawal and 28 east central Florida stations, the method is applied to one year of 24 h wind forecasts from the Global Ensemble Forecast System (GEFS). Compared to the raw GEFS, the approach improves both the deterministic and probabilistic forecast skill. Analysis of multivariate rank histograms indicate the post processed forecasts are calibrated. Two downscaling case studies are presented, a quiescent easterly flow event and a frontal passage. Strengths and weaknesses of the approach are presented and discussed.

  17. Generalized thermalization for integrable system under quantum quench.

    Science.gov (United States)

    Muralidharan, Sushruth; Lochan, Kinjalk; Shankaranarayanan, S

    2018-01-01

    We investigate equilibration and generalized thermalization of the quantum Harmonic chain under local quantum quench. The quench action we consider is connecting two disjoint harmonic chains of different sizes and the system jumps between two integrable settings. We verify the validity of the generalized Gibbs ensemble description for this infinite-dimensional Hilbert space system and also identify equilibration between the subsystems as in classical systems. Using Bogoliubov transformations, we show that the eigenstates of the system prior to the quench evolve toward the Gibbs Generalized Ensemble description. Eigenstates that are more delocalized (in the sense of inverse participation ratio) prior to the quench, tend to equilibrate more rapidly. Further, through the phase space properties of a generalized Gibbs ensemble and the strength of stimulated emission, we identify the necessary criterion on the initial states for such relaxation at late times and also find out the states that would potentially not be described by the generalized Gibbs ensemble description.

  18. Evaluation of the transport matrix method for simulation of ocean biogeochemical tracers

    Science.gov (United States)

    Kvale, Karin F.; Khatiwala, Samar; Dietze, Heiner; Kriest, Iris; Oschlies, Andreas

    2017-06-01

    Conventional integration of Earth system and ocean models can accrue considerable computational expenses, particularly for marine biogeochemical applications. Offline numerical schemes in which only the biogeochemical tracers are time stepped and transported using a pre-computed circulation field can substantially reduce the burden and are thus an attractive alternative. One such scheme is the transport matrix method (TMM), which represents tracer transport as a sequence of sparse matrix-vector products that can be performed efficiently on distributed-memory computers. While the TMM has been used for a variety of geochemical and biogeochemical studies, to date the resulting solutions have not been comprehensively assessed against their online counterparts. Here, we present a detailed comparison of the two. It is based on simulations of the state-of-the-art biogeochemical sub-model embedded within the widely used coarse-resolution University of Victoria Earth System Climate Model (UVic ESCM). The default, non-linear advection scheme was first replaced with a linear, third-order upwind-biased advection scheme to satisfy the linearity requirement of the TMM. Transport matrices were extracted from an equilibrium run of the physical model and subsequently used to integrate the biogeochemical model offline to equilibrium. The identical biogeochemical model was also run online. Our simulations show that offline integration introduces some bias to biogeochemical quantities through the omission of the polar filtering used in UVic ESCM and in the offline application of time-dependent forcing fields, with high latitudes showing the largest differences with respect to the online model. Differences in other regions and in the seasonality of nutrients and phytoplankton distributions are found to be relatively minor, giving confidence that the TMM is a reliable tool for offline integration of complex biogeochemical models. Moreover, while UVic ESCM is a serial code, the TMM can

  19. Ensembl 2002: accommodating comparative genomics.

    Science.gov (United States)

    Clamp, M; Andrews, D; Barker, D; Bevan, P; Cameron, G; Chen, Y; Clark, L; Cox, T; Cuff, J; Curwen, V; Down, T; Durbin, R; Eyras, E; Gilbert, J; Hammond, M; Hubbard, T; Kasprzyk, A; Keefe, D; Lehvaslaiho, H; Iyer, V; Melsopp, C; Mongin, E; Pettett, R; Potter, S; Rust, A; Schmidt, E; Searle, S; Slater, G; Smith, J; Spooner, W; Stabenau, A; Stalker, J; Stupka, E; Ureta-Vidal, A; Vastrik, I; Birney, E

    2003-01-01

    The Ensembl (http://www.ensembl.org/) database project provides a bioinformatics framework to organise biology around the sequences of large genomes. It is a comprehensive source of stable automatic annotation of human, mouse and other genome sequences, available as either an interactive web site or as flat files. Ensembl also integrates manually annotated gene structures from external sources where available. As well as being one of the leading sources of genome annotation, Ensembl is an open source software engineering project to develop a portable system able to handle very large genomes and associated requirements. These range from sequence analysis to data storage and visualisation and installations exist around the world in both companies and at academic sites. With both human and mouse genome sequences available and more vertebrate sequences to follow, many of the recent developments in Ensembl have focusing on developing automatic comparative genome analysis and visualisation.

  20. Products of random matrices from fixed trace and induced Ginibre ensembles

    Science.gov (United States)

    Akemann, Gernot; Cikovic, Milan

    2018-05-01

    We investigate the microcanonical version of the complex induced Ginibre ensemble, by introducing a fixed trace constraint for its second moment. Like for the canonical Ginibre ensemble, its complex eigenvalues can be interpreted as a two-dimensional Coulomb gas, which are now subject to a constraint and a modified, collective confining potential. Despite the lack of determinantal structure in this fixed trace ensemble, we compute all its density correlation functions at finite matrix size and compare to a fixed trace ensemble of normal matrices, representing a different Coulomb gas. Our main tool of investigation is the Laplace transform, that maps back the fixed trace to the induced Ginibre ensemble. Products of random matrices have been used to study the Lyapunov and stability exponents for chaotic dynamical systems, where the latter are based on the complex eigenvalues of the product matrix. Because little is known about the universality of the eigenvalue distribution of such product matrices, we then study the product of m induced Ginibre matrices with a fixed trace constraint—which are clearly non-Gaussian—and M  ‑  m such Ginibre matrices without constraint. Using an m-fold inverse Laplace transform, we obtain a concise result for the spectral density of such a mixed product matrix at finite matrix size, for arbitrary fixed m and M. Very recently local and global universality was proven by the authors and their coworker for a more general, single elliptic fixed trace ensemble in the bulk of the spectrum. Here, we argue that the spectral density of mixed products is in the same universality class as the product of M independent induced Ginibre ensembles.

  1. Data Pre-Analysis and Ensemble of Various Artificial Neural Networks for Monthly Streamflow Forecasting

    Directory of Open Access Journals (Sweden)

    Jianzhong Zhou

    2018-05-01

    Full Text Available This paper introduces three artificial neural network (ANN architectures for monthly streamflow forecasting: a radial basis function network, an extreme learning machine, and the Elman network. Three ensemble techniques, a simple average ensemble, a weighted average ensemble, and an ANN-based ensemble, were used to combine the outputs of the individual ANN models. The objective was to highlight the performance of the general regression neural network-based ensemble technique (GNE through an improvement of monthly streamflow forecasting accuracy. Before the construction of an ANN model, data preanalysis techniques, such as empirical wavelet transform (EWT, were exploited to eliminate the oscillations of the streamflow series. Additionally, a theory of chaos phase space reconstruction was used to select the most relevant and important input variables for forecasting. The proposed GNE ensemble model has been applied for the mean monthly streamflow observation data from the Wudongde hydrological station in the Jinsha River Basin, China. Comparisons and analysis of this study have demonstrated that the denoised streamflow time series was less disordered and unsystematic than was suggested by the original time series according to chaos theory. Thus, EWT can be adopted as an effective data preanalysis technique for the prediction of monthly streamflow. Concurrently, the GNE performed better when compared with other ensemble techniques.

  2. The semantic similarity ensemble

    Directory of Open Access Journals (Sweden)

    Andrea Ballatore

    2013-12-01

    Full Text Available Computational measures of semantic similarity between geographic terms provide valuable support across geographic information retrieval, data mining, and information integration. To date, a wide variety of approaches to geo-semantic similarity have been devised. A judgment of similarity is not intrinsically right or wrong, but obtains a certain degree of cognitive plausibility, depending on how closely it mimics human behavior. Thus selecting the most appropriate measure for a specific task is a significant challenge. To address this issue, we make an analogy between computational similarity measures and soliciting domain expert opinions, which incorporate a subjective set of beliefs, perceptions, hypotheses, and epistemic biases. Following this analogy, we define the semantic similarity ensemble (SSE as a composition of different similarity measures, acting as a panel of experts having to reach a decision on the semantic similarity of a set of geographic terms. The approach is evaluated in comparison to human judgments, and results indicate that an SSE performs better than the average of its parts. Although the best member tends to outperform the ensemble, all ensembles outperform the average performance of each ensemble's member. Hence, in contexts where the best measure is unknown, the ensemble provides a more cognitively plausible approach.

  3. Quantum ensembles of quantum classifiers.

    Science.gov (United States)

    Schuld, Maria; Petruccione, Francesco

    2018-02-09

    Quantum machine learning witnesses an increasing amount of quantum algorithms for data-driven decision making, a problem with potential applications ranging from automated image recognition to medical diagnosis. Many of those algorithms are implementations of quantum classifiers, or models for the classification of data inputs with a quantum computer. Following the success of collective decision making with ensembles in classical machine learning, this paper introduces the concept of quantum ensembles of quantum classifiers. Creating the ensemble corresponds to a state preparation routine, after which the quantum classifiers are evaluated in parallel and their combined decision is accessed by a single-qubit measurement. This framework naturally allows for exponentially large ensembles in which - similar to Bayesian learning - the individual classifiers do not have to be trained. As an example, we analyse an exponentially large quantum ensemble in which each classifier is weighed according to its performance in classifying the training data, leading to new results for quantum as well as classical machine learning.

  4. Biogeochemical cycling in the Taiwan Strait

    Digital Repository Service at National Institute of Oceanography (India)

    Naik, H.; Chen, C-T.A.

    Based on repeat observations made during 2001-2003 along two transects in the Taiwan Strait this study aims at understanding factors controlling primary productivity with an emphasis on biogeochemical cycling of nitrogen, the major bio...

  5. Cyclic biogeochemical processes and nitrogen fate beneath a subtropical stormwater infiltration basin.

    Science.gov (United States)

    O'Reilly, Andrew M; Chang, Ni-Bin; Wanielista, Martin P

    2012-05-15

    A stormwater infiltration basin in north-central Florida, USA, was monitored from 2007 through 2008 to identify subsurface biogeochemical processes, with emphasis on N cycling, under the highly variable hydrologic conditions common in humid, subtropical climates. Cyclic variations in biogeochemical processes generally coincided with wet and dry hydrologic conditions. Oxidizing conditions in the subsurface persisted for about one month or less at the beginning of wet periods with dissolved O(2) and NO(3)(-) showing similar temporal patterns. Reducing conditions in the subsurface evolved during prolonged flooding of the basin. At about the same time O(2) and NO(3)(-) reduction concluded, Mn, Fe and SO(4)(2-) reduction began, with the onset of methanogenesis one month later. Reducing conditions persisted up to six months, continuing into subsequent dry periods until the next major oxidizing infiltration event. Evidence of denitrification in shallow groundwater at the site is supported by median NO(3)(-)-N less than 0.016 mg L(-1), excess N(2) up to 3 mg L(-1) progressively enriched in δ(15)N during prolonged basin flooding, and isotopically heavy δ(15)N and δ(18)O of NO(3)(-) (up to 25‰ and 15‰, respectively). Isotopic enrichment of newly infiltrated stormwater suggests denitrification was partially completed within two days. Soil and water chemistry data suggest that a biogeochemically active zone exists in the upper 1.4m of soil, where organic carbon was the likely electron donor supplied by organic matter in soil solids or dissolved in infiltrating stormwater. The cyclic nature of reducing conditions effectively controlled the N cycle, switching N fate beneath the basin from NO(3)(-) leaching to reduction in the shallow saturated zone. Results can inform design of functionalized soil amendments that could replace the native soil in a stormwater infiltration basin and mitigate potential NO(3)(-) leaching to groundwater by replicating the biogeochemical

  6. Cyclic biogeochemical processes and nitrogen fate beneath a subtropical stormwater infiltration basin

    Science.gov (United States)

    O'Reilly, Andrew M.; Chang, Ni-Bin; Wanielista, Martin P.

    2012-01-01

    A stormwater infiltration basin in north–central Florida, USA, was monitored from 2007 through 2008 to identify subsurface biogeochemical processes, with emphasis on N cycling, under the highly variable hydrologic conditions common in humid, subtropical climates. Cyclic variations in biogeochemical processes generally coincided with wet and dry hydrologic conditions. Oxidizing conditions in the subsurface persisted for about one month or less at the beginning of wet periods with dissolved O2 and NO3- showing similar temporal patterns. Reducing conditions in the subsurface evolved during prolonged flooding of the basin. At about the same time O2 and NO3- reduction concluded, Mn, Fe and SO42- reduction began, with the onset of methanogenesis one month later. Reducing conditions persisted up to six months, continuing into subsequent dry periods until the next major oxidizing infiltration event. Evidence of denitrification in shallow groundwater at the site is supported by median NO3-–N less than 0.016 mg L-1, excess N2 up to 3 mg L-1 progressively enriched in δ15N during prolonged basin flooding, and isotopically heavy δ15N and δ18O of NO3- (up to 25‰ and 15‰, respectively). Isotopic enrichment of newly infiltrated stormwater suggests denitrification was partially completed within two days. Soil and water chemistry data suggest that a biogeochemically active zone exists in the upper 1.4 m of soil, where organic carbon was the likely electron donor supplied by organic matter in soil solids or dissolved in infiltrating stormwater. The cyclic nature of reducing conditions effectively controlled the N cycle, switching N fate beneath the basin from NO3- leaching to reduction in the shallow saturated zone. Results can inform design of functionalized soil amendments that could replace the native soil in a stormwater infiltration basin and mitigate potential NO3- leaching to groundwater by replicating the biogeochemical conditions under the observed basin.

  7. Musical ensembles in Ancient Mesapotamia

    NARCIS (Netherlands)

    Krispijn, T.J.H.; Dumbrill, R.; Finkel, I.

    2010-01-01

    Identification of musical instruments from ancient Mesopotamia by comparing musical ensembles attested in Sumerian and Akkadian texts with depicted ensembles. Lexicographical contributions to the Sumerian and Akkadian lexicon.

  8. PSO-Ensemble Demo Application

    DEFF Research Database (Denmark)

    2004-01-01

    Within the framework of the PSO-Ensemble project (FU2101) a demo application has been created. The application use ECMWF ensemble forecasts. Two instances of the application are running; one for Nysted Offshore and one for the total production (except Horns Rev) in the Eltra area. The output...

  9. Multilevel ensemble Kalman filter

    KAUST Repository

    Chernov, Alexey; Hoel, Haakon; Law, Kody; Nobile, Fabio; Tempone, Raul

    2016-01-01

    This work embeds a multilevel Monte Carlo (MLMC) sampling strategy into the Monte Carlo step of the ensemble Kalman filter (EnKF). In terms of computational cost vs. approximation error the asymptotic performance of the multilevel ensemble Kalman filter (MLEnKF) is superior to the EnKF s.

  10. Multilevel ensemble Kalman filter

    KAUST Repository

    Chernov, Alexey

    2016-01-06

    This work embeds a multilevel Monte Carlo (MLMC) sampling strategy into the Monte Carlo step of the ensemble Kalman filter (EnKF). In terms of computational cost vs. approximation error the asymptotic performance of the multilevel ensemble Kalman filter (MLEnKF) is superior to the EnKF s.

  11. Ensemble Kinetic Modeling of Metabolic Networks from Dynamic Metabolic Profiles

    Directory of Open Access Journals (Sweden)

    Gengjie Jia

    2012-11-01

    Full Text Available Kinetic modeling of metabolic pathways has important applications in metabolic engineering, but significant challenges still remain. The difficulties faced vary from finding best-fit parameters in a highly multidimensional search space to incomplete parameter identifiability. To meet some of these challenges, an ensemble modeling method is developed for characterizing a subset of kinetic parameters that give statistically equivalent goodness-of-fit to time series concentration data. The method is based on the incremental identification approach, where the parameter estimation is done in a step-wise manner. Numerical efficacy is achieved by reducing the dimensionality of parameter space and using efficient random parameter exploration algorithms. The shift toward using model ensembles, instead of the traditional “best-fit” models, is necessary to directly account for model uncertainty during the application of such models. The performance of the ensemble modeling approach has been demonstrated in the modeling of a generic branched pathway and the trehalose pathway in Saccharomyces cerevisiae using generalized mass action (GMA kinetics.

  12. Reservoir History Matching Using Ensemble Kalman Filters with Anamorphosis Transforms

    KAUST Repository

    Aman, Beshir M.

    2012-12-01

    This work aims to enhance the Ensemble Kalman Filter performance by transforming the non-Gaussian state variables into Gaussian variables to be a step closer to optimality. This is done by using univariate and multivariate Box-Cox transformation. Some History matching methods such as Kalman filter, particle filter and the ensemble Kalman filter are reviewed and applied to a test case in the reservoir application. The key idea is to apply the transformation before the update step and then transform back after applying the Kalman correction. In general, the results of the multivariate method was promising, despite the fact it over-estimated some variables.

  13. Multi-Model Ensemble Wake Vortex Prediction

    Science.gov (United States)

    Koerner, Stephan; Holzaepfel, Frank; Ahmad, Nash'at N.

    2015-01-01

    Several multi-model ensemble methods are investigated for predicting wake vortex transport and decay. This study is a joint effort between National Aeronautics and Space Administration and Deutsches Zentrum fuer Luft- und Raumfahrt to develop a multi-model ensemble capability using their wake models. An overview of different multi-model ensemble methods and their feasibility for wake applications is presented. The methods include Reliability Ensemble Averaging, Bayesian Model Averaging, and Monte Carlo Simulations. The methodologies are evaluated using data from wake vortex field experiments.

  14. Ensemble-based Kalman Filters in Strongly Nonlinear Dynamics

    Institute of Scientific and Technical Information of China (English)

    Zhaoxia PU; Joshua HACKER

    2009-01-01

    This study examines the effectiveness of ensemble Kalman filters in data assimilation with the strongly nonlinear dynamics of the Lorenz-63 model, and in particular their use in predicting the regime transition that occurs when the model jumps from one basin of attraction to the other. Four configurations of the ensemble-based Kalman filtering data assimilation techniques, including the ensemble Kalman filter, ensemble adjustment Kalman filter, ensemble square root filter and ensemble transform Kalman filter, are evaluated with their ability in predicting the regime transition (also called phase transition) and also are compared in terms of their sensitivity to both observational and sampling errors. The sensitivity of each ensemble-based filter to the size of the ensemble is also examined.

  15. Monthly ENSO Forecast Skill and Lagged Ensemble Size

    Science.gov (United States)

    Trenary, L.; DelSole, T.; Tippett, M. K.; Pegion, K.

    2018-04-01

    The mean square error (MSE) of a lagged ensemble of monthly forecasts of the Niño 3.4 index from the Climate Forecast System (CFSv2) is examined with respect to ensemble size and configuration. Although the real-time forecast is initialized 4 times per day, it is possible to infer the MSE for arbitrary initialization frequency and for burst ensembles by fitting error covariances to a parametric model and then extrapolating to arbitrary ensemble size and initialization frequency. Applying this method to real-time forecasts, we find that the MSE consistently reaches a minimum for a lagged ensemble size between one and eight days, when four initializations per day are included. This ensemble size is consistent with the 8-10 day lagged ensemble configuration used operationally. Interestingly, the skill of both ensemble configurations is close to the estimated skill of the infinite ensemble. The skill of the weighted, lagged, and burst ensembles are found to be comparable. Certain unphysical features of the estimated error growth were tracked down to problems with the climatology and data discontinuities.

  16. Time-optimal path planning in uncertain flow fields using ensemble method

    KAUST Repository

    Wang, Tong; Le Maitre, Olivier; Hoteit, Ibrahim; Knio, Omar

    2016-01-01

    the performance of sampling strategy, and develop insight into extensions dealing with regional or general circulation models. In particular, the ensemble method enables us to perform a statistical analysis of travel times, and consequently develop a path planning

  17. A Simple Approach to Account for Climate Model Interdependence in Multi-Model Ensembles

    Science.gov (United States)

    Herger, N.; Abramowitz, G.; Angelil, O. M.; Knutti, R.; Sanderson, B.

    2016-12-01

    Multi-model ensembles are an indispensable tool for future climate projection and its uncertainty quantification. Ensembles containing multiple climate models generally have increased skill, consistency and reliability. Due to the lack of agreed-on alternatives, most scientists use the equally-weighted multi-model mean as they subscribe to model democracy ("one model, one vote").Different research groups are known to share sections of code, parameterizations in their model, literature, or even whole model components. Therefore, individual model runs do not represent truly independent estimates. Ignoring this dependence structure might lead to a false model consensus, wrong estimation of uncertainty and effective number of independent models.Here, we present a way to partially address this problem by selecting a subset of CMIP5 model runs so that its climatological mean minimizes the RMSE compared to a given observation product. Due to the cancelling out of errors, regional biases in the ensemble mean are reduced significantly.Using a model-as-truth experiment we demonstrate that those regional biases persist into the future and we are not fitting noise, thus providing improved observationally-constrained projections of the 21st century. The optimally selected ensemble shows significantly higher global mean surface temperature projections than the original ensemble, where all the model runs are considered. Moreover, the spread is decreased well beyond that expected from the decreased ensemble size.Several previous studies have recommended an ensemble selection approach based on performance ranking of the model runs. Here, we show that this approach can perform even worse than randomly selecting ensemble members and can thus be harmful. We suggest that accounting for interdependence in the ensemble selection process is a necessary step for robust projections for use in impact assessments, adaptation and mitigation of climate change.

  18. Preface to: Indian Ocean biogeochemical processes and ecological variability

    Digital Repository Service at National Institute of Oceanography (India)

    Hood, R.R.; Naqvi, S.W.A.; Wiggert, J.D.

    monsoonal in fluence. The biogeochemical and ecological impacts of this complex physical forcing are not yet fully understood. The Indian Ocean is truly one of the last great frontiers of ocea- nographic research. In addition, it appears... to be particularly vulnerable to climate change and anthropogenic impacts, yet it has been more than a decade since the last coordinated international study of biogeochemical and ecological proc esses was undertaken in this region. To obtain a better un...

  19. Prediction of Coal Face Gas Concentration by Multi-Scale Selective Ensemble Hybrid Modeling

    Directory of Open Access Journals (Sweden)

    WU Xiang

    2014-06-01

    Full Text Available A selective ensemble hybrid modeling prediction method based on wavelet transformation is proposed to improve the fitting and generalization capability of the existing prediction models of the coal face gas concentration, which has a strong stochastic volatility. Mallat algorithm was employed for the multi-scale decomposition and single-scale reconstruction of the gas concentration time series. Then, it predicted every subsequence by sparsely weighted multi unstable ELM(extreme learning machine predictor within method SERELM(sparse ensemble regressors of ELM. At last, it superimposed the predicted values of these models to obtain the predicted values of the original sequence. The proposed method takes advantage of characteristics of multi scale analysis of wavelet transformation, accuracy and fast characteristics of ELM prediction and the generalization ability of L1 regularized selective ensemble learning method. The results show that the forecast accuracy has large increase by using the proposed method. The average relative error is 0.65%, the maximum relative error is 4.16% and the probability of relative error less than 1% reaches 0.785.

  20. New technique for ensemble dressing combining Multimodel SuperEnsemble and precipitation PDF

    Science.gov (United States)

    Cane, D.; Milelli, M.

    2009-09-01

    The Multimodel SuperEnsemble technique (Krishnamurti et al., Science 285, 1548-1550, 1999) is a postprocessing method for the estimation of weather forecast parameters reducing direct model output errors. It differs from other ensemble analysis techniques by the use of an adequate weighting of the input forecast models to obtain a combined estimation of meteorological parameters. Weights are calculated by least-square minimization of the difference between the model and the observed field during a so-called training period. Although it can be applied successfully on the continuous parameters like temperature, humidity, wind speed and mean sea level pressure (Cane and Milelli, Meteorologische Zeitschrift, 15, 2, 2006), the Multimodel SuperEnsemble gives good results also when applied on the precipitation, a parameter quite difficult to handle with standard post-processing methods. Here we present our methodology for the Multimodel precipitation forecasts applied on a wide spectrum of results over Piemonte very dense non-GTS weather station network. We will focus particularly on an accurate statistical method for bias correction and on the ensemble dressing in agreement with the observed precipitation forecast-conditioned PDF. Acknowledgement: this work is supported by the Italian Civil Defence Department.

  1. An iterative stochastic ensemble method for parameter estimation of subsurface flow models

    International Nuclear Information System (INIS)

    Elsheikh, Ahmed H.; Wheeler, Mary F.; Hoteit, Ibrahim

    2013-01-01

    Parameter estimation for subsurface flow models is an essential step for maximizing the value of numerical simulations for future prediction and the development of effective control strategies. We propose the iterative stochastic ensemble method (ISEM) as a general method for parameter estimation based on stochastic estimation of gradients using an ensemble of directional derivatives. ISEM eliminates the need for adjoint coding and deals with the numerical simulator as a blackbox. The proposed method employs directional derivatives within a Gauss–Newton iteration. The update equation in ISEM resembles the update step in ensemble Kalman filter, however the inverse of the output covariance matrix in ISEM is regularized using standard truncated singular value decomposition or Tikhonov regularization. We also investigate the performance of a set of shrinkage based covariance estimators within ISEM. The proposed method is successfully applied on several nonlinear parameter estimation problems for subsurface flow models. The efficiency of the proposed algorithm is demonstrated by the small size of utilized ensembles and in terms of error convergence rates

  2. An iterative stochastic ensemble method for parameter estimation of subsurface flow models

    KAUST Repository

    Elsheikh, Ahmed H.

    2013-06-01

    Parameter estimation for subsurface flow models is an essential step for maximizing the value of numerical simulations for future prediction and the development of effective control strategies. We propose the iterative stochastic ensemble method (ISEM) as a general method for parameter estimation based on stochastic estimation of gradients using an ensemble of directional derivatives. ISEM eliminates the need for adjoint coding and deals with the numerical simulator as a blackbox. The proposed method employs directional derivatives within a Gauss-Newton iteration. The update equation in ISEM resembles the update step in ensemble Kalman filter, however the inverse of the output covariance matrix in ISEM is regularized using standard truncated singular value decomposition or Tikhonov regularization. We also investigate the performance of a set of shrinkage based covariance estimators within ISEM. The proposed method is successfully applied on several nonlinear parameter estimation problems for subsurface flow models. The efficiency of the proposed algorithm is demonstrated by the small size of utilized ensembles and in terms of error convergence rates. © 2013 Elsevier Inc.

  3. Unsupervised Learning in an Ensemble of Spiking Neural Networks Mediated by ITDP.

    Directory of Open Access Journals (Sweden)

    Yoonsik Shim

    2016-10-01

    Full Text Available We propose a biologically plausible architecture for unsupervised ensemble learning in a population of spiking neural network classifiers. A mixture of experts type organisation is shown to be effective, with the individual classifier outputs combined via a gating network whose operation is driven by input timing dependent plasticity (ITDP. The ITDP gating mechanism is based on recent experimental findings. An abstract, analytically tractable model of the ITDP driven ensemble architecture is derived from a logical model based on the probabilities of neural firing events. A detailed analysis of this model provides insights that allow it to be extended into a full, biologically plausible, computational implementation of the architecture which is demonstrated on a visual classification task. The extended model makes use of a style of spiking network, first introduced as a model of cortical microcircuits, that is capable of Bayesian inference, effectively performing expectation maximization. The unsupervised ensemble learning mechanism, based around such spiking expectation maximization (SEM networks whose combined outputs are mediated by ITDP, is shown to perform the visual classification task well and to generalize to unseen data. The combined ensemble performance is significantly better than that of the individual classifiers, validating the ensemble architecture and learning mechanisms. The properties of the full model are analysed in the light of extensive experiments with the classification task, including an investigation into the influence of different input feature selection schemes and a comparison with a hierarchical STDP based ensemble architecture.

  4. Unsupervised Learning in an Ensemble of Spiking Neural Networks Mediated by ITDP.

    Science.gov (United States)

    Shim, Yoonsik; Philippides, Andrew; Staras, Kevin; Husbands, Phil

    2016-10-01

    We propose a biologically plausible architecture for unsupervised ensemble learning in a population of spiking neural network classifiers. A mixture of experts type organisation is shown to be effective, with the individual classifier outputs combined via a gating network whose operation is driven by input timing dependent plasticity (ITDP). The ITDP gating mechanism is based on recent experimental findings. An abstract, analytically tractable model of the ITDP driven ensemble architecture is derived from a logical model based on the probabilities of neural firing events. A detailed analysis of this model provides insights that allow it to be extended into a full, biologically plausible, computational implementation of the architecture which is demonstrated on a visual classification task. The extended model makes use of a style of spiking network, first introduced as a model of cortical microcircuits, that is capable of Bayesian inference, effectively performing expectation maximization. The unsupervised ensemble learning mechanism, based around such spiking expectation maximization (SEM) networks whose combined outputs are mediated by ITDP, is shown to perform the visual classification task well and to generalize to unseen data. The combined ensemble performance is significantly better than that of the individual classifiers, validating the ensemble architecture and learning mechanisms. The properties of the full model are analysed in the light of extensive experiments with the classification task, including an investigation into the influence of different input feature selection schemes and a comparison with a hierarchical STDP based ensemble architecture.

  5. Modeling Coordination Problems in a Music Ensemble

    DEFF Research Database (Denmark)

    Frimodt-Møller, Søren R.

    2008-01-01

    This paper considers in general terms, how musicians are able to coordinate through rational choices in a situation of (temporary) doubt in an ensemble performance. A fictitious example involving a 5-bar development in an unknown piece of music is analyzed in terms of epistemic logic, more...... to coordinate. Such coordination can be described in terms of Michael Bacharach's theory of variable frames as an aid to solve game theoretic coordination problems....

  6. Ensemble based system for whole-slide prostate cancer probability mapping using color texture features.

    LENUS (Irish Health Repository)

    DiFranco, Matthew D

    2011-01-01

    We present a tile-based approach for producing clinically relevant probability maps of prostatic carcinoma in histological sections from radical prostatectomy. Our methodology incorporates ensemble learning for feature selection and classification on expert-annotated images. Random forest feature selection performed over varying training sets provides a subset of generalized CIEL*a*b* co-occurrence texture features, while sample selection strategies with minimal constraints reduce training data requirements to achieve reliable results. Ensembles of classifiers are built using expert-annotated tiles from training images, and scores for the probability of cancer presence are calculated from the responses of each classifier in the ensemble. Spatial filtering of tile-based texture features prior to classification results in increased heat-map coherence as well as AUC values of 95% using ensembles of either random forests or support vector machines. Our approach is designed for adaptation to different imaging modalities, image features, and histological decision domains.

  7. A Theoretical Analysis of Why Hybrid Ensembles Work

    Directory of Open Access Journals (Sweden)

    Kuo-Wei Hsu

    2017-01-01

    Full Text Available Inspired by the group decision making process, ensembles or combinations of classifiers have been found favorable in a wide variety of application domains. Some researchers propose to use the mixture of two different types of classification algorithms to create a hybrid ensemble. Why does such an ensemble work? The question remains. Following the concept of diversity, which is one of the fundamental elements of the success of ensembles, we conduct a theoretical analysis of why hybrid ensembles work, connecting using different algorithms to accuracy gain. We also conduct experiments on classification performance of hybrid ensembles of classifiers created by decision tree and naïve Bayes classification algorithms, each of which is a top data mining algorithm and often used to create non-hybrid ensembles. Therefore, through this paper, we provide a complement to the theoretical foundation of creating and using hybrid ensembles.

  8. Path planning in uncertain flow fields using ensemble method

    KAUST Repository

    Wang, Tong

    2016-08-20

    An ensemble-based approach is developed to conduct optimal path planning in unsteady ocean currents under uncertainty. We focus our attention on two-dimensional steady and unsteady uncertain flows, and adopt a sampling methodology that is well suited to operational forecasts, where an ensemble of deterministic predictions is used to model and quantify uncertainty. In an operational setting, much about dynamics, topography, and forcing of the ocean environment is uncertain. To address this uncertainty, the flow field is parametrized using a finite number of independent canonical random variables with known densities, and the ensemble is generated by sampling these variables. For each of the resulting realizations of the uncertain current field, we predict the path that minimizes the travel time by solving a boundary value problem (BVP), based on the Pontryagin maximum principle. A family of backward-in-time trajectories starting at the end position is used to generate suitable initial values for the BVP solver. This allows us to examine and analyze the performance of the sampling strategy and to develop insight into extensions dealing with general circulation ocean models. In particular, the ensemble method enables us to perform a statistical analysis of travel times and consequently develop a path planning approach that accounts for these statistics. The proposed methodology is tested for a number of scenarios. We first validate our algorithms by reproducing simple canonical solutions, and then demonstrate our approach in more complex flow fields, including idealized, steady and unsteady double-gyre flows.

  9. Massively Parallel Assimilation of TOGA/TAO and Topex/Poseidon Measurements into a Quasi Isopycnal Ocean General Circulation Model Using an Ensemble Kalman Filter

    Science.gov (United States)

    Keppenne, Christian L.; Rienecker, Michele; Borovikov, Anna Y.; Suarez, Max

    1999-01-01

    A massively parallel ensemble Kalman filter (EnKF)is used to assimilate temperature data from the TOGA/TAO array and altimetry from TOPEX/POSEIDON into a Pacific basin version of the NASA Seasonal to Interannual Prediction Project (NSIPP)ls quasi-isopycnal ocean general circulation model. The EnKF is an approximate Kalman filter in which the error-covariance propagation step is modeled by the integration of multiple instances of a numerical model. An estimate of the true error covariances is then inferred from the distribution of the ensemble of model state vectors. This inplementation of the filter takes advantage of the inherent parallelism in the EnKF algorithm by running all the model instances concurrently. The Kalman filter update step also occurs in parallel by having each processor process the observations that occur in the region of physical space for which it is responsible. The massively parallel data assimilation system is validated by withholding some of the data and then quantifying the extent to which the withheld information can be inferred from the assimilation of the remaining data. The distributions of the forecast and analysis error covariances predicted by the ENKF are also examined.

  10. Modelling of transport and biogeochemical processes in pollution plumes: Vejen landfill, Denmark

    DEFF Research Database (Denmark)

    Brun, A.; Engesgaard, Peter Knudegaard; Christensen, Thomas Højlund

    2002-01-01

    A biogeochemical transport code is used to simulate leachate attenuation. biogeochemical processes. and development of redox zones in a pollution plume downstream of the Vejen landfill in Denmark. Calibration of the degradation parameters resulted in a good agreement with the observed distribution...

  11. New concept of statistical ensembles

    International Nuclear Information System (INIS)

    Gorenstein, M.I.

    2009-01-01

    An extension of the standard concept of the statistical ensembles is suggested. Namely, the statistical ensembles with extensive quantities fluctuating according to an externally given distribution is introduced. Applications in the statistical models of multiple hadron production in high energy physics are discussed.

  12. Quasi-static ensemble variational data assimilation: a theoretical and numerical study with the iterative ensemble Kalman smoother

    Science.gov (United States)

    Fillion, Anthony; Bocquet, Marc; Gratton, Serge

    2018-04-01

    The analysis in nonlinear variational data assimilation is the solution of a non-quadratic minimization. Thus, the analysis efficiency relies on its ability to locate a global minimum of the cost function. If this minimization uses a Gauss-Newton (GN) method, it is critical for the starting point to be in the attraction basin of a global minimum. Otherwise the method may converge to a local extremum, which degrades the analysis. With chaotic models, the number of local extrema often increases with the temporal extent of the data assimilation window, making the former condition harder to satisfy. This is unfortunate because the assimilation performance also increases with this temporal extent. However, a quasi-static (QS) minimization may overcome these local extrema. It accomplishes this by gradually injecting the observations in the cost function. This method was introduced by Pires et al. (1996) in a 4D-Var context. We generalize this approach to four-dimensional strong-constraint nonlinear ensemble variational (EnVar) methods, which are based on both a nonlinear variational analysis and the propagation of dynamical error statistics via an ensemble. This forces one to consider the cost function minimizations in the broader context of cycled data assimilation algorithms. We adapt this QS approach to the iterative ensemble Kalman smoother (IEnKS), an exemplar of nonlinear deterministic four-dimensional EnVar methods. Using low-order models, we quantify the positive impact of the QS approach on the IEnKS, especially for long data assimilation windows. We also examine the computational cost of QS implementations and suggest cheaper algorithms.

  13. The Ensembl genome database project.

    Science.gov (United States)

    Hubbard, T; Barker, D; Birney, E; Cameron, G; Chen, Y; Clark, L; Cox, T; Cuff, J; Curwen, V; Down, T; Durbin, R; Eyras, E; Gilbert, J; Hammond, M; Huminiecki, L; Kasprzyk, A; Lehvaslaiho, H; Lijnzaad, P; Melsopp, C; Mongin, E; Pettett, R; Pocock, M; Potter, S; Rust, A; Schmidt, E; Searle, S; Slater, G; Smith, J; Spooner, W; Stabenau, A; Stalker, J; Stupka, E; Ureta-Vidal, A; Vastrik, I; Clamp, M

    2002-01-01

    The Ensembl (http://www.ensembl.org/) database project provides a bioinformatics framework to organise biology around the sequences of large genomes. It is a comprehensive source of stable automatic annotation of the human genome sequence, with confirmed gene predictions that have been integrated with external data sources, and is available as either an interactive web site or as flat files. It is also an open source software engineering project to develop a portable system able to handle very large genomes and associated requirements from sequence analysis to data storage and visualisation. The Ensembl site is one of the leading sources of human genome sequence annotation and provided much of the analysis for publication by the international human genome project of the draft genome. The Ensembl system is being installed around the world in both companies and academic sites on machines ranging from supercomputers to laptops.

  14. Evaluation of stability of k-means cluster ensembles with respect to random initialization.

    Science.gov (United States)

    Kuncheva, Ludmila I; Vetrov, Dmitry P

    2006-11-01

    Many clustering algorithms, including cluster ensembles, rely on a random component. Stability of the results across different runs is considered to be an asset of the algorithm. The cluster ensembles considered here are based on k-means clusterers. Each clusterer is assigned a random target number of clusters, k and is started from a random initialization. Here, we use 10 artificial and 10 real data sets to study ensemble stability with respect to random k, and random initialization. The data sets were chosen to have a small number of clusters (two to seven) and a moderate number of data points (up to a few hundred). Pairwise stability is defined as the adjusted Rand index between pairs of clusterers in the ensemble, averaged across all pairs. Nonpairwise stability is defined as the entropy of the consensus matrix of the ensemble. An experimental comparison with the stability of the standard k-means algorithm was carried out for k from 2 to 20. The results revealed that ensembles are generally more stable, markedly so for larger k. To establish whether stability can serve as a cluster validity index, we first looked at the relationship between stability and accuracy with respect to the number of clusters, k. We found that such a relationship strongly depends on the data set, varying from almost perfect positive correlation (0.97, for the glass data) to almost perfect negative correlation (-0.93, for the crabs data). We propose a new combined stability index to be the sum of the pairwise individual and ensemble stabilities. This index was found to correlate better with the ensemble accuracy. Following the hypothesis that a point of stability of a clustering algorithm corresponds to a structure found in the data, we used the stability measures to pick the number of clusters. The combined stability index gave best results.

  15. Force Sensor Based Tool Condition Monitoring Using a Heterogeneous Ensemble Learning Model

    Directory of Open Access Journals (Sweden)

    Guofeng Wang

    2014-11-01

    Full Text Available Tool condition monitoring (TCM plays an important role in improving machining efficiency and guaranteeing workpiece quality. In order to realize reliable recognition of the tool condition, a robust classifier needs to be constructed to depict the relationship between tool wear states and sensory information. However, because of the complexity of the machining process and the uncertainty of the tool wear evolution, it is hard for a single classifier to fit all the collected samples without sacrificing generalization ability. In this paper, heterogeneous ensemble learning is proposed to realize tool condition monitoring in which the support vector machine (SVM, hidden Markov model (HMM and radius basis function (RBF are selected as base classifiers and a stacking ensemble strategy is further used to reflect the relationship between the outputs of these base classifiers and tool wear states. Based on the heterogeneous ensemble learning classifier, an online monitoring system is constructed in which the harmonic features are extracted from force signals and a minimal redundancy and maximal relevance (mRMR algorithm is utilized to select the most prominent features. To verify the effectiveness of the proposed method, a titanium alloy milling experiment was carried out and samples with different tool wear states were collected to build the proposed heterogeneous ensemble learning classifier. Moreover, the homogeneous ensemble learning model and majority voting strategy are also adopted to make a comparison. The analysis and comparison results show that the proposed heterogeneous ensemble learning classifier performs better in both classification accuracy and stability.

  16. Social behaviour shapes hypothalamic neural ensemble representations of conspecific sex

    Science.gov (United States)

    Remedios, Ryan; Kennedy, Ann; Zelikowsky, Moriel; Grewe, Benjamin F.; Schnitzer, Mark J.; Anderson, David J.

    2017-10-01

    All animals possess a repertoire of innate (or instinctive) behaviours, which can be performed without training. Whether such behaviours are mediated by anatomically distinct and/or genetically specified neural pathways remains unknown. Here we report that neural representations within the mouse hypothalamus, that underlie innate social behaviours, are shaped by social experience. Oestrogen receptor 1-expressing (Esr1+) neurons in the ventrolateral subdivision of the ventromedial hypothalamus (VMHvl) control mating and fighting in rodents. We used microendoscopy to image Esr1+ neuronal activity in the VMHvl of male mice engaged in these social behaviours. In sexually and socially experienced adult males, divergent and characteristic neural ensembles represented male versus female conspecifics. However, in inexperienced adult males, male and female intruders activated overlapping neuronal populations. Sex-specific neuronal ensembles gradually separated as the mice acquired social and sexual experience. In mice permitted to investigate but not to mount or attack conspecifics, ensemble divergence did not occur. However, 30 minutes of sexual experience with a female was sufficient to promote the separation of male and female ensembles and to induce an attack response 24 h later. These observations uncover an unexpected social experience-dependent component to the formation of hypothalamic neural assemblies controlling innate social behaviours. More generally, they reveal plasticity and dynamic coding in an evolutionarily ancient deep subcortical structure that is traditionally viewed as a ‘hard-wired’ system.

  17. Advanced Atmospheric Ensemble Modeling Techniques

    Energy Technology Data Exchange (ETDEWEB)

    Buckley, R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Chiswell, S. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Kurzeja, R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Maze, G. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Viner, B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Werth, D. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2017-09-29

    Ensemble modeling (EM), the creation of multiple atmospheric simulations for a given time period, has become an essential tool for characterizing uncertainties in model predictions. We explore two novel ensemble modeling techniques: (1) perturbation of model parameters (Adaptive Programming, AP), and (2) data assimilation (Ensemble Kalman Filter, EnKF). The current research is an extension to work from last year and examines transport on a small spatial scale (<100 km) in complex terrain, for more rigorous testing of the ensemble technique. Two different release cases were studied, a coastal release (SF6) and an inland release (Freon) which consisted of two release times. Observations of tracer concentration and meteorology are used to judge the ensemble results. In addition, adaptive grid techniques have been developed to reduce required computing resources for transport calculations. Using a 20- member ensemble, the standard approach generated downwind transport that was quantitatively good for both releases; however, the EnKF method produced additional improvement for the coastal release where the spatial and temporal differences due to interior valley heating lead to the inland movement of the plume. The AP technique showed improvements for both release cases, with more improvement shown in the inland release. This research demonstrated that transport accuracy can be improved when models are adapted to a particular location/time or when important local data is assimilated into the simulation and enhances SRNL’s capability in atmospheric transport modeling in support of its current customer base and local site missions, as well as our ability to attract new customers within the intelligence community.

  18. Layered Ensemble Architecture for Time Series Forecasting.

    Science.gov (United States)

    Rahman, Md Mustafizur; Islam, Md Monirul; Murase, Kazuyuki; Yao, Xin

    2016-01-01

    Time series forecasting (TSF) has been widely used in many application areas such as science, engineering, and finance. The phenomena generating time series are usually unknown and information available for forecasting is only limited to the past values of the series. It is, therefore, necessary to use an appropriate number of past values, termed lag, for forecasting. This paper proposes a layered ensemble architecture (LEA) for TSF problems. Our LEA consists of two layers, each of which uses an ensemble of multilayer perceptron (MLP) networks. While the first ensemble layer tries to find an appropriate lag, the second ensemble layer employs the obtained lag for forecasting. Unlike most previous work on TSF, the proposed architecture considers both accuracy and diversity of the individual networks in constructing an ensemble. LEA trains different networks in the ensemble by using different training sets with an aim of maintaining diversity among the networks. However, it uses the appropriate lag and combines the best trained networks to construct the ensemble. This indicates LEAs emphasis on accuracy of the networks. The proposed architecture has been tested extensively on time series data of neural network (NN)3 and NN5 competitions. It has also been tested on several standard benchmark time series data. In terms of forecasting accuracy, our experimental results have revealed clearly that LEA is better than other ensemble and nonensemble methods.

  19. Dynamical mean-field theory of noisy spiking neuron ensembles: Application to the Hodgkin-Huxley model

    International Nuclear Information System (INIS)

    Hasegawa, Hideo

    2003-01-01

    A dynamical mean-field approximation (DMA) previously proposed by the present author [H. Hasegawa, Phys. Rev E 67, 041903 (2003)] has been extended to ensembles described by a general noisy spiking neuron model. Ensembles of N-unit neurons, each of which is expressed by coupled K-dimensional differential equations (DEs), are assumed to be subject to spatially correlated white noises. The original KN-dimensional stochastic DEs have been replaced by K(K+2)-dimensional deterministic DEs expressed in terms of means and the second-order moments of local and global variables: the fourth-order contributions are taken into account by the Gaussian decoupling approximation. Our DMA has been applied to an ensemble of Hodgkin-Huxley (HH) neurons (K=4), for which effects of the noise, the coupling strength, and the ensemble size on the response to a single-spike input have been investigated. Numerical results calculated by the DMA theory are in good agreement with those obtained by direct simulations, although the former computation is about a thousand times faster than the latter for a typical HH neuron ensemble with N=100

  20. Ensembles and Experiments in Classical and Quantum Physics

    Science.gov (United States)

    Neumaier, Arnold

    A philosophically consistent axiomatic approach to classical and quantum mechanics is given. The approach realizes a strong formal implementation of Bohr's correspondence principle. In all instances, classical and quantum concepts are fully parallel: the same general theory has a classical realization and a quantum realization. Extending the ''probability via expectation'' approach of Whittle to noncommuting quantities, this paper defines quantities, ensembles, and experiments as mathematical concepts and shows how to model complementarity, uncertainty, probability, nonlocality and dynamics in these terms. The approach carries no connotation of unlimited repeatability; hence it can be applied to unique systems such as the universe. Consistent experiments provide an elegant solution to the reality problem, confirming the insistence of the orthodox Copenhagen interpretation on that there is nothing but ensembles, while avoiding its elusive reality picture. The weak law of large numbers explains the emergence of classical properties for macroscopic systems.

  1. Radionuclide release from simulated waste material after biogeochemical leaching of uraniferous mineral samples

    International Nuclear Information System (INIS)

    Williamson, Aimee Lynn; Caron, François; Spiers, Graeme

    2014-01-01

    Biogeochemical mineral dissolution is a promising method for the released of metals in low-grade host mineralization that contain sulphidic minerals. The application of biogeochemical mineral dissolution to engineered leach heap piles in the Elliot Lake region may be considered as a promising passive technology for the economic recovery of low grade Uranium-bearing ores. In the current investigation, the decrease of radiological activity of uraniferous mineral material after biogeochemical mineral dissolution is quantified by gamma spectroscopy and compared to the results from digestion/ICP-MS analysis of the ore materials to determine if gamma spectroscopy is a simple, viable alternative quantification method for heavy nuclides. The potential release of Uranium (U) and Radium-226 ( 226 Ra) to the aqueous environment from samples that have been treated to represent various stages of leaching and passive closure processes are assessed. Dissolution of U from the solid phase has occurred during biogeochemical mineral dissolution in the presence of Acidithiobacillus ferrooxidans, with gamma spectroscopy indicating an 84% decrease in Uranium-235 ( 235 U) content, a value in accordance with the data obtained by dissolution chemistry. Gamma spectroscopy data indicate that only 30% of the 226 Ra was removed during the biogeochemical mineral dissolution. Chemical inhibition and passivation treatments of waste materials following the biogeochemical mineral dissolution offer greater protection against residual U and 226 Ra leaching. Pacified samples resist the release of 226 Ra contained in the mineral phase and may offer more protection to the aqueous environment for the long term, compared to untreated or inhibited residues, and should be taken into account for future decommissioning. - Highlights: • Gamma counting showed an 84% decrease in 235 U after biogeochemical mineral leaching. • Chemical digestion/ICP-MS analysis also showed an 84% decrease in total U. • Over

  2. Rotationally invariant family of Levy-like random matrix ensembles

    International Nuclear Information System (INIS)

    Choi, Jinmyung; Muttalib, K A

    2009-01-01

    We introduce a family of rotationally invariant random matrix ensembles characterized by a parameter λ. While λ = 1 corresponds to well-known critical ensembles, we show that λ ≠ 1 describes 'Levy-like' ensembles, characterized by power-law eigenvalue densities. For λ > 1 the density is bounded, as in Gaussian ensembles, but λ < 1 describes ensembles characterized by densities with long tails. In particular, the model allows us to evaluate, in terms of a novel family of orthogonal polynomials, the eigenvalue correlations for Levy-like ensembles. These correlations differ qualitatively from those in either the Gaussian or the critical ensembles. (fast track communication)

  3. Diversity in random subspacing ensembles

    NARCIS (Netherlands)

    Tsymbal, A.; Pechenizkiy, M.; Cunningham, P.; Kambayashi, Y.; Mohania, M.K.; Wöß, W.

    2004-01-01

    Ensembles of learnt models constitute one of the main current directions in machine learning and data mining. It was shown experimentally and theoretically that in order for an ensemble to be effective, it should consist of classifiers having diversity in their predictions. A number of ways are

  4. Can decadal climate predictions be improved by ocean ensemble dispersion filtering?

    Science.gov (United States)

    Kadow, C.; Illing, S.; Kröner, I.; Ulbrich, U.; Cubasch, U.

    2017-12-01

    Decadal predictions by Earth system models aim to capture the state and phase of the climate several years inadvance. Atmosphere-ocean interaction plays an important role for such climate forecasts. While short-termweather forecasts represent an initial value problem and long-term climate projections represent a boundarycondition problem, the decadal climate prediction falls in-between these two time scales. The ocean memorydue to its heat capacity holds big potential skill on the decadal scale. In recent years, more precise initializationtechniques of coupled Earth system models (incl. atmosphere and ocean) have improved decadal predictions.Ensembles are another important aspect. Applying slightly perturbed predictions results in an ensemble. Insteadof using and evaluating one prediction, but the whole ensemble or its ensemble average, improves a predictionsystem. However, climate models in general start losing the initialized signal and its predictive skill from oneforecast year to the next. Here we show that the climate prediction skill of an Earth system model can be improvedby a shift of the ocean state toward the ensemble mean of its individual members at seasonal intervals. Wefound that this procedure, called ensemble dispersion filter, results in more accurate results than the standarddecadal prediction. Global mean and regional temperature, precipitation, and winter cyclone predictions showan increased skill up to 5 years ahead. Furthermore, the novel technique outperforms predictions with largerensembles and higher resolution. Our results demonstrate how decadal climate predictions benefit from oceanensemble dispersion filtering toward the ensemble mean. This study is part of MiKlip (fona-miklip.de) - a major project on decadal climate prediction in Germany.We focus on the Max-Planck-Institute Earth System Model using the low-resolution version (MPI-ESM-LR) andMiKlip's basic initialization strategy as in 2017 published decadal climate forecast: http

  5. Squeezing of Collective Excitations in Spin Ensembles

    DEFF Research Database (Denmark)

    Kraglund Andersen, Christian; Mølmer, Klaus

    2012-01-01

    We analyse the possibility to create two-mode spin squeezed states of two separate spin ensembles by inverting the spins in one ensemble and allowing spin exchange between the ensembles via a near resonant cavity field. We investigate the dynamics of the system using a combination of numerical an...

  6. AUC-Maximizing Ensembles through Metalearning.

    Science.gov (United States)

    LeDell, Erin; van der Laan, Mark J; Petersen, Maya

    2016-05-01

    Area Under the ROC Curve (AUC) is often used to measure the performance of an estimator in binary classification problems. An AUC-maximizing classifier can have significant advantages in cases where ranking correctness is valued or if the outcome is rare. In a Super Learner ensemble, maximization of the AUC can be achieved by the use of an AUC-maximining metalearning algorithm. We discuss an implementation of an AUC-maximization technique that is formulated as a nonlinear optimization problem. We also evaluate the effectiveness of a large number of different nonlinear optimization algorithms to maximize the cross-validated AUC of the ensemble fit. The results provide evidence that AUC-maximizing metalearners can, and often do, out-perform non-AUC-maximizing metalearning methods, with respect to ensemble AUC. The results also demonstrate that as the level of imbalance in the training data increases, the Super Learner ensemble outperforms the top base algorithm by a larger degree.

  7. Urban runoff forecasting with ensemble weather predictions

    DEFF Research Database (Denmark)

    Pedersen, Jonas Wied; Courdent, Vianney Augustin Thomas; Vezzaro, Luca

    This research shows how ensemble weather forecasts can be used to generate urban runoff forecasts up to 53 hours into the future. The results highlight systematic differences between ensemble members that needs to be accounted for when these forecasts are used in practice.......This research shows how ensemble weather forecasts can be used to generate urban runoff forecasts up to 53 hours into the future. The results highlight systematic differences between ensemble members that needs to be accounted for when these forecasts are used in practice....

  8. A MITgcm/DART ensemble analysis and prediction system with application to the Gulf of Mexico

    KAUST Repository

    Hoteit, Ibrahim

    2013-09-01

    This paper describes the development of an advanced ensemble Kalman filter (EnKF)-based ocean data assimilation system for prediction of the evolution of the loop current in the Gulf of Mexico (GoM). The system integrates the Data Assimilation Research Testbed (DART) assimilation package with the Massachusetts Institute of Technology ocean general circulation model (MITgcm). The MITgcm/DART system supports the assimilation of a wide range of ocean observations and uses an ensemble approach to solve the nonlinear assimilation problems. The GoM prediction system was implemented with an eddy-resolving 1/10th degree configuration of the MITgcm. Assimilation experiments were performed over a 6-month period between May and October during a strong loop current event in 1999. The model was sequentially constrained with weekly satellite sea surface temperature and altimetry data. Experiments results suggest that the ensemble-based assimilation system shows a high predictive skill in the GoM, with estimated ensemble spread mainly concentrated around the front of the loop current. Further analysis of the system estimates demonstrates that the ensemble assimilation accurately reproduces the observed features without imposing any negative impact on the dynamical balance of the system. Results from sensitivity experiments with respect to the ensemble filter parameters are also presented and discussed. © 2013 Elsevier B.V.

  9. Multilevel ensemble Kalman filtering

    KAUST Repository

    Hoel, Haakon

    2016-01-08

    The ensemble Kalman filter (EnKF) is a sequential filtering method that uses an ensemble of particle paths to estimate the means and covariances required by the Kalman filter by the use of sample moments, i.e., the Monte Carlo method. EnKF is often both robust and efficient, but its performance may suffer in settings where the computational cost of accurate simulations of particles is high. The multilevel Monte Carlo method (MLMC) is an extension of classical Monte Carlo methods which by sampling stochastic realizations on a hierarchy of resolutions may reduce the computational cost of moment approximations by orders of magnitude. In this work we have combined the ideas of MLMC and EnKF to construct the multilevel ensemble Kalman filter (MLEnKF) for the setting of finite dimensional state and observation spaces. The main ideas of this method is to compute particle paths on a hierarchy of resolutions and to apply multilevel estimators on the ensemble hierarchy of particles to compute Kalman filter means and covariances. Theoretical results and a numerical study of the performance gains of MLEnKF over EnKF will be presented. Some ideas on the extension of MLEnKF to settings with infinite dimensional state spaces will also be presented.

  10. Multilevel ensemble Kalman filtering

    KAUST Repository

    Hoel, Haakon; Chernov, Alexey; Law, Kody; Nobile, Fabio; Tempone, Raul

    2016-01-01

    The ensemble Kalman filter (EnKF) is a sequential filtering method that uses an ensemble of particle paths to estimate the means and covariances required by the Kalman filter by the use of sample moments, i.e., the Monte Carlo method. EnKF is often both robust and efficient, but its performance may suffer in settings where the computational cost of accurate simulations of particles is high. The multilevel Monte Carlo method (MLMC) is an extension of classical Monte Carlo methods which by sampling stochastic realizations on a hierarchy of resolutions may reduce the computational cost of moment approximations by orders of magnitude. In this work we have combined the ideas of MLMC and EnKF to construct the multilevel ensemble Kalman filter (MLEnKF) for the setting of finite dimensional state and observation spaces. The main ideas of this method is to compute particle paths on a hierarchy of resolutions and to apply multilevel estimators on the ensemble hierarchy of particles to compute Kalman filter means and covariances. Theoretical results and a numerical study of the performance gains of MLEnKF over EnKF will be presented. Some ideas on the extension of MLEnKF to settings with infinite dimensional state spaces will also be presented.

  11. Developing an approach to effectively use super ensemble experiments for the projection of hydrological extremes under climate change

    Science.gov (United States)

    Watanabe, S.; Kim, H.; Utsumi, N.

    2017-12-01

    This study aims to develop a new approach which projects hydrology under climate change using super ensemble experiments. The use of multiple ensemble is essential for the estimation of extreme, which is a major issue in the impact assessment of climate change. Hence, the super ensemble experiments are recently conducted by some research programs. While it is necessary to use multiple ensemble, the multiple calculations of hydrological simulation for each output of ensemble simulations needs considerable calculation costs. To effectively use the super ensemble experiments, we adopt a strategy to use runoff projected by climate models directly. The general approach of hydrological projection is to conduct hydrological model simulations which include land-surface and river routing process using atmospheric boundary conditions projected by climate models as inputs. This study, on the other hand, simulates only river routing model using runoff projected by climate models. In general, the climate model output is systematically biased so that a preprocessing which corrects such bias is necessary for impact assessments. Various bias correction methods have been proposed, but, to the best of our knowledge, no method has proposed for variables other than surface meteorology. Here, we newly propose a method for utilizing the projected future runoff directly. The developed method estimates and corrects the bias based on the pseudo-observation which is a result of retrospective offline simulation. We show an application of this approach to the super ensemble experiments conducted under the program of Half a degree Additional warming, Prognosis and Projected Impacts (HAPPI). More than 400 ensemble experiments from multiple climate models are available. The results of the validation using historical simulations by HAPPI indicates that the output of this approach can effectively reproduce retrospective runoff variability. Likewise, the bias of runoff from super ensemble climate

  12. Functional Enzyme-Based Approach for Linking Microbial Community Functions with Biogeochemical Process Kinetics

    Energy Technology Data Exchange (ETDEWEB)

    Li, Minjing [School; Qian, Wei-jun [Pacific Northwest National Laboratory, Richland, Washington 99354, United States; Gao, Yuqian [Pacific Northwest National Laboratory, Richland, Washington 99354, United States; Shi, Liang [School; Liu, Chongxuan [Pacific Northwest National Laboratory, Richland, Washington 99354, United States; School

    2017-09-28

    The kinetics of biogeochemical processes in natural and engineered environmental systems are typically described using Monod-type or modified Monod-type models. These models rely on biomass as surrogates for functional enzymes in microbial community that catalyze biogeochemical reactions. A major challenge to apply such models is the difficulty to quantitatively measure functional biomass for constraining and validating the models. On the other hand, omics-based approaches have been increasingly used to characterize microbial community structure, functions, and metabolites. Here we proposed an enzyme-based model that can incorporate omics-data to link microbial community functions with biogeochemical process kinetics. The model treats enzymes as time-variable catalysts for biogeochemical reactions and applies biogeochemical reaction network to incorporate intermediate metabolites. The sequences of genes and proteins from metagenomes, as well as those from the UniProt database, were used for targeted enzyme quantification and to provide insights into the dynamic linkage among functional genes, enzymes, and metabolites that are necessary to be incorporated in the model. The application of the model was demonstrated using denitrification as an example by comparing model-simulated with measured functional enzymes, genes, denitrification substrates and intermediates

  13. Multimodel Ensembles of Wheat Growth: Many Models are Better than One

    Science.gov (United States)

    Martre, Pierre; Wallach, Daniel; Asseng, Senthold; Ewert, Frank; Jones, James W.; Rotter, Reimund P.; Boote, Kenneth J.; Ruane, Alexander C.; Thorburn, Peter J.; Cammarano, Davide; hide

    2015-01-01

    Crop models of crop growth are increasingly used to quantify the impact of global changes due to climate or crop management. Therefore, accuracy of simulation results is a major concern. Studies with ensembles of crop model scan give valuable information about model accuracy and uncertainty, but such studies are difficult to organize and have only recently begun. We report on the largest ensemble study to date, of 27 wheat models tested in four contrasting locations for their accuracy in simulating multiple crop growth and yield variables. The relative error averaged over models was 2438 for the different end-of-season variables including grain yield (GY) and grain protein concentration (GPC). There was little relation between error of a model for GY or GPC and error for in-season variables. Thus, most models did not arrive at accurate simulations of GY and GPC by accurately simulating preceding growth dynamics. Ensemble simulations, taking either the mean (e-mean) or median (e-median) of simulated values, gave better estimates than any individual model when all variables were considered. Compared to individual models, e-median ranked first in simulating measured GY and third in GPC. The error of e-mean and e-median declined with an increasing number of ensemble members, with little decrease beyond 10 models. We conclude that multimodel ensembles can be used to create new estimators with improved accuracy and consistency in simulating growth dynamics. We argue that these results are applicable to other crop species, and hypothesize that they apply more generally to ecological system models.

  14. Multimodel Ensembles of Wheat Growth: More Models are Better than One

    Science.gov (United States)

    Martre, Pierre; Wallach, Daniel; Asseng, Senthold; Ewert, Frank; Jones, James W.; Rotter, Reimund P.; Boote, Kenneth J.; Ruane, Alex C.; Thorburn, Peter J.; Cammarano, Davide; hide

    2015-01-01

    Crop models of crop growth are increasingly used to quantify the impact of global changes due to climate or crop management. Therefore, accuracy of simulation results is a major concern. Studies with ensembles of crop models can give valuable information about model accuracy and uncertainty, but such studies are difficult to organize and have only recently begun. We report on the largest ensemble study to date, of 27 wheat models tested in four contrasting locations for their accuracy in simulating multiple crop growth and yield variables. The relative error averaged over models was 24-38% for the different end-of-season variables including grain yield (GY) and grain protein concentration (GPC). There was little relation between error of a model for GY or GPC and error for in-season variables. Thus, most models did not arrive at accurate simulations of GY and GPC by accurately simulating preceding growth dynamics. Ensemble simulations, taking either the mean (e-mean) or median (e-median) of simulated values, gave better estimates than any individual model when all variables were considered. Compared to individual models, e-median ranked first in simulating measured GY and third in GPC. The error of e-mean and e-median declined with an increasing number of ensemble members, with little decrease beyond 10 models. We conclude that multimodel ensembles can be used to create new estimators with improved accuracy and consistency in simulating growth dynamics. We argue that these results are applicable to other crop species, and hypothesize that they apply more generally to ecological system models.

  15. Bayesian ensemble refinement by replica simulations and reweighting

    Science.gov (United States)

    Hummer, Gerhard; Köfinger, Jürgen

    2015-12-01

    We describe different Bayesian ensemble refinement methods, examine their interrelation, and discuss their practical application. With ensemble refinement, the properties of dynamic and partially disordered (bio)molecular structures can be characterized by integrating a wide range of experimental data, including measurements of ensemble-averaged observables. We start from a Bayesian formulation in which the posterior is a functional that ranks different configuration space distributions. By maximizing this posterior, we derive an optimal Bayesian ensemble distribution. For discrete configurations, this optimal distribution is identical to that obtained by the maximum entropy "ensemble refinement of SAXS" (EROS) formulation. Bayesian replica ensemble refinement enhances the sampling of relevant configurations by imposing restraints on averages of observables in coupled replica molecular dynamics simulations. We show that the strength of the restraints should scale linearly with the number of replicas to ensure convergence to the optimal Bayesian result in the limit of infinitely many replicas. In the "Bayesian inference of ensembles" method, we combine the replica and EROS approaches to accelerate the convergence. An adaptive algorithm can be used to sample directly from the optimal ensemble, without replicas. We discuss the incorporation of single-molecule measurements and dynamic observables such as relaxation parameters. The theoretical analysis of different Bayesian ensemble refinement approaches provides a basis for practical applications and a starting point for further investigations.

  16. Intra- versus inter-site macroscale variation in biogeochemical properties along a paddy soil chronosequence

    Directory of Open Access Journals (Sweden)

    C. Mueller-Niggemann

    2012-03-01

    Full Text Available In order to assess the intrinsic heterogeneity of paddy soils, a set of biogeochemical soil parameters was investigated in five field replicates of seven paddy fields (50, 100, 300, 500, 700, 1000, and 2000 yr of wetland rice cultivation, one flooded paddy nursery, one tidal wetland (TW, and one freshwater site (FW from a coastal area at Hangzhou Bay, Zhejiang Province, China. All soils evolved from a marine tidal flat substrate due to land reclamation. The biogeochemical parameters based on their properties were differentiated into (i a group behaving conservatively (TC, TOC, TN, TS, magnetic susceptibility, soil lightness and colour parameters, δ13C, δ15N, lipids and n-alkanes and (ii one encompassing more labile properties or fast cycling components (Nmic, Cmic, nitrate, ammonium, DON and DOC. The macroscale heterogeneity in paddy soils was assessed by evaluating intra- versus inter-site spatial variability of biogeochemical properties using statistical data analysis (descriptive, explorative and non-parametric. Results show that the intrinsic heterogeneity of paddy soil organic and minerogenic components per field is smaller than between study sites. The coefficient of variation (CV values of conservative parameters varied in a low range (10% to 20%, decreasing from younger towards older paddy soils. This indicates a declining variability of soil biogeochemical properties in longer used cropping sites according to progress in soil evolution. A generally higher variation of CV values (>20–40% observed for labile parameters implies a need for substantially higher sampling frequency when investigating these as compared to more conservative parameters. Since the representativeness of the sampling strategy could be sufficiently demonstrated, an investigation of long-term carbon accumulation/sequestration trends in topsoils of the 2000 yr paddy chronosequence under wetland rice cultivation

  17. Biogeochemical aspects of aquifer thermal energy storage

    NARCIS (Netherlands)

    Brons, H.J.

    1992-01-01

    During the process of aquifer thermal energy storage the in situ temperature of the groundwater- sediment system may fluctuate significantly. As a result the groundwater characteristics can be considerably affected by a variety of chemical, biogeochemical and microbiological

  18. Ensemble method for dengue prediction.

    Science.gov (United States)

    Buczak, Anna L; Baugher, Benjamin; Moniz, Linda J; Bagley, Thomas; Babin, Steven M; Guven, Erhan

    2018-01-01

    In the 2015 NOAA Dengue Challenge, participants made three dengue target predictions for two locations (Iquitos, Peru, and San Juan, Puerto Rico) during four dengue seasons: 1) peak height (i.e., maximum weekly number of cases during a transmission season; 2) peak week (i.e., week in which the maximum weekly number of cases occurred); and 3) total number of cases reported during a transmission season. A dengue transmission season is the 12-month period commencing with the location-specific, historical week with the lowest number of cases. At the beginning of the Dengue Challenge, participants were provided with the same input data for developing the models, with the prediction testing data provided at a later date. Our approach used ensemble models created by combining three disparate types of component models: 1) two-dimensional Method of Analogues models incorporating both dengue and climate data; 2) additive seasonal Holt-Winters models with and without wavelet smoothing; and 3) simple historical models. Of the individual component models created, those with the best performance on the prior four years of data were incorporated into the ensemble models. There were separate ensembles for predicting each of the three targets at each of the two locations. Our ensemble models scored higher for peak height and total dengue case counts reported in a transmission season for Iquitos than all other models submitted to the Dengue Challenge. However, the ensemble models did not do nearly as well when predicting the peak week. The Dengue Challenge organizers scored the dengue predictions of the Challenge participant groups. Our ensemble approach was the best in predicting the total number of dengue cases reported for transmission season and peak height for Iquitos, Peru.

  19. Ensemble method for dengue prediction.

    Directory of Open Access Journals (Sweden)

    Anna L Buczak

    Full Text Available In the 2015 NOAA Dengue Challenge, participants made three dengue target predictions for two locations (Iquitos, Peru, and San Juan, Puerto Rico during four dengue seasons: 1 peak height (i.e., maximum weekly number of cases during a transmission season; 2 peak week (i.e., week in which the maximum weekly number of cases occurred; and 3 total number of cases reported during a transmission season. A dengue transmission season is the 12-month period commencing with the location-specific, historical week with the lowest number of cases. At the beginning of the Dengue Challenge, participants were provided with the same input data for developing the models, with the prediction testing data provided at a later date.Our approach used ensemble models created by combining three disparate types of component models: 1 two-dimensional Method of Analogues models incorporating both dengue and climate data; 2 additive seasonal Holt-Winters models with and without wavelet smoothing; and 3 simple historical models. Of the individual component models created, those with the best performance on the prior four years of data were incorporated into the ensemble models. There were separate ensembles for predicting each of the three targets at each of the two locations.Our ensemble models scored higher for peak height and total dengue case counts reported in a transmission season for Iquitos than all other models submitted to the Dengue Challenge. However, the ensemble models did not do nearly as well when predicting the peak week.The Dengue Challenge organizers scored the dengue predictions of the Challenge participant groups. Our ensemble approach was the best in predicting the total number of dengue cases reported for transmission season and peak height for Iquitos, Peru.

  20. Ensemble stacking mitigates biases in inference of synaptic connectivity.

    Science.gov (United States)

    Chambers, Brendan; Levy, Maayan; Dechery, Joseph B; MacLean, Jason N

    2018-01-01

    A promising alternative to directly measuring the anatomical connections in a neuronal population is inferring the connections from the activity. We employ simulated spiking neuronal networks to compare and contrast commonly used inference methods that identify likely excitatory synaptic connections using statistical regularities in spike timing. We find that simple adjustments to standard algorithms improve inference accuracy: A signing procedure improves the power of unsigned mutual-information-based approaches and a correction that accounts for differences in mean and variance of background timing relationships, such as those expected to be induced by heterogeneous firing rates, increases the sensitivity of frequency-based methods. We also find that different inference methods reveal distinct subsets of the synaptic network and each method exhibits different biases in the accurate detection of reciprocity and local clustering. To correct for errors and biases specific to single inference algorithms, we combine methods into an ensemble. Ensemble predictions, generated as a linear combination of multiple inference algorithms, are more sensitive than the best individual measures alone, and are more faithful to ground-truth statistics of connectivity, mitigating biases specific to single inference methods. These weightings generalize across simulated datasets, emphasizing the potential for the broad utility of ensemble-based approaches.

  1. Contact planarization of ensemble nanowires

    Science.gov (United States)

    Chia, A. C. E.; LaPierre, R. R.

    2011-06-01

    The viability of four organic polymers (S1808, SC200, SU8 and Cyclotene) as filling materials to achieve planarization of ensemble nanowire arrays is reported. Analysis of the porosity, surface roughness and thermal stability of each filling material was performed. Sonication was used as an effective method to remove the tops of the nanowires (NWs) to achieve complete planarization. Ensemble nanowire devices were fully fabricated and I-V measurements confirmed that Cyclotene effectively planarizes the NWs while still serving the role as an insulating layer between the top and bottom contacts. These processes and analysis can be easily implemented into future characterization and fabrication of ensemble NWs for optoelectronic device applications.

  2. Ensemble forecasting of species distributions.

    Science.gov (United States)

    Araújo, Miguel B; New, Mark

    2007-01-01

    Concern over implications of climate change for biodiversity has led to the use of bioclimatic models to forecast the range shifts of species under future climate-change scenarios. Recent studies have demonstrated that projections by alternative models can be so variable as to compromise their usefulness for guiding policy decisions. Here, we advocate the use of multiple models within an ensemble forecasting framework and describe alternative approaches to the analysis of bioclimatic ensembles, including bounding box, consensus and probabilistic techniques. We argue that, although improved accuracy can be delivered through the traditional tasks of trying to build better models with improved data, more robust forecasts can also be achieved if ensemble forecasts are produced and analysed appropriately.

  3. The Advantage of Using International Multimodel Ensemble for Seasonal Precipitation Forecast over Israel

    Directory of Open Access Journals (Sweden)

    Amir Givati

    2017-01-01

    Full Text Available This study analyzes the results of monthly and seasonal precipitation forecasting from seven different global climate forecast models for major basins in Israel within October–April 1982–2010. The six National Multimodel Ensemble (NMME models and the ECMWF seasonal model were used to calculate an International Multimodel Ensemble (IMME. The study presents the performance of both monthly and seasonal predictions of precipitation accumulated over three months, with respect to different lead times for the ensemble mean values, one per individual model. Additionally, we analyzed the performance of different combinations of models. We present verification of seasonal forecasting using real forecasts, focusing on a small domain characterized by complex terrain, high annual precipitation variability, and a sharp precipitation gradient from west to east as well as from south to north. The results in this study show that, in general, the monthly analysis does not provide very accurate results, even when using the IMME for one-month lead time. We found that the IMME outperformed any single model prediction. Our analysis indicates that the optimal combinations with the high correlation values contain at least three models. Moreover, prediction with larger number of models in the ensemble produces more robust predictions. The results obtained in this study highlight the advantages of using an ensemble of global models over single models for small domain.

  4. Assessment of managed aquifer recharge potential using ensembles of local models.

    Science.gov (United States)

    Smith, Anthony J; Pollock, Daniel W

    2012-01-01

    A simple quantitative approach for assessing the artificial recharge potential of large regions using spatial ensembles of local models is proposed. The method extends existing qualitative approaches and enables rapid assessments within a programmable environment. Spatial discretization of a water resource region into continuous local domains allows simple local models to be applied independently in each domain using lumped parameters. The ensemble results can be analyzed directly or combined with other quantitative and thematic information and visualized as regional suitability maps. A case study considers the hydraulic potential for surface infiltration across a large water resource region using a published analytic model for basin recharge. The model solution was implemented within a geographic information system and evaluated independently in >21,000 local domains using lumped parameters derived from existing regional datasets. Computer execution times to run the whole ensemble and process the results were in the order of a few minutes. Relevant aspects of the case study results and general conclusions concerning the utility and limitations of the method are discussed. © 2011, CSIRO. Ground Water © 2011, National Ground Water Association.

  5. Assessment of Surface Air Temperature over China Using Multi-criterion Model Ensemble Framework

    Science.gov (United States)

    Li, J.; Zhu, Q.; Su, L.; He, X.; Zhang, X.

    2017-12-01

    The General Circulation Models (GCMs) are designed to simulate the present climate and project future trends. It has been noticed that the performances of GCMs are not always in agreement with each other over different regions. Model ensemble techniques have been developed to post-process the GCMs' outputs and improve their prediction reliabilities. To evaluate the performances of GCMs, root-mean-square error, correlation coefficient, and uncertainty are commonly used statistical measures. However, the simultaneous achievements of these satisfactory statistics cannot be guaranteed when using many model ensemble techniques. Meanwhile, uncertainties and future scenarios are critical for Water-Energy management and operation. In this study, a new multi-model ensemble framework was proposed. It uses a state-of-art evolutionary multi-objective optimization algorithm, termed Multi-Objective Complex Evolution Global Optimization with Principle Component Analysis and Crowding Distance (MOSPD), to derive optimal GCM ensembles and demonstrate the trade-offs among various solutions. Such trade-off information was further analyzed with a robust Pareto front with respect to different statistical measures. A case study was conducted to optimize the surface air temperature (SAT) ensemble solutions over seven geographical regions of China for the historical period (1900-2005) and future projection (2006-2100). The results showed that the ensemble solutions derived with MOSPD algorithm are superior over the simple model average and any single model output during the historical simulation period. For the future prediction, the proposed ensemble framework identified that the largest SAT change would occur in the South Central China under RCP 2.6 scenario, North Eastern China under RCP 4.5 scenario, and North Western China under RCP 8.5 scenario, while the smallest SAT change would occur in the Inner Mongolia under RCP 2.6 scenario, South Central China under RCP 4.5 scenario, and

  6. Reproducing multi-model ensemble average with Ensemble-averaged Reconstructed Forcings (ERF) in regional climate modeling

    Science.gov (United States)

    Erfanian, A.; Fomenko, L.; Wang, G.

    2016-12-01

    Multi-model ensemble (MME) average is considered the most reliable for simulating both present-day and future climates. It has been a primary reference for making conclusions in major coordinated studies i.e. IPCC Assessment Reports and CORDEX. The biases of individual models cancel out each other in MME average, enabling the ensemble mean to outperform individual members in simulating the mean climate. This enhancement however comes with tremendous computational cost, which is especially inhibiting for regional climate modeling as model uncertainties can originate from both RCMs and the driving GCMs. Here we propose the Ensemble-based Reconstructed Forcings (ERF) approach to regional climate modeling that achieves a similar level of bias reduction at a fraction of cost compared with the conventional MME approach. The new method constructs a single set of initial and boundary conditions (IBCs) by averaging the IBCs of multiple GCMs, and drives the RCM with this ensemble average of IBCs to conduct a single run. Using a regional climate model (RegCM4.3.4-CLM4.5), we tested the method over West Africa for multiple combination of (up to six) GCMs. Our results indicate that the performance of the ERF method is comparable to that of the MME average in simulating the mean climate. The bias reduction seen in ERF simulations is achieved by using more realistic IBCs in solving the system of equations underlying the RCM physics and dynamics. This endows the new method with a theoretical advantage in addition to reducing computational cost. The ERF output is an unaltered solution of the RCM as opposed to a climate state that might not be physically plausible due to the averaging of multiple solutions with the conventional MME approach. The ERF approach should be considered for use in major international efforts such as CORDEX. Key words: Multi-model ensemble, ensemble analysis, ERF, regional climate modeling

  7. Conductor gestures influence evaluations of ensemble performance.

    Science.gov (United States)

    Morrison, Steven J; Price, Harry E; Smedley, Eric M; Meals, Cory D

    2014-01-01

    Previous research has found that listener evaluations of ensemble performances vary depending on the expressivity of the conductor's gestures, even when performances are otherwise identical. It was the purpose of the present study to test whether this effect of visual information was evident in the evaluation of specific aspects of ensemble performance: articulation and dynamics. We constructed a set of 32 music performances that combined auditory and visual information and were designed to feature a high degree of contrast along one of two target characteristics: articulation and dynamics. We paired each of four music excerpts recorded by a chamber ensemble in both a high- and low-contrast condition with video of four conductors demonstrating high- and low-contrast gesture specifically appropriate to either articulation or dynamics. Using one of two equivalent test forms, college music majors and non-majors (N = 285) viewed sixteen 30 s performances and evaluated the quality of the ensemble's articulation, dynamics, technique, and tempo along with overall expressivity. Results showed significantly higher evaluations for performances featuring high rather than low conducting expressivity regardless of the ensemble's performance quality. Evaluations for both articulation and dynamics were strongly and positively correlated with evaluations of overall ensemble expressivity.

  8. Ensemble stacking mitigates biases in inference of synaptic connectivity

    Directory of Open Access Journals (Sweden)

    Brendan Chambers

    2018-03-01

    Full Text Available A promising alternative to directly measuring the anatomical connections in a neuronal population is inferring the connections from the activity. We employ simulated spiking neuronal networks to compare and contrast commonly used inference methods that identify likely excitatory synaptic connections using statistical regularities in spike timing. We find that simple adjustments to standard algorithms improve inference accuracy: A signing procedure improves the power of unsigned mutual-information-based approaches and a correction that accounts for differences in mean and variance of background timing relationships, such as those expected to be induced by heterogeneous firing rates, increases the sensitivity of frequency-based methods. We also find that different inference methods reveal distinct subsets of the synaptic network and each method exhibits different biases in the accurate detection of reciprocity and local clustering. To correct for errors and biases specific to single inference algorithms, we combine methods into an ensemble. Ensemble predictions, generated as a linear combination of multiple inference algorithms, are more sensitive than the best individual measures alone, and are more faithful to ground-truth statistics of connectivity, mitigating biases specific to single inference methods. These weightings generalize across simulated datasets, emphasizing the potential for the broad utility of ensemble-based approaches. Mapping the routing of spikes through local circuitry is crucial for understanding neocortical computation. Under appropriate experimental conditions, these maps can be used to infer likely patterns of synaptic recruitment, linking activity to underlying anatomical connections. Such inferences help to reveal the synaptic implementation of population dynamics and computation. We compare a number of standard functional measures to infer underlying connectivity. We find that regularization impacts measures

  9. IIASA's climate-vegetation-biogeochemical cycle module as a part of an integrated model for climate change

    International Nuclear Information System (INIS)

    Ganopolski, A.V.; Jonas, M.; Krabec, J.; Olendrzynski, K.; Petoukhov, V.K.; Venevsky, S.V.

    1994-01-01

    The main objective of this study is the development of a hierarchy of coupled climate biosphere models with a full description of the global biogeochemical cycles. These models are planned for use as the core of a set of integrated models of climate change and they will incorporate the main elements of the Earth system (atmosphere, hydrosphere, pedosphere and biosphere) linked with each other (and eventually with the antroposphere) through the fluxes of heat, momentum, water and through the global biogeochemical cycles of carbon and nitrogen. This set of integrated models can be considered to fill the gap between highly simplified integrated models of climate change and very sophisticated and computationally expensive coupled models, developed on the basis of general circulation models (GCMs). It is anticipated that this range of integrated models will be an effective tool for investigating the broad spectrum of problems connected with the coexistence of human society and biosphere

  10. Incorporating nitrogen fixing cyanobacteria in the global biogeochemical model HAMOCC

    Science.gov (United States)

    Paulsen, Hanna; Ilyina, Tatiana; Six, Katharina

    2015-04-01

    Nitrogen fixation by marine diazotrophs plays a fundamental role in the oceanic nitrogen and carbon cycle as it provides a major source of 'new' nitrogen to the euphotic zone that supports biological carbon export and sequestration. Since most global biogeochemical models include nitrogen fixation only diagnostically, they are not able to capture its spatial pattern sufficiently. Here we present the incorporation of an explicit, dynamic representation of diazotrophic cyanobacteria and the corresponding nitrogen fixation in the global ocean biogeochemical model HAMOCC (Hamburg Ocean Carbon Cycle model), which is part of the Max Planck Institute for Meteorology Earth system model (MPI-ESM). The parameterization of the diazotrophic growth is thereby based on available knowledge about the cyanobacterium Trichodesmium spp., which is considered as the most significant pelagic nitrogen fixer. Evaluation against observations shows that the model successfully reproduces the main spatial distribution of cyanobacteria and nitrogen fixation, covering large parts of the tropical and subtropical oceans. Besides the role of cyanobacteria in marine biogeochemical cycles, their capacity to form extensive surface blooms induces a number of bio-physical feedback mechanisms in the Earth system. The processes driving these interactions, which are related to the alteration of heat absorption, surface albedo and momentum input by wind, are incorporated in the biogeochemical and physical model of the MPI-ESM in order to investigate their impacts on a global scale. First preliminary results will be shown.

  11. Topological quantization of ensemble averages

    International Nuclear Information System (INIS)

    Prodan, Emil

    2009-01-01

    We define the current of a quantum observable and, under well-defined conditions, we connect its ensemble average to the index of a Fredholm operator. The present work builds on a formalism developed by Kellendonk and Schulz-Baldes (2004 J. Funct. Anal. 209 388) to study the quantization of edge currents for continuous magnetic Schroedinger operators. The generalization given here may be a useful tool to scientists looking for novel manifestations of the topological quantization. As a new application, we show that the differential conductance of atomic wires is given by the index of a certain operator. We also comment on how the formalism can be used to probe the existence of edge states

  12. Sequential ensemble-based optimal design for parameter estimation: SEQUENTIAL ENSEMBLE-BASED OPTIMAL DESIGN

    Energy Technology Data Exchange (ETDEWEB)

    Man, Jun [Zhejiang Provincial Key Laboratory of Agricultural Resources and Environment, Institute of Soil and Water Resources and Environmental Science, College of Environmental and Resource Sciences, Zhejiang University, Hangzhou China; Zhang, Jiangjiang [Zhejiang Provincial Key Laboratory of Agricultural Resources and Environment, Institute of Soil and Water Resources and Environmental Science, College of Environmental and Resource Sciences, Zhejiang University, Hangzhou China; Li, Weixuan [Pacific Northwest National Laboratory, Richland Washington USA; Zeng, Lingzao [Zhejiang Provincial Key Laboratory of Agricultural Resources and Environment, Institute of Soil and Water Resources and Environmental Science, College of Environmental and Resource Sciences, Zhejiang University, Hangzhou China; Wu, Laosheng [Department of Environmental Sciences, University of California, Riverside California USA

    2016-10-01

    The ensemble Kalman filter (EnKF) has been widely used in parameter estimation for hydrological models. The focus of most previous studies was to develop more efficient analysis (estimation) algorithms. On the other hand, it is intuitively understandable that a well-designed sampling (data-collection) strategy should provide more informative measurements and subsequently improve the parameter estimation. In this work, a Sequential Ensemble-based Optimal Design (SEOD) method, coupled with EnKF, information theory and sequential optimal design, is proposed to improve the performance of parameter estimation. Based on the first-order and second-order statistics, different information metrics including the Shannon entropy difference (SD), degrees of freedom for signal (DFS) and relative entropy (RE) are used to design the optimal sampling strategy, respectively. The effectiveness of the proposed method is illustrated by synthetic one-dimensional and two-dimensional unsaturated flow case studies. It is shown that the designed sampling strategies can provide more accurate parameter estimation and state prediction compared with conventional sampling strategies. Optimal sampling designs based on various information metrics perform similarly in our cases. The effect of ensemble size on the optimal design is also investigated. Overall, larger ensemble size improves the parameter estimation and convergence of optimal sampling strategy. Although the proposed method is applied to unsaturated flow problems in this study, it can be equally applied in any other hydrological problems.

  13. Aqueous complexation reactions governing the rate and extent of biogeochemical U(VI) reduction

    International Nuclear Information System (INIS)

    Kemner, K.M.; Kelly, S.D.; Brooks, Scott C.; Dong, Wenming; Carroll, Sue; Fredrickson, James K.

    2006-01-01

    The proposed research will elucidate the principal biogeochemical reactions that govern the concentration, chemical speciation, and reactivity of the redox-sensitive contaminant uranium. The results will provide an improved understanding and predictive capability of the mechanisms that govern the biogeochemical reduction of uranium in subsurface environments

  14. A Pseudoproxy-Ensemble Study of Late-Holocene Climate Field Reconstructions Using CCA

    Science.gov (United States)

    Amrhein, D. E.; Smerdon, J. E.

    2009-12-01

    Recent evaluations of late-Holocene multi-proxy reconstruction methods have used pseudoproxy experiments derived from millennial General Circulation Model (GCM) integrations. These experiments assess the performance of a reconstruction technique by comparing pseudoproxy reconstructions, which use restricted subsets of model data, against complete GCM data fields. Most previous studies have tested methodologies using different pseudoproxy noise levels, but only with single realizations for each noise classification. A more robust evaluation of performance is to create an ensemble of pseudoproxy networks with distinct sets of noise realizations and a corresponding reconstruction ensemble that can be evaluated for consistency and sensitivity to random error. This work investigates canonical correlation analysis (CCA) as a late-Holocene climate field reconstruction (CFR) technique using ensembles of pseudoproxy experiments derived from the NCAR CSM 1.4 millennial integration. Three 200-member reconstruction ensembles are computed using pseudoproxies with signal-to-noise ratios (by standard deviation) of 1, 0.5, and 0.25 and locations that approximate the spatial distribution of real-world multiproxy networks. An important component of these ensemble calculations is the independent optimization of the three CCA truncation parameters for each ensemble member. This task is accomplished using an inexpensive discrete optimization algorithm that minimizes both RMS error in the calibration interval and the number of free parameters in the reconstruction model to avoid artificial skill. Within this framework, CCA is investigated for its sensitivity to the level of noise in the pseudoproxy network and the spatial distribution of the network. Warm biases, variance losses, and validation-interval error increase with noise level and vary spatially within the reconstructed fields. Reconstruction skill, measured as grid-point correlations during the validation interval, is lowest in

  15. Biogeochemical speciation of Fe in ocean water

    NARCIS (Netherlands)

    Hiemstra, T.; Riemsdijk, van W.H.

    2006-01-01

    The biogeochemical speciation of Fe in seawater has been evaluated using the consistent Non-Ideal Competitive Adsorption model (NICA¿Donnan model). Two types of data sets were used, i.e. Fe-hydroxide solubility data and competitive ligand equilibration/cathodic stripping voltammetry (CLE/CSV) Fe

  16. Modality-Driven Classification and Visualization of Ensemble Variance

    Energy Technology Data Exchange (ETDEWEB)

    Bensema, Kevin; Gosink, Luke; Obermaier, Harald; Joy, Kenneth I.

    2016-10-01

    Advances in computational power now enable domain scientists to address conceptual and parametric uncertainty by running simulations multiple times in order to sufficiently sample the uncertain input space. While this approach helps address conceptual and parametric uncertainties, the ensemble datasets produced by this technique present a special challenge to visualization researchers as the ensemble dataset records a distribution of possible values for each location in the domain. Contemporary visualization approaches that rely solely on summary statistics (e.g., mean and variance) cannot convey the detailed information encoded in ensemble distributions that are paramount to ensemble analysis; summary statistics provide no information about modality classification and modality persistence. To address this problem, we propose a novel technique that classifies high-variance locations based on the modality of the distribution of ensemble predictions. Additionally, we develop a set of confidence metrics to inform the end-user of the quality of fit between the distribution at a given location and its assigned class. We apply a similar method to time-varying ensembles to illustrate the relationship between peak variance and bimodal or multimodal behavior. These classification schemes enable a deeper understanding of the behavior of the ensemble members by distinguishing between distributions that can be described by a single tendency and distributions which reflect divergent trends in the ensemble.

  17. Canonical Ensemble Model for Black Hole Radiation Jingyi Zhang

    Indian Academy of Sciences (India)

    Canonical Ensemble Model for Black Hole Radiation. 575. For entropy, there is no corresponding thermodynamical quantity, without loss of generalization. Let us define an entropy operator. ˆS = −KB ln ˆρ. (11). Then, the mean value of entropy is. S ≡〈ˆS〉 = tr( ˆρ ˆS) = −KBtr( ˆρ ln ˆρ). (12). For ideal gases, let y = V , then the ...

  18. Molecular biogeochemical provinces in the Atlantic Surface Ocean

    Science.gov (United States)

    Koch, B. P.; Flerus, R.; Schmitt-Kopplin, P.; Lechtenfeld, O. J.; Bracher, A.; Cooper, W.; Frka, S.; Gašparović, B.; Gonsior, M.; Hertkorn, N.; Jaffe, R.; Jenkins, A.; Kuss, J.; Lara, R. J.; Lucio, M.; McCallister, S. L.; Neogi, S. B.; Pohl, C.; Roettgers, R.; Rohardt, G.; Schmitt, B. B.; Stuart, A.; Theis, A.; Ying, W.; Witt, M.; Xie, Z.; Yamashita, Y.; Zhang, L.; Zhu, Z. Y.; Kattner, G.

    2010-12-01

    One of the most important aspects to understand marine organic carbon fluxes is to resolve the molecular mechanisms which convert fresh, labile biomolecules into semi-labile and refractory dissolved and particulate organic compounds in the ocean. In this interdisciplinary project, which was performed on a cruise with RV Polarstern, we carried out a detailed molecular characterisation of dissolved organic matter (DOM) on a North-South transect in the Atlantic surface ocean in order to relate the data to different biological, climatic, oceanographic, and meteorological regimes as well as to terrestrial input from riverine and atmospheric sources. Our goal was to achieve a high resolution data set for the biogeochemical characterisation of the sources and reactivity of DOM. We applied ultrahigh resolution Fourier Transform Ion Cyclotron Resonance Mass Spectrometry (FT-ICR-MS), nutrient, trace element, amino acid, and lipid analyses and other biogeochemical measurements for 220 samples from the upper water column (0-200m) and eight deep profiles. Various spectroscopic techniques were applied continuously in a constant sample water flow supplied by a fish system and the moon pool. Radiocarbon dating enabled assessing DOC residence time. Bacterial abundance and production provided a metabolic context for the DOM characterization work and pCO2 concentrations. Combining molecular organic techniques and inductively coupled plasma mass spectrometry (ICP-MS) established an important link between organic and inorganic biogeochemical studies. Multivariate statistics, primarily based on FT-ICR-MS data for 220 samples, allowed identifying geographical clusters which matched ecological provinces proposed previously by Longhurst (2007). Our study demonstrated that marine DOM carries molecular information reflecting the “history” of ocean water masses. This information can be used to define molecular biogeochemical provinces and to improve our understanding of element fluxes in

  19. Bioactive focus in conformational ensembles: a pluralistic approach

    Science.gov (United States)

    Habgood, Matthew

    2017-12-01

    Computational generation of conformational ensembles is key to contemporary drug design. Selecting the members of the ensemble that will approximate the conformation most likely to bind to a desired target (the bioactive conformation) is difficult, given that the potential energy usually used to generate and rank the ensemble is a notoriously poor discriminator between bioactive and non-bioactive conformations. In this study an approach to generating a focused ensemble is proposed in which each conformation is assigned multiple rankings based not just on potential energy but also on solvation energy, hydrophobic or hydrophilic interaction energy, radius of gyration, and on a statistical potential derived from Cambridge Structural Database data. The best ranked structures derived from each system are then assembled into a new ensemble that is shown to be better focused on bioactive conformations. This pluralistic approach is tested on ensembles generated by the Molecular Operating Environment's Low Mode Molecular Dynamics module, and by the Cambridge Crystallographic Data Centre's conformation generator software.

  20. Biogeochemical cycles and continental ecosystems - Report on Science and Technology no. 27

    International Nuclear Information System (INIS)

    Pedro, Georges; Blanzat, Bernard; Albrecht, Pierre; Berthelin, Jacques; Boudot, Jean-Pierre; Munier-Lamy, Colette; Cossa, Daniel; Feix, Isabelle; Guillaumont, Robert; HUC, Alain Yves; Lavelle, Patrick; Lebrun, Michel; Lucas, Yves; Metivier, Henri; Ourisson, Guy; Raimbault, Patrick; Ranger, Jacques; Gerard, Frederic; Schmidt-Laine, Claudine; Dercourt, Jean; Gaillardet, Jerome; Bourrie, Guilhem; Trolard, Fabienne; Gerard, Frederic; Dambrine, Etienne; Meunier, Jean Dominique; Benoit, Marc; Breda, Nathalie; Dupouey, Jean-Luc; Granier, Andre; Franc, Alain; GARBAYE, Jean; Martin, Francis; Landmann, Guy; Loustau, Denis; Martinez, Jose; Crochon, Philippe; Gay, Jean-Didier; Peres, Jean-Marc; Tamponnet, Christian; Andreux, Francis; Tusseauvuillemin, Marie-Helene; Barker, Evelyne; Bouisset, Patrick; Germain, Pierre; Masson, Olivier; Boust, Dominique; Bailly du Bois, Pascal; Abdelouas, Abdesselam; Grambow, Bernd; Ansoborlo, Erich; Chiappini, Remo; Lobinski, Ryzsard; Montavon, Gilles; Moulin, Christophe; Moulin, Valerie; Ollivier, Bernard; Haeseler, Franck; Prieur, Daniel; Magot, Michel; Charmasson, Sabine; Poss, Roland; Grimaldi, Catherine; Grimaldi, Michel; Malet, Caroline

    2007-11-01

    environments is described very briefly to show that in this field everything at the surface of the planet is linked (chapter 5.2). In environments that are highly influenced by Man where the excess and overload of certain elements can lead to dysfunctions, we have focused on three types of environments: - the agricultural sector, where we looked at the practice of manuring using exogenous organic matter (EOM) which is found in areas of intensive breeding farms (chapter 6.1); - the industrial sector; we present a number of environments where certain elements are in excess, such as can happen in mining and metallurgy areas (chapter 6.2); - Finally, we looked at estuary environments because they are the collection point for numerous contaminants coming from the continents (chapter 6.3). The second unit proceeds from all we have learnt in the first, briefly that biogeochemical cycles are the consequence of complex systems that necessitate many scientific fields and numerous data. This consideration implies: a) further studies in certain scientific areas; b) a particular interest in data collection; c) the reliance on model building. Point a) constitutes the third part which stresses the development of all venues concerning the role of the soil as a living environment. Knowledge is still lacking in this area. Concerning this issue, two points are considered in particular: - research on micro-organisms that are crucial for transforming waste and other residues from living beings and their return to the mineral state (chapter 7); - research on soil organic matter, which plays a fundamental role because it is an intermediate between the mineral world and the living world (chapter 8). Points b) and c) are treated in the fourth part of the report which discusses model operations (chapter 9) and problems linked to a generalized collection of data (for instance, observatories, networks) (chapter 10)

  1. Demonstrating the value of larger ensembles in forecasting physical systems

    Directory of Open Access Journals (Sweden)

    Reason L. Machete

    2016-12-01

    Full Text Available Ensemble simulation propagates a collection of initial states forward in time in a Monte Carlo fashion. Depending on the fidelity of the model and the properties of the initial ensemble, the goal of ensemble simulation can range from merely quantifying variations in the sensitivity of the model all the way to providing actionable probability forecasts of the future. Whatever the goal is, success depends on the properties of the ensemble, and there is a longstanding discussion in meteorology as to the size of initial condition ensemble most appropriate for Numerical Weather Prediction. In terms of resource allocation: how is one to divide finite computing resources between model complexity, ensemble size, data assimilation and other components of the forecast system. One wishes to avoid undersampling information available from the model's dynamics, yet one also wishes to use the highest fidelity model available. Arguably, a higher fidelity model can better exploit a larger ensemble; nevertheless it is often suggested that a relatively small ensemble, say ~16 members, is sufficient and that larger ensembles are not an effective investment of resources. This claim is shown to be dubious when the goal is probabilistic forecasting, even in settings where the forecast model is informative but imperfect. Probability forecasts for a ‘simple’ physical system are evaluated at different lead times; ensembles of up to 256 members are considered. The pure density estimation context (where ensemble members are drawn from the same underlying distribution as the target differs from the forecasting context, where one is given a high fidelity (but imperfect model. In the forecasting context, the information provided by additional members depends also on the fidelity of the model, the ensemble formation scheme (data assimilation, the ensemble interpretation and the nature of the observational noise. The effect of increasing the ensemble size is quantified by

  2. Ensemble Weight Enumerators for Protograph LDPC Codes

    Science.gov (United States)

    Divsalar, Dariush

    2006-01-01

    Recently LDPC codes with projected graph, or protograph structures have been proposed. In this paper, finite length ensemble weight enumerators for LDPC codes with protograph structures are obtained. Asymptotic results are derived as the block size goes to infinity. In particular we are interested in obtaining ensemble average weight enumerators for protograph LDPC codes which have minimum distance that grows linearly with block size. As with irregular ensembles, linear minimum distance property is sensitive to the proportion of degree-2 variable nodes. In this paper the derived results on ensemble weight enumerators show that linear minimum distance condition on degree distribution of unstructured irregular LDPC codes is a sufficient but not a necessary condition for protograph LDPC codes.

  3. Classification of premalignant pancreatic cancer mass-spectrometry data using decision tree ensembles

    Directory of Open Access Journals (Sweden)

    Wong G William

    2008-06-01

    Full Text Available Abstract Background Pancreatic cancer is the fourth leading cause of cancer death in the United States. Consequently, identification of clinically relevant biomarkers for the early detection of this cancer type is urgently needed. In recent years, proteomics profiling techniques combined with various data analysis methods have been successfully used to gain critical insights into processes and mechanisms underlying pathologic conditions, particularly as they relate to cancer. However, the high dimensionality of proteomics data combined with their relatively small sample sizes poses a significant challenge to current data mining methodology where many of the standard methods cannot be applied directly. Here, we propose a novel methodological framework using machine learning method, in which decision tree based classifier ensembles coupled with feature selection methods, is applied to proteomics data generated from premalignant pancreatic cancer. Results This study explores the utility of three different feature selection schemas (Student t test, Wilcoxon rank sum test and genetic algorithm to reduce the high dimensionality of a pancreatic cancer proteomic dataset. Using the top features selected from each method, we compared the prediction performances of a single decision tree algorithm C4.5 with six different decision-tree based classifier ensembles (Random forest, Stacked generalization, Bagging, Adaboost, Logitboost and Multiboost. We show that ensemble classifiers always outperform single decision tree classifier in having greater accuracies and smaller prediction errors when applied to a pancreatic cancer proteomics dataset. Conclusion In our cross validation framework, classifier ensembles generally have better classification accuracies compared to that of a single decision tree when applied to a pancreatic cancer proteomic dataset, thus suggesting its utility in future proteomics data analysis. Additionally, the use of feature selection

  4. Physical/biogeochemical coupled model : impact of an offline vs online strategy

    Science.gov (United States)

    Hameau, Angélique; Perruche, Coralie; Bricaud, Clément; Gutknecht, Elodie; Reffray, Guillaume

    2014-05-01

    Mercator-Ocean, the French ocean forecasting center, has been developing several operational forecasting systems and reanalysis of the physical and biogeochemical 3D-Ocean. Here we study the impact of an offline vs online strategy to couple the physical (OPA) and biogeochemical (PISCES) modules included in the NEMO platform. For this purpose, we perform global one-year long simulations at 1° resolution. The model was initialized with global climatologies. The spin-up involved 10 years of biogeochemical off-line simulation forced by a climatology of ocean physics. The online mode consists in running physical and biogeochemical models simultaneously whereas in the offline mode, the biogeochemical model is launched alone, forced by averaged physical forcing (1 day, 7 days,… ). The Mercator operational biogeochemical system is currently using the offline mode with a weekly physical forcing. A special treatment is applied to the vertical diffusivity coefficient (Kz): as it varies of several orders of magnitude, we compute the mean of the LOG10 of Kz. Moreover, a threshold value is applied to remove the highest values corresponding to enhanced convection. To improve this system, 2 directions are explored. First, 3 physical forcing frequencies are compared to quantify errors due to the offline mode: 1 hour (online mode), 1 day and 1 week (offline modes). Secondly, sensitivity tests to the threshold value applied to Kz are performed. The simulations are evaluated by systematically comparing model fields to observations (Globcolour product and World Ocean Atlas 2005) at global and regional scales. We show first that offline simulations are in good agreement with online simulation. As expected, the lower the physical forcing frequency is, the closer to the online solution is the offline simulation. The threshold value on the vertical diffusivity coefficient manages the mixing strength within the mixed layer. A value of 1 m2.s-1 appears to be a good compromise to approach

  5. Lessons Learned from Assimilating Altimeter Data into a Coupled General Circulation Model with the GMAO Augmented Ensemble Kalman Filter

    Science.gov (United States)

    Keppenne, Christian; Vernieres, Guillaume; Rienecker, Michele; Jacob, Jossy; Kovach, Robin

    2011-01-01

    Satellite altimetry measurements have provided global, evenly distributed observations of the ocean surface since 1993. However, the difficulties introduced by the presence of model biases and the requirement that data assimilation systems extrapolate the sea surface height (SSH) information to the subsurface in order to estimate the temperature, salinity and currents make it difficult to optimally exploit these measurements. This talk investigates the potential of the altimetry data assimilation once the biases are accounted for with an ad hoc bias estimation scheme. Either steady-state or state-dependent multivariate background-error covariances from an ensemble of model integrations are used to address the problem of extrapolating the information to the sub-surface. The GMAO ocean data assimilation system applied to an ensemble of coupled model instances using the GEOS-5 AGCM coupled to MOM4 is used in the investigation. To model the background error covariances, the system relies on a hybrid ensemble approach in which a small number of dynamically evolved model trajectories is augmented on the one hand with past instances of the state vector along each trajectory and, on the other, with a steady state ensemble of error estimates from a time series of short-term model forecasts. A state-dependent adaptive error-covariance localization and inflation algorithm controls how the SSH information is extrapolated to the sub-surface. A two-step predictor corrector approach is used to assimilate future information. Independent (not-assimilated) temperature and salinity observations from Argo floats are used to validate the assimilation. A two-step projection method in which the system first calculates a SSH increment and then projects this increment vertically onto the temperature, salt and current fields is found to be most effective in reconstructing the sub-surface information. The performance of the system in reconstructing the sub-surface fields is particularly

  6. Ensemble atmospheric dispersion calculations for decision support systems

    International Nuclear Information System (INIS)

    Borysiewicz, M.; Potempski, S.; Galkowski, A.; Zelazny, R.

    2003-01-01

    This document describes two approaches to long-range atmospheric dispersion of pollutants based on the ensemble concept. In the first part of the report some experiences related to the exercises undertaken under the ENSEMBLE project of the European Union are presented. The second part is devoted to the implementation of mesoscale numerical prediction models RAMS and atmospheric dispersion model HYPACT on Beowulf cluster and theirs usage for ensemble forecasting and long range atmospheric ensemble dispersion calculations based on available meteorological data from NCEO, NOAA (USA). (author)

  7. Aqueous Complexation Reactions Governing the Rate and Extent of Biogeochemical U(VI) Reduction

    International Nuclear Information System (INIS)

    Scott C. Brooks; Wenming Dong; Sue Carroll; James K. Fredrickson; Kenneth M. Kemner; Shelly D. Kelly

    2006-01-01

    The proposed research will elucidate the principal biogeochemical reactions that govern the concentration, chemical speciation, and reactivity of the redox-sensitive contaminant uranium. The results will provide an improved understanding and predictive capability of the mechanisms that govern the biogeochemical reduction of uranium in subsurface environments. In addition, the work plan is designed to: (1) Generate fundamental scientific understanding on the relationship between U(VI) chemical speciation and its susceptibility to biogeochemical reduction reactions. (2) Elucidate the controls on the rate and extent of contaminant reactivity. (3) Provide new insights into the aqueous and solid speciation of U(VI)/U(IV) under representative groundwater conditions

  8. Consequences of climate change for biogeochemical cycling in forests of northeastern North America

    Science.gov (United States)

    John L. Campbell; Lindsey E. Rustad; Elizabeth W. Boyer; Sheila F. Christopher; Charles T. Driscoll; Ivan .J. Fernandez; Peter M. Groffman; Daniel Houle; Jana Kiekbusch; Alison H. Magill; Myron J. Mitchell; Scott V. Ollinger

    2009-01-01

    A critical component of assessing the impacts of climate change on forest ecosystems involves understanding associated changes in biogeochemical cycling of elements. Evidence from research on northeastern North American forests shows that direct effects of climate change will evoke changes in biogeochemical cycling by altering plant physiology forest productivity, and...

  9. The role of ensemble-based statistics in variational assimilation of cloud-affected observations from infrared imagers

    Science.gov (United States)

    Hacker, Joshua; Vandenberghe, Francois; Jung, Byoung-Jo; Snyder, Chris

    2017-04-01

    Effective assimilation of cloud-affected radiance observations from space-borne imagers, with the aim of improving cloud analysis and forecasting, has proven to be difficult. Large observation biases, nonlinear observation operators, and non-Gaussian innovation statistics present many challenges. Ensemble-variational data assimilation (EnVar) systems offer the benefits of flow-dependent background error statistics from an ensemble, and the ability of variational minimization to handle nonlinearity. The specific benefits of ensemble statistics, relative to static background errors more commonly used in variational systems, have not been quantified for the problem of assimilating cloudy radiances. A simple experiment framework is constructed with a regional NWP model and operational variational data assimilation system, to provide the basis understanding the importance of ensemble statistics in cloudy radiance assimilation. Restricting the observations to those corresponding to clouds in the background forecast leads to innovations that are more Gaussian. The number of large innovations is reduced compared to the more general case of all observations, but not eliminated. The Huber norm is investigated to handle the fat tails of the distributions, and allow more observations to be assimilated without the need for strict background checks that eliminate them. Comparing assimilation using only ensemble background error statistics with assimilation using only static background error statistics elucidates the importance of the ensemble statistics. Although the cost functions in both experiments converge to similar values after sufficient outer-loop iterations, the resulting cloud water, ice, and snow content are greater in the ensemble-based analysis. The subsequent forecasts from the ensemble-based analysis also retain more condensed water species, indicating that the local environment is more supportive of clouds. In this presentation we provide details that explain the

  10. Ensemble-Based Data Assimilation in Reservoir Characterization: A Review

    Directory of Open Access Journals (Sweden)

    Seungpil Jung

    2018-02-01

    Full Text Available This paper presents a review of ensemble-based data assimilation for strongly nonlinear problems on the characterization of heterogeneous reservoirs with different production histories. It concentrates on ensemble Kalman filter (EnKF and ensemble smoother (ES as representative frameworks, discusses their pros and cons, and investigates recent progress to overcome their drawbacks. The typical weaknesses of ensemble-based methods are non-Gaussian parameters, improper prior ensembles and finite population size. Three categorized approaches, to mitigate these limitations, are reviewed with recent accomplishments; improvement of Kalman gains, add-on of transformation functions, and independent evaluation of observed data. The data assimilation in heterogeneous reservoirs, applying the improved ensemble methods, is discussed on predicting unknown dynamic data in reservoir characterization.

  11. Internal Spin Control, Squeezing and Decoherence in Ensembles of Alkali Atomic Spins

    Science.gov (United States)

    Norris, Leigh Morgan

    particular, we find that state preparation using control of the internal hyperfine spin increases the entangling power of squeezing protocols when f>1/2. Post-processing of the ensemble using additional internal spin control converts this entanglement into metrologically useful spin squeezing. By employing a variation of the Holstein-Primakoff approximation, in which the collective spin observables of the atomic ensemble are treated as quadratures of a bosonic mode, we model entanglement generation, spin squeezing and the effects of internal spin control. The Holstein-Primakoff formalism also enables us to take into account the decoherence of the ensemble due to optical pumping. While most works ignore or treat optical pumping phenomenologically, we employ a master equation derived from first principles. Our analysis shows that state preparation and the hyperfine spin size have a substantial impact upon both the generation of spin squeezing and the decoherence of the ensemble. Through a numerical search, we determine state preparations that enhance squeezing protocols while remaining robust to optical pumping. Finally, most work on spin squeezing in atomic ensembles has treated the light as a plane wave that couples identically to all atoms. In the final part of this dissertation, we go beyond the customary plane wave approximation on the light and employ focused paraxial beams, which are more efficiently mode matched to the radiation pattern of the atomic ensemble. The mathematical formalism and the internal spin control techniques that we applied in the plane wave case are generalized to accommodate the non-homogeneous paraxial probe. We find the optimal geometries of the atomic ensemble and the probe for mode matching and generation of spin squeezing.

  12. Various multistage ensembles for prediction of heating energy consumption

    Directory of Open Access Journals (Sweden)

    Radisa Jovanovic

    2015-04-01

    Full Text Available Feedforward neural network models are created for prediction of daily heating energy consumption of a NTNU university campus Gloshaugen using actual measured data for training and testing. Improvement of prediction accuracy is proposed by using neural network ensemble. Previously trained feed-forward neural networks are first separated into clusters, using k-means algorithm, and then the best network of each cluster is chosen as member of an ensemble. Two conventional averaging methods for obtaining ensemble output are applied; simple and weighted. In order to achieve better prediction results, multistage ensemble is investigated. As second level, adaptive neuro-fuzzy inference system with various clustering and membership functions are used to aggregate the selected ensemble members. Feedforward neural network in second stage is also analyzed. It is shown that using ensemble of neural networks can predict heating energy consumption with better accuracy than the best trained single neural network, while the best results are achieved with multistage ensemble.

  13. Ensemble Machine Learning Methods and Applications

    CERN Document Server

    Ma, Yunqian

    2012-01-01

    It is common wisdom that gathering a variety of views and inputs improves the process of decision making, and, indeed, underpins a democratic society. Dubbed “ensemble learning” by researchers in computational intelligence and machine learning, it is known to improve a decision system’s robustness and accuracy. Now, fresh developments are allowing researchers to unleash the power of ensemble learning in an increasing range of real-world applications. Ensemble learning algorithms such as “boosting” and “random forest” facilitate solutions to key computational issues such as face detection and are now being applied in areas as diverse as object trackingand bioinformatics.   Responding to a shortage of literature dedicated to the topic, this volume offers comprehensive coverage of state-of-the-art ensemble learning techniques, including various contributions from researchers in leading industrial research labs. At once a solid theoretical study and a practical guide, the volume is a windfall for r...

  14. Biogeochemical cycling of radionuclides in the environment

    International Nuclear Information System (INIS)

    Livens, F.R.

    1990-01-01

    The biogeochemical cycling of radionuclides with other components such as nutrients around ecosystems is discussed. In particular the behaviour of cesium in freshwater ecosystems since the Chernobyl accident and the behaviour of technetium in the form of pertechnetate anions, TcO 4 , in marine ecosystems is considered. (UK)

  15. An Improved Ensemble of Random Vector Functional Link Networks Based on Particle Swarm Optimization with Double Optimization Strategy.

    Science.gov (United States)

    Ling, Qing-Hua; Song, Yu-Qing; Han, Fei; Yang, Dan; Huang, De-Shuang

    2016-01-01

    For ensemble learning, how to select and combine the candidate classifiers are two key issues which influence the performance of the ensemble system dramatically. Random vector functional link networks (RVFL) without direct input-to-output links is one of suitable base-classifiers for ensemble systems because of its fast learning speed, simple structure and good generalization performance. In this paper, to obtain a more compact ensemble system with improved convergence performance, an improved ensemble of RVFL based on attractive and repulsive particle swarm optimization (ARPSO) with double optimization strategy is proposed. In the proposed method, ARPSO is applied to select and combine the candidate RVFL. As for using ARPSO to select the optimal base RVFL, ARPSO considers both the convergence accuracy on the validation data and the diversity of the candidate ensemble system to build the RVFL ensembles. In the process of combining RVFL, the ensemble weights corresponding to the base RVFL are initialized by the minimum norm least-square method and then further optimized by ARPSO. Finally, a few redundant RVFL is pruned, and thus the more compact ensemble of RVFL is obtained. Moreover, in this paper, theoretical analysis and justification on how to prune the base classifiers on classification problem is presented, and a simple and practically feasible strategy for pruning redundant base classifiers on both classification and regression problems is proposed. Since the double optimization is performed on the basis of the single optimization, the ensemble of RVFL built by the proposed method outperforms that built by some single optimization methods. Experiment results on function approximation and classification problems verify that the proposed method could improve its convergence accuracy as well as reduce the complexity of the ensemble system.

  16. New approach to information fusion for Lipschitz classifiers ensembles: Application in multi-channel C-OTDR-monitoring systems

    Energy Technology Data Exchange (ETDEWEB)

    Timofeev, Andrey V.; Egorov, Dmitry V. [LPP “EqualiZoom”, Astana, 010000 (Kazakhstan)

    2016-06-08

    This paper presents new results concerning selection of an optimal information fusion formula for an ensemble of Lipschitz classifiers. The goal of information fusion is to create an integral classificatory which could provide better generalization ability of the ensemble while achieving a practically acceptable level of effectiveness. The problem of information fusion is very relevant for data processing in multi-channel C-OTDR-monitoring systems. In this case we have to effectively classify targeted events which appear in the vicinity of the monitored object. Solution of this problem is based on usage of an ensemble of Lipschitz classifiers each of which corresponds to a respective channel. We suggest a brand new method for information fusion in case of ensemble of Lipschitz classifiers. This method is called “The Weighing of Inversely as Lipschitz Constants” (WILC). Results of WILC-method practical usage in multichannel C-OTDR monitoring systems are presented.

  17. Popular Music and the Instrumental Ensemble.

    Science.gov (United States)

    Boespflug, George

    1999-01-01

    Discusses popular music, the role of the musical performer as a creator, and the styles of jazz and popular music. Describes the pop ensemble at the college level, focusing on improvisation, rehearsals, recording, and performance. Argues that pop ensembles be used in junior and senior high school. (CMK)

  18. Ensemble Kalman methods for inverse problems

    International Nuclear Information System (INIS)

    Iglesias, Marco A; Law, Kody J H; Stuart, Andrew M

    2013-01-01

    The ensemble Kalman filter (EnKF) was introduced by Evensen in 1994 (Evensen 1994 J. Geophys. Res. 99 10143–62) as a novel method for data assimilation: state estimation for noisily observed time-dependent problems. Since that time it has had enormous impact in many application domains because of its robustness and ease of implementation, and numerical evidence of its accuracy. In this paper we propose the application of an iterative ensemble Kalman method for the solution of a wide class of inverse problems. In this context we show that the estimate of the unknown function that we obtain with the ensemble Kalman method lies in a subspace A spanned by the initial ensemble. Hence the resulting error may be bounded above by the error found from the best approximation in this subspace. We provide numerical experiments which compare the error incurred by the ensemble Kalman method for inverse problems with the error of the best approximation in A, and with variants on traditional least-squares approaches, restricted to the subspace A. In so doing we demonstrate that the ensemble Kalman method for inverse problems provides a derivative-free optimization method with comparable accuracy to that achieved by traditional least-squares approaches. Furthermore, we also demonstrate that the accuracy is of the same order of magnitude as that achieved by the best approximation. Three examples are used to demonstrate these assertions: inversion of a compact linear operator; inversion of piezometric head to determine hydraulic conductivity in a Darcy model of groundwater flow; and inversion of Eulerian velocity measurements at positive times to determine the initial condition in an incompressible fluid. (paper)

  19. Ensemble-based data assimilation and optimal sensor placement for scalar source reconstruction

    Science.gov (United States)

    Mons, Vincent; Wang, Qi; Zaki, Tamer

    2017-11-01

    Reconstructing the characteristics of a scalar source from limited remote measurements in a turbulent flow is a problem of great interest for environmental monitoring, and is challenging due to several aspects. Firstly, the numerical estimation of the scalar dispersion in a turbulent flow requires significant computational resources. Secondly, in actual practice, only a limited number of observations are available, which generally makes the corresponding inverse problem ill-posed. Ensemble-based variational data assimilation techniques are adopted to solve the problem of scalar source localization in a turbulent channel flow at Reτ = 180 . This approach combines the components of variational data assimilation and ensemble Kalman filtering, and inherits the robustness from the former and the ease of implementation from the latter. An ensemble-based methodology for optimal sensor placement is also proposed in order to improve the condition of the inverse problem, which enhances the performances of the data assimilation scheme. This work has been partially funded by the Office of Naval Research (Grant N00014-16-1-2542) and by the National Science Foundation (Grant 1461870).

  20. Probability Maps for the Visualization of Assimilation Ensemble Flow Data

    KAUST Repository

    Hollt, Thomas

    2015-05-25

    Ocean forecasts nowadays are created by running ensemble simulations in combination with data assimilation techniques. Most of these techniques resample the ensemble members after each assimilation cycle. This means that in a time series, after resampling, every member can follow up on any of the members before resampling. Tracking behavior over time, such as all possible paths of a particle in an ensemble vector field, becomes very difficult, as the number of combinations rises exponentially with the number of assimilation cycles. In general a single possible path is not of interest but only the probabilities that any point in space might be reached by a particle at some point in time. In this work we present an approach using probability-weighted piecewise particle trajectories to allow such a mapping interactively, instead of tracing quadrillions of individual particles. We achieve interactive rates by binning the domain and splitting up the tracing process into the individual assimilation cycles, so that particles that fall into the same bin after a cycle can be treated as a single particle with a larger probability as input for the next time step. As a result we loose the possibility to track individual particles, but can create probability maps for any desired seed at interactive rates.

  1. Determination of the vapor–liquid transition of square-well particles using a novel generalized-canonical-ensemble-based method

    International Nuclear Information System (INIS)

    Zhao Liang; Tu Yu-Song; Xu Shun; Zhou Xin

    2017-01-01

    The square-well (SW) potential is one of the simplest pair potential models and its phase behavior has been clearly revealed, therefore it has become a benchmark for checking new theories or numerical methods. We introduce the generalized canonical ensemble (GCE) into the isobaric replica exchange Monte Carlo (REMC) algorithm to form a novel isobaric GCE-REMC method, and apply it to the study of vapor–liquid transition of SW particles. It is validated that this method can reproduce the vapor–liquid diagram of SW particles by comparing the estimated vapor–liquid binodals and the critical point with those from the literature. The notable advantage of this method is that the unstable vapor–liquid coexisting states, which cannot be detected using conventional sampling techniques, are accessed with a high sampling efficiency. Besides, the isobaric GCE-REMC method can visit all the possible states, including stable, metastable or unstable states during the phase transition over a wide pressure range, providing an effective pathway to understand complex phase transitions during the nucleation or crystallization process in physical or biological systems. (paper)

  2. Ensemble of different approaches for a reliable person re-identification system

    Directory of Open Access Journals (Sweden)

    Loris Nanni

    2016-07-01

    Full Text Available An ensemble of approaches for reliable person re-identification is proposed in this paper. The proposed ensemble is built combining widely used person re-identification systems using different color spaces and some variants of state-of-the-art approaches that are proposed in this paper. Different descriptors are tested, and both texture and color features are extracted from the images; then the different descriptors are compared using different distance measures (e.g., the Euclidean distance, angle, and the Jeffrey distance. To improve performance, a method based on skeleton detection, extracted from the depth map, is also applied when the depth map is available. The proposed ensemble is validated on three widely used datasets (CAVIAR4REID, IAS, and VIPeR, keeping the same parameter set of each approach constant across all tests to avoid overfitting and to demonstrate that the proposed system can be considered a general-purpose person re-identification system. Our experimental results show that the proposed system offers significant improvements over baseline approaches. The source code used for the approaches tested in this paper will be available at https://www.dei.unipd.it/node/2357 and http://robotics.dei.unipd.it/reid/.

  3. Using support vector machine ensembles for target audience classification on Twitter.

    Science.gov (United States)

    Lo, Siaw Ling; Chiong, Raymond; Cornforth, David

    2015-01-01

    The vast amount and diversity of the content shared on social media can pose a challenge for any business wanting to use it to identify potential customers. In this paper, our aim is to investigate the use of both unsupervised and supervised learning methods for target audience classification on Twitter with minimal annotation efforts. Topic domains were automatically discovered from contents shared by followers of an account owner using Twitter Latent Dirichlet Allocation (LDA). A Support Vector Machine (SVM) ensemble was then trained using contents from different account owners of the various topic domains identified by Twitter LDA. Experimental results show that the methods presented are able to successfully identify a target audience with high accuracy. In addition, we show that using a statistical inference approach such as bootstrapping in over-sampling, instead of using random sampling, to construct training datasets can achieve a better classifier in an SVM ensemble. We conclude that such an ensemble system can take advantage of data diversity, which enables real-world applications for differentiating prospective customers from the general audience, leading to business advantage in the crowded social media space.

  4. Using support vector machine ensembles for target audience classification on Twitter.

    Directory of Open Access Journals (Sweden)

    Siaw Ling Lo

    Full Text Available The vast amount and diversity of the content shared on social media can pose a challenge for any business wanting to use it to identify potential customers. In this paper, our aim is to investigate the use of both unsupervised and supervised learning methods for target audience classification on Twitter with minimal annotation efforts. Topic domains were automatically discovered from contents shared by followers of an account owner using Twitter Latent Dirichlet Allocation (LDA. A Support Vector Machine (SVM ensemble was then trained using contents from different account owners of the various topic domains identified by Twitter LDA. Experimental results show that the methods presented are able to successfully identify a target audience with high accuracy. In addition, we show that using a statistical inference approach such as bootstrapping in over-sampling, instead of using random sampling, to construct training datasets can achieve a better classifier in an SVM ensemble. We conclude that such an ensemble system can take advantage of data diversity, which enables real-world applications for differentiating prospective customers from the general audience, leading to business advantage in the crowded social media space.

  5. Genetic Algorithm Optimized Neural Networks Ensemble as ...

    African Journals Online (AJOL)

    Marquardt algorithm by varying conditions such as inputs, hidden neurons, initialization, training sets and random Gaussian noise injection to ... Several such ensembles formed the population which was evolved to generate the fittest ensemble.

  6. Estimating Convection Parameters in the GFDL CM2.1 Model Using Ensemble Data Assimilation

    Science.gov (United States)

    Li, Shan; Zhang, Shaoqing; Liu, Zhengyu; Lu, Lv; Zhu, Jiang; Zhang, Xuefeng; Wu, Xinrong; Zhao, Ming; Vecchi, Gabriel A.; Zhang, Rong-Hua; Lin, Xiaopei

    2018-04-01

    Parametric uncertainty in convection parameterization is one major source of model errors that cause model climate drift. Convection parameter tuning has been widely studied in atmospheric models to help mitigate the problem. However, in a fully coupled general circulation model (CGCM), convection parameters which impact the ocean as well as the climate simulation may have different optimal values. This study explores the possibility of estimating convection parameters with an ensemble coupled data assimilation method in a CGCM. Impacts of the convection parameter estimation on climate analysis and forecast are analyzed. In a twin experiment framework, five convection parameters in the GFDL coupled model CM2.1 are estimated individually and simultaneously under both perfect and imperfect model regimes. Results show that the ensemble data assimilation method can help reduce the bias in convection parameters. With estimated convection parameters, the analyses and forecasts for both the atmosphere and the ocean are generally improved. It is also found that information in low latitudes is relatively more important for estimating convection parameters. This study further suggests that when important parameters in appropriate physical parameterizations are identified, incorporating their estimation into traditional ensemble data assimilation procedure could improve the final analysis and climate prediction.

  7. Global Ensemble Forecast System (GEFS) [1 Deg.

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The Global Ensemble Forecast System (GEFS) is a weather forecast model made up of 21 separate forecasts, or ensemble members. The National Centers for Environmental...

  8. Natural and drought scenarios in an east central Amazon forest: Fidelity of the Community Land Model 3.5 with three biogeochemical models

    Science.gov (United States)

    Sakaguchi, Koichi; Zeng, Xubin; Christoffersen, Bradley J.; Restrepo-Coupe, Natalia; Saleska, Scott R.; Brando, Paulo M.

    2011-03-01

    Recent development of general circulation models involves biogeochemical cycles: flows of carbon and other chemical species that circulate through the Earth system. Such models are valuable tools for future projections of climate, but still bear large uncertainties in the model simulations. One of the regions with especially high uncertainty is the Amazon forest where large-scale dieback associated with the changing climate is predicted by several models. In order to better understand the capability and weakness of global-scale land-biogeochemical models in simulating a tropical ecosystem under the present day as well as significantly drier climates, we analyzed the off-line simulations for an east central Amazon forest by the Community Land Model version 3.5 of the National Center for Atmospheric Research and its three independent biogeochemical submodels (CASA', CN, and DGVM). Intense field measurements carried out under Large Scale Biosphere-Atmosphere Experiment in Amazonia, including forest response to drought from a throughfall exclusion experiment, are utilized to evaluate the whole spectrum of biogeophysical and biogeochemical aspects of the models. Our analysis shows reasonable correspondence in momentum and energy turbulent fluxes, but it highlights three processes that are not in agreement with observations: (1) inconsistent seasonality in carbon fluxes, (2) biased biomass size and allocation, and (3) overestimation of vegetation stress to short-term drought but underestimation of biomass loss from long-term drought. Without resolving these issues the modeled feedbacks from the biosphere in future climate projections would be questionable. We suggest possible directions for model improvements and also emphasize the necessity of more studies using a variety of in situ data for both driving and evaluating land-biogeochemical models.

  9. Bidirectional Modulation of Intrinsic Excitability in Rat Prelimbic Cortex Neuronal Ensembles and Non-Ensembles after Operant Learning.

    Science.gov (United States)

    Whitaker, Leslie R; Warren, Brandon L; Venniro, Marco; Harte, Tyler C; McPherson, Kylie B; Beidel, Jennifer; Bossert, Jennifer M; Shaham, Yavin; Bonci, Antonello; Hope, Bruce T

    2017-09-06

    Learned associations between environmental stimuli and rewards drive goal-directed learning and motivated behavior. These memories are thought to be encoded by alterations within specific patterns of sparsely distributed neurons called neuronal ensembles that are activated selectively by reward-predictive stimuli. Here, we use the Fos promoter to identify strongly activated neuronal ensembles in rat prelimbic cortex (PLC) and assess altered intrinsic excitability after 10 d of operant food self-administration training (1 h/d). First, we used the Daun02 inactivation procedure in male FosLacZ-transgenic rats to ablate selectively Fos-expressing PLC neurons that were active during operant food self-administration. Selective ablation of these neurons decreased food seeking. We then used male FosGFP-transgenic rats to assess selective alterations of intrinsic excitability in Fos-expressing neuronal ensembles (FosGFP + ) that were activated during food self-administration and compared these with alterations in less activated non-ensemble neurons (FosGFP - ). Using whole-cell recordings of layer V pyramidal neurons in an ex vivo brain slice preparation, we found that operant self-administration increased excitability of FosGFP + neurons and decreased excitability of FosGFP - neurons. Increased excitability of FosGFP + neurons was driven by increased steady-state input resistance. Decreased excitability of FosGFP - neurons was driven by increased contribution of small-conductance calcium-activated potassium (SK) channels. Injections of the specific SK channel antagonist apamin into PLC increased Fos expression but had no effect on food seeking. Overall, operant learning increased intrinsic excitability of PLC Fos-expressing neuronal ensembles that play a role in food seeking but decreased intrinsic excitability of Fos - non-ensembles. SIGNIFICANCE STATEMENT Prefrontal cortex activity plays a critical role in operant learning, but the underlying cellular mechanisms are

  10. Assessment of the GHG Reduction Potential from Energy Crops Using a Combined LCA and Biogeochemical Process Models: A Review

    Directory of Open Access Journals (Sweden)

    Dong Jiang

    2014-01-01

    Full Text Available The main purpose for developing biofuel is to reduce GHG (greenhouse gas emissions, but the comprehensive environmental impact of such fuels is not clear. Life cycle analysis (LCA, as a complete comprehensive analysis method, has been widely used in bioenergy assessment studies. Great efforts have been directed toward establishing an efficient method for comprehensively estimating the greenhouse gas (GHG emission reduction potential from the large-scale cultivation of energy plants by combining LCA with ecosystem/biogeochemical process models. LCA presents a general framework for evaluating the energy consumption and GHG emission from energy crop planting, yield acquisition, production, product use, and postprocessing. Meanwhile, ecosystem/biogeochemical process models are adopted to simulate the fluxes and storage of energy, water, carbon, and nitrogen in the soil-plant (energy crops soil continuum. Although clear progress has been made in recent years, some problems still exist in current studies and should be addressed. This paper reviews the state-of-the-art method for estimating GHG emission reduction through developing energy crops and introduces in detail a new approach for assessing GHG emission reduction by combining LCA with biogeochemical process models. The main achievements of this study along with the problems in current studies are described and discussed.

  11. Multiscale Investigation on Biofilm Distribution and Its Impact on Macroscopic Biogeochemical Reaction Rates

    Science.gov (United States)

    Yan, Zhifeng; Liu, Chongxuan; Liu, Yuanyuan; Bailey, Vanessa L.

    2017-11-01

    Biofilms are critical locations for biogeochemical reactions in the subsurface environment. The occurrence and distribution of biofilms at microscale as well as their impacts on macroscopic biogeochemical reaction rates are still poorly understood. This paper investigated the formation and distributions of biofilms in heterogeneous sediments using multiscale models and evaluated the effects of biofilm heterogeneity on local and macroscopic biogeochemical reaction rates. Sediment pore structures derived from X-ray computed tomography were used to simulate the microscale flow dynamics and biofilm distribution in the sediment column. The response of biofilm formation and distribution to the variations in hydraulic and chemical properties was first examined. One representative biofilm distribution was then utilized to evaluate its effects on macroscopic reaction rates using nitrate reduction as an example. The results revealed that microorganisms primarily grew on the surfaces of grains and aggregates near preferential flow paths where both electron donor and acceptor were readily accessible, leading to the heterogeneous distribution of biofilms in the sediments. The heterogeneous biofilm distribution decreased the macroscopic rate of biogeochemical reactions as compared with those in homogeneous cases. Operationally considering the heterogeneous biofilm distribution in macroscopic reactive transport models such as using dual porosity domain concept can significantly improve the prediction of biogeochemical reaction rates. Overall, this study provided important insights into the biofilm formation and distribution in soils and sediments as well as their impacts on the macroscopic manifestation of reaction rates.

  12. A MITgcm/DART ensemble analysis and prediction system with application to the Gulf of Mexico

    KAUST Repository

    Hoteit, Ibrahim; Hoar, Timothy J.; Gopalakrishnan, Ganesh; Collins, Nancy S.; Anderson, Jeffrey L.; Cornuelle, Bruce D.; Kö hl, Armin; Heimbach, Patrick

    2013-01-01

    Research Testbed (DART) assimilation package with the Massachusetts Institute of Technology ocean general circulation model (MITgcm). The MITgcm/DART system supports the assimilation of a wide range of ocean observations and uses an ensemble approach

  13. A second-order unconstrained optimization method for canonical-ensemble density-functional methods

    Science.gov (United States)

    Nygaard, Cecilie R.; Olsen, Jeppe

    2013-03-01

    A second order converging method of ensemble optimization (SOEO) in the framework of Kohn-Sham Density-Functional Theory is presented, where the energy is minimized with respect to an ensemble density matrix. It is general in the sense that the number of fractionally occupied orbitals is not predefined, but rather it is optimized by the algorithm. SOEO is a second order Newton-Raphson method of optimization, where both the form of the orbitals and the occupation numbers are optimized simultaneously. To keep the occupation numbers between zero and two, a set of occupation angles is defined, from which the occupation numbers are expressed as trigonometric functions. The total number of electrons is controlled by a built-in second order restriction of the Newton-Raphson equations, which can be deactivated in the case of a grand-canonical ensemble (where the total number of electrons is allowed to change). To test the optimization method, dissociation curves for diatomic carbon are produced using different functionals for the exchange-correlation energy. These curves show that SOEO favors symmetry broken pure-state solutions when using functionals with exact exchange such as Hartree-Fock and Becke three-parameter Lee-Yang-Parr. This is explained by an unphysical contribution to the exact exchange energy from interactions between fractional occupations. For functionals without exact exchange, such as local density approximation or Becke Lee-Yang-Parr, ensemble solutions are favored at interatomic distances larger than the equilibrium distance. Calculations on the chromium dimer are also discussed. They show that SOEO is able to converge to ensemble solutions for systems that are more complicated than diatomic carbon.

  14. Derivation of Mayer Series from Canonical Ensemble

    International Nuclear Information System (INIS)

    Wang Xian-Zhi

    2016-01-01

    Mayer derived the Mayer series from both the canonical ensemble and the grand canonical ensemble by use of the cluster expansion method. In 2002, we conjectured a recursion formula of the canonical partition function of a fluid (X.Z. Wang, Phys. Rev. E 66 (2002) 056102). In this paper we give a proof for this formula by developing an appropriate expansion of the integrand of the canonical partition function. We further derive the Mayer series solely from the canonical ensemble by use of this recursion formula. (paper)

  15. Derivation of Mayer Series from Canonical Ensemble

    Science.gov (United States)

    Wang, Xian-Zhi

    2016-02-01

    Mayer derived the Mayer series from both the canonical ensemble and the grand canonical ensemble by use of the cluster expansion method. In 2002, we conjectured a recursion formula of the canonical partition function of a fluid (X.Z. Wang, Phys. Rev. E 66 (2002) 056102). In this paper we give a proof for this formula by developing an appropriate expansion of the integrand of the canonical partition function. We further derive the Mayer series solely from the canonical ensemble by use of this recursion formula.

  16. Large- to submesoscale surface circulation and its implications on biogeochemical/biological horizontal distributions during the OUTPACE cruise (southwest Pacific)

    Science.gov (United States)

    Rousselet, Louise; de Verneil, Alain; Doglioli, Andrea M.; Petrenko, Anne A.; Duhamel, Solange; Maes, Christophe; Blanke, Bruno

    2018-04-01

    The patterns of the large-scale, meso- and submesoscale surface circulation on biogeochemical and biological distributions are examined in the western tropical South Pacific (WTSP) in the context of the OUTPACE cruise (February-April 2015). Multi-disciplinary original in situ observations were achieved along a zonal transect through the WTSP and their analysis was coupled with satellite data. The use of Lagrangian diagnostics allows for the identification of water mass pathways, mesoscale structures, and submesoscale features such as fronts. In particular, we confirmed the existence of a global wind-driven southward circulation of surface waters in the entire WTSP, using a new high-resolution altimetry-derived product, validated by in situ drifters, that includes cyclogeostrophy and Ekman components with geostrophy. The mesoscale activity is shown to be responsible for counter-intuitive water mass trajectories in two subregions: (i) the Coral Sea, with surface exchanges between the North Vanuatu Jet and the North Caledonian Jet, and (ii) around 170° W, with an eastward pathway, whereas a westward general direction dominates. Fronts and small-scale features, detected with finite-size Lyapunov exponents (FSLEs), are correlated with 25 % of surface tracer gradients, which reveals the significance of such structures in the generation of submesoscale surface gradients. Additionally, two high-frequency sampling transects of biogeochemical parameters and microorganism abundances demonstrate the influence of fronts in controlling the spatial distribution of bacteria and phytoplankton, and as a consequence the microbial community structure. All circulation scales play an important role that has to be taken into account not only when analysing the data from OUTPACE but also, more generally, for understanding the global distribution of biogeochemical components.

  17. Biogeochemical Processes Regulating the Mobility of Uranium in Sediments

    Energy Technology Data Exchange (ETDEWEB)

    Belli, Keaton M.; Taillefert, Martial

    2016-07-01

    This book chapters reviews the latest knowledge on the biogeochemical processes regulating the mobility of uranium in sediments. It contains both data from the literature and new data from the authors.

  18. Regionalization of post-processed ensemble runoff forecasts

    Directory of Open Access Journals (Sweden)

    J. O. Skøien

    2016-05-01

    Full Text Available For many years, meteorological models have been run with perturbated initial conditions or parameters to produce ensemble forecasts that are used as a proxy of the uncertainty of the forecasts. However, the ensembles are usually both biased (the mean is systematically too high or too low, compared with the observed weather, and has dispersion errors (the ensemble variance indicates a too low or too high confidence in the forecast, compared with the observed weather. The ensembles are therefore commonly post-processed to correct for these shortcomings. Here we look at one of these techniques, referred to as Ensemble Model Output Statistics (EMOS (Gneiting et al., 2005. Originally, the post-processing parameters were identified as a fixed set of parameters for a region. The application of our work is the European Flood Awareness System (http://www.efas.eu, where a distributed model is run with meteorological ensembles as input. We are therefore dealing with a considerably larger data set than previous analyses. We also want to regionalize the parameters themselves for other locations than the calibration gauges. The post-processing parameters are therefore estimated for each calibration station, but with a spatial penalty for deviations from neighbouring stations, depending on the expected semivariance between the calibration catchment and these stations. The estimated post-processed parameters can then be used for regionalization of the postprocessing parameters also for uncalibrated locations using top-kriging in the rtop-package (Skøien et al., 2006, 2014. We will show results from cross-validation of the methodology and although our interest is mainly in identifying exceedance probabilities for certain return levels, we will also show how the rtop package can be used for creating a set of post-processed ensembles through simulations.

  19. Pauci ex tanto numero: reducing redundancy in multi-model ensembles

    Science.gov (United States)

    Solazzo, E.; Riccio, A.; Kioutsioukis, I.; Galmarini, S.

    2013-02-01

    We explicitly address the fundamental issue of member diversity in multi-model ensembles. To date no attempts in this direction are documented within the air quality (AQ) community, although the extensive use of ensembles in this field. Common biases and redundancy are the two issues directly deriving from lack of independence, undermining the significance of a multi-model ensemble, and are the subject of this study. Shared biases among models will determine a biased ensemble, making therefore essential the errors of the ensemble members to be independent so that bias can cancel out. Redundancy derives from having too large a portion of common variance among the members of the ensemble, producing overconfidence in the predictions and underestimation of the uncertainty. The two issues of common biases and redundancy are analysed in detail using the AQMEII ensemble of AQ model results for four air pollutants in two European regions. We show that models share large portions of bias and variance, extending well beyond those induced by common inputs. We make use of several techniques to further show that subsets of models can explain the same amount of variance as the full ensemble with the advantage of being poorly correlated. Selecting the members for generating skilful, non-redundant ensembles from such subsets proved, however, non-trivial. We propose and discuss various methods of member selection and rate the ensemble performance they produce. In most cases, the full ensemble is outscored by the reduced ones. We conclude that, although independence of outputs may not always guarantee enhancement of scores (but this depends upon the skill being investigated) we discourage selecting the members of the ensemble simply on the basis of scores, that is, independence and skills need to be considered disjointly.

  20. Engineering Pseudomonas stutzeri as a biogeochemical biosensor

    Science.gov (United States)

    Boynton, L.; Cheng, H. Y.; Del Valle, I.; Masiello, C. A.; Silberg, J. J.

    2016-12-01

    Biogeochemical cycles are being drastically altered as a result of anthropogenic activities, such as the burning of fossil fuels and the industrial production of ammonia. We know microbes play a major part in these cycles, but the extent of their biogeochemical roles remains largely uncharacterized due to inadequacies with culturing and measurement. While metagenomics and other -omics methods offer ways to reconstruct microbial communities, these approaches can only give an indication of the functional roles of microbes in a community. These -omics approaches are rapidly being expanded to the point of outpacing our knowledge of functional genes, which highlights an inherent need for analytical methods that non-invasively monitor Earth's processes in real time. Here we aim to exploit synthetic biology methods in order to engineer a ubiquitous denitrifying microbe, Pseudomonas stutzeri that can act as a biosensor in soil and marine environments. By using an easily cultivated microbe that is also common in many environments, we hope to develop a tool that allows us to zoom in on specific aspects of the nitrogen cycle. In order to monitor processes occurring at the genetic level in environments that cannot be resolved with fluorescence-based methods, such as soils, we have developed a system that instead relies on gas production by engineered microbial biosensors. P. stutzeri has been successfully engineered to release a gas, methyl bromide, which can continuously and non-invasively be measured by GC-MS. Similar to using Green Fluorescent Protein, GFP, in the biological sciences, the gene controlling gas production can be linked to those involved in denitrification, thereby creating a quantifiable gas signal that is correlated with microbial activity in the soil. Synthetically engineered microbial biosensors could reveal key aspects of metabolism in soil systems and offer a tool for characterizing the scope and degree of microbial impact on major biogeochemical cycles.

  1. The Southern Ocean biogeochemical divide.

    Science.gov (United States)

    Marinov, I; Gnanadesikan, A; Toggweiler, J R; Sarmiento, J L

    2006-06-22

    Modelling studies have demonstrated that the nutrient and carbon cycles in the Southern Ocean play a central role in setting the air-sea balance of CO(2) and global biological production. Box model studies first pointed out that an increase in nutrient utilization in the high latitudes results in a strong decrease in the atmospheric carbon dioxide partial pressure (pCO2). This early research led to two important ideas: high latitude regions are more important in determining atmospheric pCO2 than low latitudes, despite their much smaller area, and nutrient utilization and atmospheric pCO2 are tightly linked. Subsequent general circulation model simulations show that the Southern Ocean is the most important high latitude region in controlling pre-industrial atmospheric CO(2) because it serves as a lid to a larger volume of the deep ocean. Other studies point out the crucial role of the Southern Ocean in the uptake and storage of anthropogenic carbon dioxide and in controlling global biological production. Here we probe the system to determine whether certain regions of the Southern Ocean are more critical than others for air-sea CO(2) balance and the biological export production, by increasing surface nutrient drawdown in an ocean general circulation model. We demonstrate that atmospheric CO(2) and global biological export production are controlled by different regions of the Southern Ocean. The air-sea balance of carbon dioxide is controlled mainly by the biological pump and circulation in the Antarctic deep-water formation region, whereas global export production is controlled mainly by the biological pump and circulation in the Subantarctic intermediate and mode water formation region. The existence of this biogeochemical divide separating the Antarctic from the Subantarctic suggests that it may be possible for climate change or human intervention to modify one of these without greatly altering the other.

  2. A Comparison of Ensemble Kalman Filters for Storm Surge Assimilation

    KAUST Repository

    Altaf, Muhammad

    2014-08-01

    This study evaluates and compares the performances of several variants of the popular ensembleKalman filter for the assimilation of storm surge data with the advanced circulation (ADCIRC) model. Using meteorological data from Hurricane Ike to force the ADCIRC model on a domain including the Gulf ofMexico coastline, the authors implement and compare the standard stochastic ensembleKalman filter (EnKF) and three deterministic square root EnKFs: the singular evolutive interpolated Kalman (SEIK) filter, the ensemble transform Kalman filter (ETKF), and the ensemble adjustment Kalman filter (EAKF). Covariance inflation and localization are implemented in all of these filters. The results from twin experiments suggest that the square root ensemble filters could lead to very comparable performances with appropriate tuning of inflation and localization, suggesting that practical implementation details are at least as important as the choice of the square root ensemble filter itself. These filters also perform reasonably well with a relatively small ensemble size, whereas the stochastic EnKF requires larger ensemble sizes to provide similar accuracy for forecasts of storm surge.

  3. A Comparison of Ensemble Kalman Filters for Storm Surge Assimilation

    KAUST Repository

    Altaf, Muhammad; Butler, T.; Mayo, T.; Luo, X.; Dawson, C.; Heemink, A. W.; Hoteit, Ibrahim

    2014-01-01

    This study evaluates and compares the performances of several variants of the popular ensembleKalman filter for the assimilation of storm surge data with the advanced circulation (ADCIRC) model. Using meteorological data from Hurricane Ike to force the ADCIRC model on a domain including the Gulf ofMexico coastline, the authors implement and compare the standard stochastic ensembleKalman filter (EnKF) and three deterministic square root EnKFs: the singular evolutive interpolated Kalman (SEIK) filter, the ensemble transform Kalman filter (ETKF), and the ensemble adjustment Kalman filter (EAKF). Covariance inflation and localization are implemented in all of these filters. The results from twin experiments suggest that the square root ensemble filters could lead to very comparable performances with appropriate tuning of inflation and localization, suggesting that practical implementation details are at least as important as the choice of the square root ensemble filter itself. These filters also perform reasonably well with a relatively small ensemble size, whereas the stochastic EnKF requires larger ensemble sizes to provide similar accuracy for forecasts of storm surge.

  4. HIGH-RESOLUTION ATMOSPHERIC ENSEMBLE MODELING AT SRNL

    Energy Technology Data Exchange (ETDEWEB)

    Buckley, R.; Werth, D.; Chiswell, S.; Etherton, B.

    2011-05-10

    The High-Resolution Mid-Atlantic Forecasting Ensemble (HME) is a federated effort to improve operational forecasts related to precipitation, convection and boundary layer evolution, and fire weather utilizing data and computing resources from a diverse group of cooperating institutions in order to create a mesoscale ensemble from independent members. Collaborating organizations involved in the project include universities, National Weather Service offices, and national laboratories, including the Savannah River National Laboratory (SRNL). The ensemble system is produced from an overlapping numerical weather prediction model domain and parameter subsets provided by each contributing member. The coordination, synthesis, and dissemination of the ensemble information are performed by the Renaissance Computing Institute (RENCI) at the University of North Carolina-Chapel Hill. This paper discusses background related to the HME effort, SRNL participation, and example results available from the RENCI website.

  5. Comparison of projection skills of deterministic ensemble methods using pseudo-simulation data generated from multivariate Gaussian distribution

    Science.gov (United States)

    Oh, Seok-Geun; Suh, Myoung-Seok

    2017-07-01

    The projection skills of five ensemble methods were analyzed according to simulation skills, training period, and ensemble members, using 198 sets of pseudo-simulation data (PSD) produced by random number generation assuming the simulated temperature of regional climate models. The PSD sets were classified into 18 categories according to the relative magnitude of bias, variance ratio, and correlation coefficient, where each category had 11 sets (including 1 truth set) with 50 samples. The ensemble methods used were as follows: equal weighted averaging without bias correction (EWA_NBC), EWA with bias correction (EWA_WBC), weighted ensemble averaging based on root mean square errors and correlation (WEA_RAC), WEA based on the Taylor score (WEA_Tay), and multivariate linear regression (Mul_Reg). The projection skills of the ensemble methods improved generally as compared with the best member for each category. However, their projection skills are significantly affected by the simulation skills of the ensemble member. The weighted ensemble methods showed better projection skills than non-weighted methods, in particular, for the PSD categories having systematic biases and various correlation coefficients. The EWA_NBC showed considerably lower projection skills than the other methods, in particular, for the PSD categories with systematic biases. Although Mul_Reg showed relatively good skills, it showed strong sensitivity to the PSD categories, training periods, and number of members. On the other hand, the WEA_Tay and WEA_RAC showed relatively superior skills in both the accuracy and reliability for all the sensitivity experiments. This indicates that WEA_Tay and WEA_RAC are applicable even for simulation data with systematic biases, a short training period, and a small number of ensemble members.

  6. The Hydrologic Ensemble Prediction Experiment (HEPEX)

    Science.gov (United States)

    Wood, A. W.; Thielen, J.; Pappenberger, F.; Schaake, J. C.; Hartman, R. K.

    2012-12-01

    The Hydrologic Ensemble Prediction Experiment was established in March, 2004, at a workshop hosted by the European Center for Medium Range Weather Forecasting (ECMWF). With support from the US National Weather Service (NWS) and the European Commission (EC), the HEPEX goal was to bring the international hydrological and meteorological communities together to advance the understanding and adoption of hydrological ensemble forecasts for decision support in emergency management and water resources sectors. The strategy to meet this goal includes meetings that connect the user, forecast producer and research communities to exchange ideas, data and methods; the coordination of experiments to address specific challenges; and the formation of testbeds to facilitate shared experimentation. HEPEX has organized about a dozen international workshops, as well as sessions at scientific meetings (including AMS, AGU and EGU) and special issues of scientific journals where workshop results have been published. Today, the HEPEX mission is to demonstrate the added value of hydrological ensemble prediction systems (HEPS) for emergency management and water resources sectors to make decisions that have important consequences for economy, public health, safety, and the environment. HEPEX is now organised around six major themes that represent core elements of a hydrologic ensemble prediction enterprise: input and pre-processing, ensemble techniques, data assimilation, post-processing, verification, and communication and use in decision making. This poster presents an overview of recent and planned HEPEX activities, highlighting case studies that exemplify the focus and objectives of HEPEX.

  7. Ensemble support vector machine classification of dementia using structural MRI and mini-mental state examination.

    Science.gov (United States)

    Sørensen, Lauge; Nielsen, Mads

    2018-05-15

    The International Challenge for Automated Prediction of MCI from MRI data offered independent, standardized comparison of machine learning algorithms for multi-class classification of normal control (NC), mild cognitive impairment (MCI), converting MCI (cMCI), and Alzheimer's disease (AD) using brain imaging and general cognition. We proposed to use an ensemble of support vector machines (SVMs) that combined bagging without replacement and feature selection. SVM is the most commonly used algorithm in multivariate classification of dementia, and it was therefore valuable to evaluate the potential benefit of ensembling this type of classifier. The ensemble SVM, using either a linear or a radial basis function (RBF) kernel, achieved multi-class classification accuracies of 55.6% and 55.0% in the challenge test set (60 NC, 60 MCI, 60 cMCI, 60 AD), resulting in a third place in the challenge. Similar feature subset sizes were obtained for both kernels, and the most frequently selected MRI features were the volumes of the two hippocampal subregions left presubiculum and right subiculum. Post-challenge analysis revealed that enforcing a minimum number of selected features and increasing the number of ensemble classifiers improved classification accuracy up to 59.1%. The ensemble SVM outperformed single SVM classifications consistently in the challenge test set. Ensemble methods using bagging and feature selection can improve the performance of the commonly applied SVM classifier in dementia classification. This resulted in competitive classification accuracies in the International Challenge for Automated Prediction of MCI from MRI data. Copyright © 2018 Elsevier B.V. All rights reserved.

  8. Arctic sea ice area changes in CMIP3 and CMIP5 climate models’ ensembles

    Directory of Open Access Journals (Sweden)

    V. A. Semenov

    2017-01-01

    Full Text Available The shrinking Arctic sea ice cover observed during the last decades is probably the clearest manifestation of ongoing climate change. While climate models in general reproduce the sea ice retreat in the Arctic during the 20th century and simulate further sea ice area loss during the 21st century in response to anthropogenic forcing, the models suffer from large biases and the results exhibit considerable spread. Here, we compare results from the two last generations of climate models, CMIP3 and CMIP5, with respect to total and regional Arctic sea ice change. Different characteristics of sea ice area (SIA in March and September have been analysed for the Entire Arctic, Central Arctic and Barents Sea. Further, the sensitivity of SIA to changes in Northern Hemisphere (NH temperature is investigated and dynamical links between SIA and some atmospheric variability modes are assessed.CMIP3 (SRES A1B and CMIP5 (RCP8.5 models not only simulate a coherent decline of the Arctic SIA but also depict consistent changes in the SIA seasonal cycle. The spatial patterns of SIC variability improve in CMIP5 ensemble, most noticeably in summer when compared to HadISST1 data. A better simulation of summer SIA in the Entire Arctic by CMIP5 models is accompanied by a slightly increased bias for winter season in comparison to CMIP3 ensemble. SIA in the Barents Sea is strongly overestimated by the majority of CMIP3 and CMIP5 models, and projected SIA changes are characterized by a high uncertainty. Both CMIP ensembles depict a significant link between the SIA and NH temperature changes indicating that a part of inter-ensemble SIA spread comes from different temperature sensitivity to anthropogenic forcing. The results suggest that, in general, a sensitivity of SIA to external forcing is enhanced in CMIP5 models. Arctic SIA interannual variability in the end of the 20th century is on average well simulated by both ensembles. To the end of the 21st century, September

  9. Pauci ex tanto numero: reduce redundancy in multi-model ensembles

    Science.gov (United States)

    Solazzo, E.; Riccio, A.; Kioutsioukis, I.; Galmarini, S.

    2013-08-01

    We explicitly address the fundamental issue of member diversity in multi-model ensembles. To date, no attempts in this direction have been documented within the air quality (AQ) community despite the extensive use of ensembles in this field. Common biases and redundancy are the two issues directly deriving from lack of independence, undermining the significance of a multi-model ensemble, and are the subject of this study. Shared, dependant biases among models do not cancel out but will instead determine a biased ensemble. Redundancy derives from having too large a portion of common variance among the members of the ensemble, producing overconfidence in the predictions and underestimation of the uncertainty. The two issues of common biases and redundancy are analysed in detail using the AQMEII ensemble of AQ model results for four air pollutants in two European regions. We show that models share large portions of bias and variance, extending well beyond those induced by common inputs. We make use of several techniques to further show that subsets of models can explain the same amount of variance as the full ensemble with the advantage of being poorly correlated. Selecting the members for generating skilful, non-redundant ensembles from such subsets proved, however, non-trivial. We propose and discuss various methods of member selection and rate the ensemble performance they produce. In most cases, the full ensemble is outscored by the reduced ones. We conclude that, although independence of outputs may not always guarantee enhancement of scores (but this depends upon the skill being investigated), we discourage selecting the members of the ensemble simply on the basis of scores; that is, independence and skills need to be considered disjointly.

  10. Data assimilation in integrated hydrological modeling using ensemble Kalman filtering

    DEFF Research Database (Denmark)

    Rasmussen, Jørn; Madsen, H.; Jensen, Karsten Høgh

    2015-01-01

    Groundwater head and stream discharge is assimilated using the ensemble transform Kalman filter in an integrated hydrological model with the aim of studying the relationship between the filter performance and the ensemble size. In an attempt to reduce the required number of ensemble members...... and estimating parameters requires a much larger ensemble size than just assimilating groundwater head observations. However, the required ensemble size can be greatly reduced with the use of adaptive localization, which by far outperforms distance-based localization. The study is conducted using synthetic data...

  11. Dynamic principle for ensemble control tools.

    Science.gov (United States)

    Samoletov, A; Vasiev, B

    2017-11-28

    Dynamical equations describing physical systems in contact with a thermal bath are commonly extended by mathematical tools called "thermostats." These tools are designed for sampling ensembles in statistical mechanics. Here we propose a dynamic principle underlying a range of thermostats which is derived using fundamental laws of statistical physics and ensures invariance of the canonical measure. The principle covers both stochastic and deterministic thermostat schemes. Our method has a clear advantage over a range of proposed and widely used thermostat schemes that are based on formal mathematical reasoning. Following the derivation of the proposed principle, we show its generality and illustrate its applications including design of temperature control tools that differ from the Nosé-Hoover-Langevin scheme.

  12. Ensemble-based Probabilistic Forecasting at Horns Rev

    DEFF Research Database (Denmark)

    Pinson, Pierre; Madsen, Henrik

    2009-01-01

    forecasting methodology. In a first stage, ensemble forecasts of meteorological variables are converted to power through a suitable power curve model. This modelemploys local polynomial regression, and is adoptively estimated with an orthogonal fitting method. The obtained ensemble forecasts of wind power...

  13. DroidEnsemble: Detecting Android Malicious Applications with Ensemble of String and Structural Static Features

    KAUST Repository

    Wang, Wei

    2018-05-11

    Android platform has dominated the Operating System of mobile devices. However, the dramatic increase of Android malicious applications (malapps) has caused serious software failures to Android system and posed a great threat to users. The effective detection of Android malapps has thus become an emerging yet crucial issue. Characterizing the behaviors of Android applications (apps) is essential to detecting malapps. Most existing work on detecting Android malapps was mainly based on string static features such as permissions and API usage extracted from apps. There also exists work on the detection of Android malapps with structural features, such as Control Flow Graph (CFG) and Data Flow Graph (DFG). As Android malapps have become increasingly polymorphic and sophisticated, using only one type of static features may result in false negatives. In this work, we propose DroidEnsemble that takes advantages of both string features and structural features to systematically and comprehensively characterize the static behaviors of Android apps and thus build a more accurate detection model for the detection of Android malapps. We extract each app’s string features, including permissions, hardware features, filter intents, restricted API calls, used permissions, code patterns, as well as structural features like function call graph. We then use three machine learning algorithms, namely, Support Vector Machine (SVM), k-Nearest Neighbor (kNN) and Random Forest (RF), to evaluate the performance of these two types of features and of their ensemble. In the experiments, We evaluate our methods and models with 1386 benign apps and 1296 malapps. Extensive experimental results demonstrate the effectiveness of DroidEnsemble. It achieves the detection accuracy as 95.8% with only string features and as 90.68% with only structural features. DroidEnsemble reaches the detection accuracy as 98.4% with the ensemble of both types of features, reducing 9 false positives and 12 false

  14. Hydrological and associated biogeochemical consequences of rapid global warming during the Paleocene-Eocene Thermal Maximum

    Science.gov (United States)

    Carmichael, Matthew J.; Inglis, Gordon N.; Badger, Marcus P. S.; Naafs, B. David A.; Behrooz, Leila; Remmelzwaal, Serginio; Monteiro, Fanny M.; Rohrssen, Megan; Farnsworth, Alexander; Buss, Heather L.; Dickson, Alexander J.; Valdes, Paul J.; Lunt, Daniel J.; Pancost, Richard D.

    2017-10-01

    The Paleocene-Eocene Thermal Maximum (PETM) hyperthermal, 56 million years ago (Ma), is the most dramatic example of abrupt Cenozoic global warming. During the PETM surface temperatures increased between 5 and 9 °C and the onset likely took hydrological and associated biogeochemical feedbacks, and proxy data from the PETM can provide constraints on changes in warm climate hydrology simulated by general circulation models (GCMs). In this paper, we provide a critical review of biological and geochemical signatures interpreted as direct or indirect indicators of hydrological change at the PETM, explore the importance of adopting multi-proxy approaches, and present a preliminary model-data comparison. Hydrological records complement those of temperature and indicate that the climatic response at the PETM was complex, with significant regional and temporal variability. This is further illustrated by the biogeochemical consequences of inferred changes in hydrology and, in fact, changes in precipitation and the biogeochemical consequences are often conflated in geochemical signatures. There is also strong evidence in many regions for changes in the episodic and/or intra-annual distribution of precipitation that has not widely been considered when comparing proxy data to GCM output. Crucially, GCM simulations indicate that the response of the hydrological cycle to the PETM was heterogeneous - some regions are associated with increased precipitation - evaporation (P - E), whilst others are characterised by a decrease. Interestingly, the majority of proxy data come from the regions where GCMs predict an increase in PETM precipitation. We propose that comparison of hydrological proxies to GCM output can be an important test of model skill, but this will be enhanced by further data from regions of model-simulated aridity and simulation of extreme precipitation events.

  15. A probabilistic parametrization for geological uncertainty estimation using the ensemble Kalman filter (EnKF)

    NARCIS (Netherlands)

    Sebacher, B.; Hanea, R.G.; Heemink, A.

    2013-01-01

    In the past years, many applications of historymatching methods in general and ensemble Kalman filter in particular have been proposed, especially in order to estimate fields that provide uncertainty in the stochastic process defined by the dynamical system of hydrocarbon recovery. Such fields can

  16. The significance of biogeochemical cycles of macro- and microelements in connection with man-made evolution of the living matter

    International Nuclear Information System (INIS)

    Ermakov, V.V.

    2008-01-01

    Biogeochemistry as an integrated science studying the elemental composition of the living matter and its role in migration, transformation, accumulation of chemical elements and their compounds in the biosphere, has again become the leading scientific branch highlighting the man-made evolution of the planet and the pathways of interaction between the man and environment. Nowadays the central problem of biogeochemistry as science about the biosphere is that of pollution of the different taxons of the biosphere. In the most case man-made factors effect on the different organisms and the flow of chemical elements changing their local, regional and global biogeochemical cycles. The concept of balance of O 2 , CO 2 and H 2 O as general condition of the sustained development of the biosphere is considered. The questions of biological rhythms, appearance of microelementhoses and modern systemic biogeochemical methodology of assessment of taxons of the biosphere are considered too

  17. Radial electric field and rotation of the ensemble of plasma particles in tokamak

    International Nuclear Information System (INIS)

    Sorokina, E. A.; Ilgisonis, V. I.

    2012-01-01

    The velocity of macroscopic rotation of an ensemble of charged particles in a tokamak in the presence of an electric field has been calculated in a collisionless approximation. It is shown that the velocity of toroidal rotation does not reduce to a local velocity of electric drift and has opposite directions on the inner and outer sides of the torus. This result is supplemented by an analysis of the trajectories of motion of individual particles in the ensemble, which shows that the passing and trapped particles of the ensemble acquire in the electric field, on the average, different toroidal velocities. For the trapped particles, this velocity is equal to that of electric drift in the poloidal magnetic field, while the velocity of passing particles is significantly different. It is shown that, although the electric-field-induced shift of the boundaries between trapped and passing particles in the phase space depends on the particle mass and charge and is, in the general case, asymmetric, this does not lead to current generation.

  18. The Microbial Engines That Drive Earth’s Biogeochemical Cycles

    Science.gov (United States)

    Falkowski, Paul G.; Fenchel, Tom; Delong, Edward F.

    2008-05-01

    Virtually all nonequilibrium electron transfers on Earth are driven by a set of nanobiological machines composed largely of multimeric protein complexes associated with a small number of prosthetic groups. These machines evolved exclusively in microbes early in our planet’s history yet, despite their antiquity, are highly conserved. Hence, although there is enormous genetic diversity in nature, there remains a relatively stable set of core genes coding for the major redox reactions essential for life and biogeochemical cycles. These genes created and coevolved with biogeochemical cycles and were passed from microbe to microbe primarily by horizontal gene transfer. A major challenge in the coming decades is to understand how these machines evolved, how they work, and the processes that control their activity on both molecular and planetary scales.

  19. Representing Color Ensembles.

    Science.gov (United States)

    Chetverikov, Andrey; Campana, Gianluca; Kristjánsson, Árni

    2017-10-01

    Colors are rarely uniform, yet little is known about how people represent color distributions. We introduce a new method for studying color ensembles based on intertrial learning in visual search. Participants looked for an oddly colored diamond among diamonds with colors taken from either uniform or Gaussian color distributions. On test trials, the targets had various distances in feature space from the mean of the preceding distractor color distribution. Targets on test trials therefore served as probes into probabilistic representations of distractor colors. Test-trial response times revealed a striking similarity between the physical distribution of colors and their internal representations. The results demonstrate that the visual system represents color ensembles in a more detailed way than previously thought, coding not only mean and variance but, most surprisingly, the actual shape (uniform or Gaussian) of the distribution of colors in the environment.

  20. Efficient Kernel-Based Ensemble Gaussian Mixture Filtering

    KAUST Repository

    Liu, Bo

    2015-11-11

    We consider the Bayesian filtering problem for data assimilation following the kernel-based ensemble Gaussian-mixture filtering (EnGMF) approach introduced by Anderson and Anderson (1999). In this approach, the posterior distribution of the system state is propagated with the model using the ensemble Monte Carlo method, providing a forecast ensemble that is then used to construct a prior Gaussian-mixture (GM) based on the kernel density estimator. This results in two update steps: a Kalman filter (KF)-like update of the ensemble members and a particle filter (PF)-like update of the weights, followed by a resampling step to start a new forecast cycle. After formulating EnGMF for any observational operator, we analyze the influence of the bandwidth parameter of the kernel function on the covariance of the posterior distribution. We then focus on two aspects: i) the efficient implementation of EnGMF with (relatively) small ensembles, where we propose a new deterministic resampling strategy preserving the first two moments of the posterior GM to limit the sampling error; and ii) the analysis of the effect of the bandwidth parameter on contributions of KF and PF updates and on the weights variance. Numerical results using the Lorenz-96 model are presented to assess the behavior of EnGMF with deterministic resampling, study its sensitivity to different parameters and settings, and evaluate its performance against ensemble KFs. The proposed EnGMF approach with deterministic resampling suggests improved estimates in all tested scenarios, and is shown to require less localization and to be less sensitive to the choice of filtering parameters.

  1. Design ensemble machine learning model for breast cancer diagnosis.

    Science.gov (United States)

    Hsieh, Sheau-Ling; Hsieh, Sung-Huai; Cheng, Po-Hsun; Chen, Chi-Huang; Hsu, Kai-Ping; Lee, I-Shun; Wang, Zhenyu; Lai, Feipei

    2012-10-01

    In this paper, we classify the breast cancer of medical diagnostic data. Information gain has been adapted for feature selections. Neural fuzzy (NF), k-nearest neighbor (KNN), quadratic classifier (QC), each single model scheme as well as their associated, ensemble ones have been developed for classifications. In addition, a combined ensemble model with these three schemes has been constructed for further validations. The experimental results indicate that the ensemble learning performs better than individual single ones. Moreover, the combined ensemble model illustrates the highest accuracy of classifications for the breast cancer among all models.

  2. Ensemble manifold regularization.

    Science.gov (United States)

    Geng, Bo; Tao, Dacheng; Xu, Chao; Yang, Linjun; Hua, Xian-Sheng

    2012-06-01

    We propose an automatic approximation of the intrinsic manifold for general semi-supervised learning (SSL) problems. Unfortunately, it is not trivial to define an optimization function to obtain optimal hyperparameters. Usually, cross validation is applied, but it does not necessarily scale up. Other problems derive from the suboptimality incurred by discrete grid search and the overfitting. Therefore, we develop an ensemble manifold regularization (EMR) framework to approximate the intrinsic manifold by combining several initial guesses. Algorithmically, we designed EMR carefully so it 1) learns both the composite manifold and the semi-supervised learner jointly, 2) is fully automatic for learning the intrinsic manifold hyperparameters implicitly, 3) is conditionally optimal for intrinsic manifold approximation under a mild and reasonable assumption, and 4) is scalable for a large number of candidate manifold hyperparameters, from both time and space perspectives. Furthermore, we prove the convergence property of EMR to the deterministic matrix at rate root-n. Extensive experiments over both synthetic and real data sets demonstrate the effectiveness of the proposed framework.

  3. Extension of the GHJW theorem for operator ensembles

    International Nuclear Information System (INIS)

    Choi, Jeong Woon; Hong, Dowon; Chang, Ku-Young; Chi, Dong Pyo; Lee, Soojoon

    2011-01-01

    The Gisin-Hughston-Jozsa-Wootters theorem plays an important role in analyzing various theories about quantum information, quantum communication, and quantum cryptography. It means that any purifications on the extended system which yield indistinguishable state ensembles on their subsystem should have a specific local unitary relation. In this Letter, we show that the local relation is also established even when the indistinguishability of state ensembles is extended to that of operator ensembles.

  4. Skill prediction of local weather forecasts based on the ECMWF ensemble

    Directory of Open Access Journals (Sweden)

    C. Ziehmann

    2001-01-01

    Full Text Available Ensemble Prediction has become an essential part of numerical weather forecasting. In this paper we investigate the ability of ensemble forecasts to provide an a priori estimate of the expected forecast skill. Several quantities derived from the local ensemble distribution are investigated for a two year data set of European Centre for Medium-Range Weather Forecasts (ECMWF temperature and wind speed ensemble forecasts at 30 German stations. The results indicate that the population of the ensemble mode provides useful information for the uncertainty in temperature forecasts. The ensemble entropy is a similar good measure. This is not true for the spread if it is simply calculated as the variance of the ensemble members with respect to the ensemble mean. The number of clusters in the C regions is almost unrelated to the local skill. For wind forecasts, the results are less promising.

  5. Ensemble methods for seasonal limited area forecasts

    DEFF Research Database (Denmark)

    Arritt, Raymond W.; Anderson, Christopher J.; Takle, Eugene S.

    2004-01-01

    The ensemble prediction methods used for seasonal limited area forecasts were examined by comparing methods for generating ensemble simulations of seasonal precipitation. The summer 1993 model over the north-central US was used as a test case. The four methods examined included the lagged-average...

  6. Class-specific Error Bounds for Ensemble Classifiers

    Energy Technology Data Exchange (ETDEWEB)

    Prenger, R; Lemmond, T; Varshney, K; Chen, B; Hanley, W

    2009-10-06

    The generalization error, or probability of misclassification, of ensemble classifiers has been shown to be bounded above by a function of the mean correlation between the constituent (i.e., base) classifiers and their average strength. This bound suggests that increasing the strength and/or decreasing the correlation of an ensemble's base classifiers may yield improved performance under the assumption of equal error costs. However, this and other existing bounds do not directly address application spaces in which error costs are inherently unequal. For applications involving binary classification, Receiver Operating Characteristic (ROC) curves, performance curves that explicitly trade off false alarms and missed detections, are often utilized to support decision making. To address performance optimization in this context, we have developed a lower bound for the entire ROC curve that can be expressed in terms of the class-specific strength and correlation of the base classifiers. We present empirical analyses demonstrating the efficacy of these bounds in predicting relative classifier performance. In addition, we specify performance regions of the ROC curve that are naturally delineated by the class-specific strengths of the base classifiers and show that each of these regions can be associated with a unique set of guidelines for performance optimization of binary classifiers within unequal error cost regimes.

  7. Creating ensembles of decision trees through sampling

    Science.gov (United States)

    Kamath, Chandrika; Cantu-Paz, Erick

    2005-08-30

    A system for decision tree ensembles that includes a module to read the data, a module to sort the data, a module to evaluate a potential split of the data according to some criterion using a random sample of the data, a module to split the data, and a module to combine multiple decision trees in ensembles. The decision tree method is based on statistical sampling techniques and includes the steps of reading the data; sorting the data; evaluating a potential split according to some criterion using a random sample of the data, splitting the data, and combining multiple decision trees in ensembles.

  8. An Efficient Ensemble Learning Method for Gene Microarray Classification

    Directory of Open Access Journals (Sweden)

    Alireza Osareh

    2013-01-01

    Full Text Available The gene microarray analysis and classification have demonstrated an effective way for the effective diagnosis of diseases and cancers. However, it has been also revealed that the basic classification techniques have intrinsic drawbacks in achieving accurate gene classification and cancer diagnosis. On the other hand, classifier ensembles have received increasing attention in various applications. Here, we address the gene classification issue using RotBoost ensemble methodology. This method is a combination of Rotation Forest and AdaBoost techniques which in turn preserve both desirable features of an ensemble architecture, that is, accuracy and diversity. To select a concise subset of informative genes, 5 different feature selection algorithms are considered. To assess the efficiency of the RotBoost, other nonensemble/ensemble techniques including Decision Trees, Support Vector Machines, Rotation Forest, AdaBoost, and Bagging are also deployed. Experimental results have revealed that the combination of the fast correlation-based feature selection method with ICA-based RotBoost ensemble is highly effective for gene classification. In fact, the proposed method can create ensemble classifiers which outperform not only the classifiers produced by the conventional machine learning but also the classifiers generated by two widely used conventional ensemble learning methods, that is, Bagging and AdaBoost.

  9. Sub-Ensemble Coastal Flood Forecasting: A Case Study of Hurricane Sandy

    Directory of Open Access Journals (Sweden)

    Justin A. Schulte

    2017-12-01

    Full Text Available In this paper, it is proposed that coastal flood ensemble forecasts be partitioned into sub-ensemble forecasts using cluster analysis in order to produce representative statistics and to measure forecast uncertainty arising from the presence of clusters. After clustering the ensemble members, the ability to predict the cluster into which the observation will fall can be measured using a cluster skill score. Additional sub-ensemble and composite skill scores are proposed for assessing the forecast skill of a clustered ensemble forecast. A recently proposed method for statistically increasing the number of ensemble members is used to improve sub-ensemble probabilistic estimates. Through the application of the proposed methodology to Sandy coastal flood reforecasts, it is demonstrated that statistics computed using only ensemble members belonging to a specific cluster are more representative than those computed using all ensemble members simultaneously. A cluster skill-cluster uncertainty index relationship is identified, which is the cluster analog of the documented spread-skill relationship. Two sub-ensemble skill scores are shown to be positively correlated with cluster forecast skill, suggesting that skillfully forecasting the cluster into which the observation will fall is important to overall forecast skill. The identified relationships also suggest that the number of ensemble members within in each cluster can be used as guidance for assessing the potential for forecast error. The inevitable existence of ensemble member clusters in tidally dominated total water level prediction systems suggests that clustering is a necessary post-processing step for producing representative and skillful total water level forecasts.

  10. Ensemble Bayesian forecasting system Part I: Theory and algorithms

    Science.gov (United States)

    Herr, Henry D.; Krzysztofowicz, Roman

    2015-05-01

    The ensemble Bayesian forecasting system (EBFS), whose theory was published in 2001, is developed for the purpose of quantifying the total uncertainty about a discrete-time, continuous-state, non-stationary stochastic process such as a time series of stages, discharges, or volumes at a river gauge. The EBFS is built of three components: an input ensemble forecaster (IEF), which simulates the uncertainty associated with random inputs; a deterministic hydrologic model (of any complexity), which simulates physical processes within a river basin; and a hydrologic uncertainty processor (HUP), which simulates the hydrologic uncertainty (an aggregate of all uncertainties except input). It works as a Monte Carlo simulator: an ensemble of time series of inputs (e.g., precipitation amounts) generated by the IEF is transformed deterministically through a hydrologic model into an ensemble of time series of outputs, which is next transformed stochastically by the HUP into an ensemble of time series of predictands (e.g., river stages). Previous research indicated that in order to attain an acceptable sampling error, the ensemble size must be on the order of hundreds (for probabilistic river stage forecasts and probabilistic flood forecasts) or even thousands (for probabilistic stage transition forecasts). The computing time needed to run the hydrologic model this many times renders the straightforward simulations operationally infeasible. This motivates the development of the ensemble Bayesian forecasting system with randomization (EBFSR), which takes full advantage of the analytic meta-Gaussian HUP and generates multiple ensemble members after each run of the hydrologic model; this auxiliary randomization reduces the required size of the meteorological input ensemble and makes it operationally feasible to generate a Bayesian ensemble forecast of large size. Such a forecast quantifies the total uncertainty, is well calibrated against the prior (climatic) distribution of

  11. A soil-landscape framework for understanding spatial and temporal variability in biogeochemical processes in catchments

    Science.gov (United States)

    McGuire, K. J.; Bailey, S. W.; Ross, D. S.

    2017-12-01

    Heterogeneity in biophysical properties within catchments challenges how we quantify and characterize biogeochemical processes and interpret catchment outputs. Interactions between the spatiotemporal variability of hydrological states and fluxes and soil development can spatially structure catchments, leading to a framework for understanding patterns in biogeochemical processes. In an upland, glaciated landscape at the Hubbard Brook Experimental Forest (HBEF) in New Hampshire, USA, we are embracing the structure and organization of soils to understand the spatial relations between runoff production zones, distinct soil-biogeochemical environments, and solute retention and release. This presentation will use observations from the HBEF to demonstrate that a soil-landscape framework is essential in understanding the spatial and temporal variability of biogeochemical processes in this catchment. Specific examples will include how laterally developed soils reveal the location of active runoff production zones and lead to gradients in primary mineral dissolution and the distribution of weathering products along hillslopes. Soil development patterns also highlight potential carbon and nitrogen cycling hotspots, differentiate acidic conditions, and affect the regulation of surface water quality. Overall, this work demonstrates the importance of understanding the landscape-level structural organization of soils in characterizing the variation and extent of biogeochemical processes that occur in catchments.

  12. Supersymmetry applied to the spectrum edge of random matrix ensembles

    International Nuclear Information System (INIS)

    Andreev, A.V.; Simons, B.D.; Taniguchi, N.

    1994-01-01

    A new matrix ensemble has recently been proposed to describe the transport properties in mesoscopic quantum wires. Both analytical and numerical studies have shown that the ensemble of Laguerre or of chiral random matrices provides a good description of scattering properties in this class of systems. Until now only conventional methods of random matrix theory have been used to study statistical properties within this ensemble. We demonstrate that the supersymmetry method, already employed in the study Dyson ensembles, can be extended to treat this class of random matrix ensembles. In developing this approach we investigate both new, as well as verify known statistical measures. Although we focus on ensembles in which T-invariance is violated our approach lays the foundation for future studies of T-invariant systems. ((orig.))

  13. The acclimative biogeochemical model of the southern North Sea

    Science.gov (United States)

    Kerimoglu, Onur; Hofmeister, Richard; Maerz, Joeran; Riethmüller, Rolf; Wirtz, Kai W.

    2017-10-01

    Ecosystem models often rely on heuristic descriptions of autotrophic growth that fail to reproduce various stationary and dynamic states of phytoplankton cellular composition observed in laboratory experiments. Here, we present the integration of an advanced phytoplankton growth model within a coupled three-dimensional physical-biogeochemical model and the application of the model system to the southern North Sea (SNS) defined on a relatively high resolution (˜ 1.5-4.5 km) curvilinear grid. The autotrophic growth model, recently introduced by Wirtz and Kerimoglu (2016), is based on a set of novel concepts for the allocation of internal resources and operation of cellular metabolism. The coupled model system consists of the General Estuarine Transport Model (GETM) as the hydrodynamical driver, a lower-trophic-level model and a simple sediment diagenesis model. We force the model system with realistic atmospheric and riverine fluxes, background turbidity caused by suspended particulate matter (SPM) and open ocean boundary conditions. For a simulation for the period 2000-2010, we show that the model system satisfactorily reproduces the physical and biogeochemical states of the system within the German Bight characterized by steep salinity; nutrient and chlorophyll (Chl) gradients, as inferred from comparisons against observation data from long-term monitoring stations; sparse in situ measurements; continuous transects; and satellites. The model also displays skill in capturing the formation of thin chlorophyll layers at the pycnocline, which is frequently observed within the stratified regions during summer. A sensitivity analysis reveals that the vertical distributions of phytoplankton concentrations estimated by the model can be qualitatively sensitive to the description of the light climate and dependence of sinking rates on the internal nutrient reserves. A non-acclimative (fixed-physiology) version of the model predicted entirely different vertical profiles

  14. Modeling task-specific neuronal ensembles improves decoding of grasp

    Science.gov (United States)

    Smith, Ryan J.; Soares, Alcimar B.; Rouse, Adam G.; Schieber, Marc H.; Thakor, Nitish V.

    2018-06-01

    Objective. Dexterous movement involves the activation and coordination of networks of neuronal populations across multiple cortical regions. Attempts to model firing of individual neurons commonly treat the firing rate as directly modulating with motor behavior. However, motor behavior may additionally be associated with modulations in the activity and functional connectivity of neurons in a broader ensemble. Accounting for variations in neural ensemble connectivity may provide additional information about the behavior being performed. Approach. In this study, we examined neural ensemble activity in primary motor cortex (M1) and premotor cortex (PM) of two male rhesus monkeys during performance of a center-out reach, grasp and manipulate task. We constructed point process encoding models of neuronal firing that incorporated task-specific variations in the baseline firing rate as well as variations in functional connectivity with the neural ensemble. Models were evaluated both in terms of their encoding capabilities and their ability to properly classify the grasp being performed. Main results. Task-specific ensemble models correctly predicted the performed grasp with over 95% accuracy and were shown to outperform models of neuronal activity that assume only a variable baseline firing rate. Task-specific ensemble models exhibited superior decoding performance in 82% of units in both monkeys (p  <  0.01). Inclusion of ensemble activity also broadly improved the ability of models to describe observed spiking. Encoding performance of task-specific ensemble models, measured by spike timing predictability, improved upon baseline models in 62% of units. Significance. These results suggest that additional discriminative information about motor behavior found in the variations in functional connectivity of neuronal ensembles located in motor-related cortical regions is relevant to decode complex tasks such as grasping objects, and may serve the basis for more

  15. The complex Laguerre symplectic ensemble of non-Hermitian matrices

    International Nuclear Information System (INIS)

    Akemann, G.

    2005-01-01

    We solve the complex extension of the chiral Gaussian symplectic ensemble, defined as a Gaussian two-matrix model of chiral non-Hermitian quaternion real matrices. This leads to the appearance of Laguerre polynomials in the complex plane and we prove their orthogonality. Alternatively, a complex eigenvalue representation of this ensemble is given for general weight functions. All k-point correlation functions of complex eigenvalues are given in terms of the corresponding skew orthogonal polynomials in the complex plane for finite-N, where N is the matrix size or number of eigenvalues, respectively. We also allow for an arbitrary number of complex conjugate pairs of characteristic polynomials in the weight function, corresponding to massive quark flavours in applications to field theory. Explicit expressions are given in the large-N limit at both weak and strong non-Hermiticity for the weight of the Gaussian two-matrix model. This model can be mapped to the complex Dirac operator spectrum with non-vanishing chemical potential. It belongs to the symmetry class of either the adjoint representation or two colours in the fundamental representation using staggered lattice fermions

  16. Towards an Australian ensemble streamflow forecasting system for flood prediction and water management

    Science.gov (United States)

    Bennett, J.; David, R. E.; Wang, Q.; Li, M.; Shrestha, D. L.

    2016-12-01

    Flood forecasting in Australia has historically relied on deterministic forecasting models run only when floods are imminent, with considerable forecaster input and interpretation. These now co-existed with a continually available 7-day streamflow forecasting service (also deterministic) aimed at operational water management applications such as environmental flow releases. The 7-day service is not optimised for flood prediction. We describe progress on developing a system for ensemble streamflow forecasting that is suitable for both flood prediction and water management applications. Precipitation uncertainty is handled through post-processing of Numerical Weather Prediction (NWP) output with a Bayesian rainfall post-processor (RPP). The RPP corrects biases, downscales NWP output, and produces reliable ensemble spread. Ensemble precipitation forecasts are used to force a semi-distributed conceptual rainfall-runoff model. Uncertainty in precipitation forecasts is insufficient to reliably describe streamflow forecast uncertainty, particularly at shorter lead-times. We characterise hydrological prediction uncertainty separately with a 4-stage error model. The error model relies on data transformation to ensure residuals are homoscedastic and symmetrically distributed. To ensure streamflow forecasts are accurate and reliable, the residuals are modelled using a mixture-Gaussian distribution with distinct parameters for the rising and falling limbs of the forecast hydrograph. In a case study of the Murray River in south-eastern Australia, we show ensemble predictions of floods generally have lower errors than deterministic forecasting methods. We also discuss some of the challenges in operationalising short-term ensemble streamflow forecasts in Australia, including meeting the needs for accurate predictions across all flow ranges and comparing forecasts generated by event and continuous hydrological models.

  17. Bacterial Production and Enzymatic Activities in Deep-Sea Sediments of the Pacific Ocean: Biogeochemical Implications of Different Temperature Constraints

    Science.gov (United States)

    Danovaro, R.; Corinaldesi, C.; dell'Anno, A.

    2002-12-01

    The deep-sea bed, acting as the ultimate sink for organic material derived from the upper oceans primary production, is now assumed to play a key role in biogeochemical cycling of organic matter on global scale. Early diagenesis of organic matter in marine sediments is dependent upon biological processes (largely mediated by bacterial activity) and by molecular diffusion. Organic matter reaching the sea floor by sedimentation is subjected to complex biogeochemical transformations that make organic matter largely unsuitable for direct utilization by benthic heterotrophs. Extracellular enzymatic activities in the sediment is generally recognized as the key step in the degradation and utilization of organic polymers by bacteria and a key role in biopolymeric carbon mobilization is played by aminopeptidase, alkaline phosphatase and glucosidase activities. In the present study we investigated bacterial density, bacterial C production and exo-enzymatic activities (aminopeptidase, glucosidase and phosphatase activity) in deep-sea sediments of the Pacific Ocean in relation with the biochemical composition of sediment organic matter (proteins, carbohydrates and lipids), in order to gather information on organic matter cycling and diagenesis. Benthic viral abundance was also measured to investigate the potential role of viruses on microbial loop functioning. Sediment samples were collected at eight stations (depth ranging from 2070-3100 m) along two transects located at the opposite side (north and south) of ocean seismic ridge Juan Fernandez (along latitudes 33° 20' - 33° 40'), constituted by the submerged vulcanoes, which connects the Chilean coasts to Rapa Nui Island. Since the northern and southern sides of this ridge apparently displayed small but significant differences in deep-sea temperature (related to the general ocean circulation), this sampling strategy allowed also investigating the role of different temperature constraints on bacterial activity and

  18. Generation of scenarios from calibrated ensemble forecasts with a dual ensemble copula coupling approach

    DEFF Research Database (Denmark)

    Ben Bouallègue, Zied; Heppelmann, Tobias; Theis, Susanne E.

    2016-01-01

    the original ensemble forecasts. Based on the assumption of error stationarity, parametric methods aim to fully describe the forecast dependence structures. In this study, the concept of ECC is combined with past data statistics in order to account for the autocorrelation of the forecast error. The new...... approach, called d-ECC, is applied to wind forecasts from the high resolution ensemble system COSMO-DE-EPS run operationally at the German weather service. Scenarios generated by ECC and d-ECC are compared and assessed in the form of time series by means of multivariate verification tools and in a product...

  19. Encoding of Spatial Attention by Primate Prefrontal Cortex Neuronal Ensembles

    Science.gov (United States)

    Treue, Stefan

    2018-01-01

    Abstract Single neurons in the primate lateral prefrontal cortex (LPFC) encode information about the allocation of visual attention and the features of visual stimuli. However, how this compares to the performance of neuronal ensembles at encoding the same information is poorly understood. Here, we recorded the responses of neuronal ensembles in the LPFC of two macaque monkeys while they performed a task that required attending to one of two moving random dot patterns positioned in different hemifields and ignoring the other pattern. We found single units selective for the location of the attended stimulus as well as for its motion direction. To determine the coding of both variables in the population of recorded units, we used a linear classifier and progressively built neuronal ensembles by iteratively adding units according to their individual performance (best single units), or by iteratively adding units based on their contribution to the ensemble performance (best ensemble). For both methods, ensembles of relatively small sizes (n decoding performance relative to individual single units. However, the decoder reached similar performance using fewer neurons with the best ensemble building method compared with the best single units method. Our results indicate that neuronal ensembles within the LPFC encode more information about the attended spatial and nonspatial features of visual stimuli than individual neurons. They further suggest that efficient coding of attention can be achieved by relatively small neuronal ensembles characterized by a certain relationship between signal and noise correlation structures. PMID:29568798

  20. Flood Forecasting Based on TIGGE Precipitation Ensemble Forecast

    Directory of Open Access Journals (Sweden)

    Jinyin Ye

    2016-01-01

    Full Text Available TIGGE (THORPEX International Grand Global Ensemble was a major part of the THORPEX (Observing System Research and Predictability Experiment. It integrates ensemble precipitation products from all the major forecast centers in the world and provides systematic evaluation on the multimodel ensemble prediction system. Development of meteorologic-hydrologic coupled flood forecasting model and early warning model based on the TIGGE precipitation ensemble forecast can provide flood probability forecast, extend the lead time of the flood forecast, and gain more time for decision-makers to make the right decision. In this study, precipitation ensemble forecast products from ECMWF, NCEP, and CMA are used to drive distributed hydrologic model TOPX. We focus on Yi River catchment and aim to build a flood forecast and early warning system. The results show that the meteorologic-hydrologic coupled model can satisfactorily predict the flow-process of four flood events. The predicted occurrence time of peak discharges is close to the observations. However, the magnitude of the peak discharges is significantly different due to various performances of the ensemble prediction systems. The coupled forecasting model can accurately predict occurrence of the peak time and the corresponding risk probability of peak discharge based on the probability distribution of peak time and flood warning, which can provide users a strong theoretical foundation and valuable information as a promising new approach.

  1. Three-model ensemble wind prediction in southern Italy

    Science.gov (United States)

    Torcasio, Rosa Claudia; Federico, Stefano; Calidonna, Claudia Roberta; Avolio, Elenio; Drofa, Oxana; Landi, Tony Christian; Malguzzi, Piero; Buzzi, Andrea; Bonasoni, Paolo

    2016-03-01

    Quality of wind prediction is of great importance since a good wind forecast allows the prediction of available wind power, improving the penetration of renewable energies into the energy market. Here, a 1-year (1 December 2012 to 30 November 2013) three-model ensemble (TME) experiment for wind prediction is considered. The models employed, run operationally at National Research Council - Institute of Atmospheric Sciences and Climate (CNR-ISAC), are RAMS (Regional Atmospheric Modelling System), BOLAM (BOlogna Limited Area Model), and MOLOCH (MOdello LOCale in H coordinates). The area considered for the study is southern Italy and the measurements used for the forecast verification are those of the GTS (Global Telecommunication System). Comparison with observations is made every 3 h up to 48 h of forecast lead time. Results show that the three-model ensemble outperforms the forecast of each individual model. The RMSE improvement compared to the best model is between 22 and 30 %, depending on the season. It is also shown that the three-model ensemble outperforms the IFS (Integrated Forecasting System) of the ECMWF (European Centre for Medium-Range Weather Forecast) for the surface wind forecasts. Notably, the three-model ensemble forecast performs better than each unbiased model, showing the added value of the ensemble technique. Finally, the sensitivity of the three-model ensemble RMSE to the length of the training period is analysed.

  2. Biogeochemical impacts of wildfires over four millennia in a Rocky Mountain subalpine watershed.

    Science.gov (United States)

    Dunnette, Paul V; Higuera, Philip E; McLauchlan, Kendra K; Derr, Kelly M; Briles, Christy E; Keefe, Margaret H

    2014-08-01

    Wildfires can significantly alter forest carbon (C) storage and nitrogen (N) availability, but the long-term biogeochemical legacy of wildfires is poorly understood. We obtained a lake-sediment record of fire and biogeochemistry from a subalpine forest in Colorado, USA, to examine the nature, magnitude, and duration of decadal-scale, fire-induced ecosystem change over the past c. 4250 yr. The high-resolution record contained 34 fires, including 13 high-severity events within the watershed. High-severity fires were followed by increased sedimentary N stable isotope ratios (δ15N) and bulk density, and decreased C and N concentrations--reflecting forest floor destruction, terrestrial C and N losses, and erosion. Sustained low sediment C : N c. 20-50 yr post-fire indicates reduced terrestrial organic matter subsidies to the lake. Low sedimentary δ15N c. 50-70 yr post-fire, coincident with C and N recovery, suggests diminishing terrestrial N availability during stand development. The magnitude of post-fire changes generally scaled directly with inferred fire severity. Our results support modern studies of forest successional C and N accumulation and indicate pronounced, long-lasting biogeochemical impacts of wildfires in subalpine forests. However, even repeated high-severity fires over millennia probably did not deplete C or N stocks, because centuries between high-severity fires allowed for sufficient biomass recovery. © 2014 The Authors. New Phytologist © 2014 New Phytologist Trust.

  3. Ensembles of gustatory cortical neurons anticipate and discriminate between tastants in a single lick

    Directory of Open Access Journals (Sweden)

    Jennifer R Stapleton

    2007-10-01

    Full Text Available The gustatory cortex (GC processes chemosensory and somatosensory information and is involved in learning and anticipation. Previously we found that a subpopulation of GC neurons responded to tastants in a single lick (Stapleton et al., 2006. Here we extend this investigation to determine if small ensembles of GC neurons, obtained while rats received blocks of tastants on a fixed ratio schedule (FR5, can discriminate between tastants and their concentrations after a single 50 µL delivery. In the FR5 schedule subjects received tastants every fifth (reinforced lick and the intervening licks were unreinforced. The ensemble firing patterns were analyzed with a Bayesian generalized linear model whose parameters included the firing rates and temporal patterns of the spike trains. We found that when both the temporal and rate parameters were included, 12 of 13 ensembles correctly identified single tastant deliveries. We also found that the activity during the unreinforced licks contained signals regarding the identity of the upcoming tastant, which suggests that GC neurons contain anticipatory information about the next tastant delivery. To support this finding we performed experiments in which tastant delivery was randomized within each block and found that the neural activity following the unreinforced licks did not predict the upcoming tastant. Collectively, these results suggest that after a single lick ensembles of GC neurons can discriminate between tastants, that they may utilize both temporal and rate information, and when the tastant delivery is repetitive ensembles contain information about the identity of the upcoming tastant delivery.

  4. A Fault-Tolerant HPC Scheduler Extension for Large and Operational Ensemble Data Assimilation:Application to the Red Sea

    KAUST Repository

    Toye, Habib

    2018-04-26

    A fully parallel ensemble data assimilation and forecasting system has been developed for the Red Sea based on the MIT general circulation model (MITgcm) to simulate the Red Sea circulation and the Data Assimilation Research Testbed (DART) ensemble assimilation software. An important limitation of operational ensemble assimilation systems is the risk of ensemble members’ collapse. This could happen in those situations when the filter update step imposes large corrections on one, or more, of the forecasted ensemble members that are not fully consistent with the model physics. Increasing the ensemble size is expected to improve the assimilation system performances, but obviously increases the risk of members’ collapse. Hardware failure or slow numerical convergence encountered for some members should also occur more frequently. In this context, the manual steering of the whole process appears as a real challenge and makes the implementation of the ensemble assimilation procedure uneasy and extremely time consuming.This paper presents our efforts to build an efficient and fault-tolerant MITgcm-DART ensemble assimilation system capable of operationally running thousands of members. Built on top of Decimate, a scheduler extension developed to ease the submission, monitoring and dynamic steering of workflow of dependent jobs in a fault-tolerant environment, we describe the assimilation system implementation and discuss in detail its coupling strategies. Within Decimate, only a few additional lines of Python is needed to define flexible convergence criteria and to implement any necessary actions to the forecast ensemble members, as for instance (i) restarting faulty job in case of job failure, (ii) changing the random seed in case of poor convergence or numerical instability, (iii) adjusting (reducing or increasing) the number of parallel forecasts on the fly, (iv) replacing members on the fly to enrich the ensemble with new members, etc.We demonstrate the efficiency

  5. Surrogate-Based Optimization of Biogeochemical Transport Models

    Science.gov (United States)

    Prieß, Malte; Slawig, Thomas

    2010-09-01

    First approaches towards a surrogate-based optimization method for a one-dimensional marine biogeochemical model of NPZD type are presented. The model, developed by Oschlies and Garcon [1], simulates the distribution of nitrogen, phytoplankton, zooplankton and detritus in a water column and is driven by ocean circulation data. A key issue is to minimize the misfit between the model output and given observational data. Our aim is to reduce the overall optimization cost avoiding expensive function and derivative evaluations by using a surrogate model replacing the high-fidelity model in focus. This in particular becomes important for more complex three-dimensional models. We analyse a coarsening in the discretization of the model equations as one way to create such a surrogate. Here the numerical stability crucially depends upon the discrete stepsize in time and space and the biochemical terms. We show that for given model parameters the level of grid coarsening can be choosen accordingly yielding a stable and satisfactory surrogate. As one example of a surrogate-based optimization method we present results of the Aggressive Space Mapping technique (developed by John W. Bandler [2, 3]) applied to the optimization of this one-dimensional biogeochemical transport model.

  6. Linking Chaotic Advection with Subsurface Biogeochemical Processes

    Science.gov (United States)

    Mays, D. C.; Freedman, V. L.; White, S. K.; Fang, Y.; Neupauer, R.

    2017-12-01

    This work investigates the extent to which groundwater flow kinematics drive subsurface biogeochemical processes. In terms of groundwater flow kinematics, we consider chaotic advection, whose essential ingredient is stretching and folding of plumes. Chaotic advection is appealing within the context of groundwater remediation because it has been shown to optimize plume spreading in the laminar flows characteristic of aquifers. In terms of subsurface biogeochemical processes, we consider an existing model for microbially-mediated reduction of relatively mobile uranium(VI) to relatively immobile uranium(IV) following injection of acetate into a floodplain aquifer beneath a former uranium mill in Rifle, Colorado. This model has been implemented in the reactive transport code eSTOMP, the massively parallel version of STOMP (Subsurface Transport Over Multiple Phases). This presentation will report preliminary numerical simulations in which the hydraulic boundary conditions in the eSTOMP model are manipulated to simulate chaotic advection resulting from engineered injection and extraction of water through a manifold of wells surrounding the plume of injected acetate. This approach provides an avenue to simulate the impact of chaotic advection within the existing framework of the eSTOMP code.

  7. Global Ensemble Forecast System (GEFS) [2.5 Deg.

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The Global Ensemble Forecast System (GEFS) is a weather forecast model made up of 21 separate forecasts, or ensemble members. The National Centers for Environmental...

  8. An ensemble classifier to predict track geometry degradation

    International Nuclear Information System (INIS)

    Cárdenas-Gallo, Iván; Sarmiento, Carlos A.; Morales, Gilberto A.; Bolivar, Manuel A.; Akhavan-Tabatabaei, Raha

    2017-01-01

    Railway operations are inherently complex and source of several problems. In particular, track geometry defects are one of the leading causes of train accidents in the United States. This paper presents a solution approach which entails the construction of an ensemble classifier to forecast the degradation of track geometry. Our classifier is constructed by solving the problem from three different perspectives: deterioration, regression and classification. We considered a different model from each perspective and our results show that using an ensemble method improves the predictive performance. - Highlights: • We present an ensemble classifier to forecast the degradation of track geometry. • Our classifier considers three perspectives: deterioration, regression and classification. • We construct and test three models and our results show that using an ensemble method improves the predictive performance.

  9. The effect of biogeochemical processes on pH

    NARCIS (Netherlands)

    Soetaert, K.E.R.; Hofmann, A.F.; Middelburg, J.J.; Meysman, F.J.R.; Greenwood, J.E.

    2007-01-01

    The impact of biogeochemical and physical processes on aquatic chemistry is usually expressed in terms of alkalinity. Here we show how to directly calculate the effect of single processes on pH. Under the assumptions of equilibrium and electroneutrality, the rate of change of pH can be calculated as

  10. Time-optimal path planning in uncertain flow fields using ensemble method

    KAUST Repository

    Wang, Tong

    2016-01-06

    An ensemble-based approach is developed to conduct time-optimal path planning in unsteady ocean currents under uncertainty. We focus our attention on two-dimensional steady and unsteady uncertain flows, and adopt a sampling methodology that is well suited to operational forecasts, where a set deterministic predictions is used to model and quantify uncertainty in the predictions. In the operational setting, much about dynamics, topography and forcing of the ocean environment is uncertain, and as a result a single path produced by a model simulation has limited utility. To overcome this limitation, we rely on a finitesize ensemble of deterministic forecasts to quantify the impact of variability in the dynamics. The uncertainty of flow field is parametrized using a finite number of independent canonical random variables with known densities, and the ensemble is generated by sampling these variables. For each the resulting realizations of the uncertain current field, we predict the optimal path by solving a boundary value problem (BVP), based on the Pontryagin maximum principle. A family of backward-in-time trajectories starting at the end position is used to generate suitable initial values for the BVP solver. This allows us to examine and analyze the performance of sampling strategy, and develop insight into extensions dealing with regional or general circulation models. In particular, the ensemble method enables us to perform a statistical analysis of travel times, and consequently develop a path planning approach that accounts for these statistics. The proposed methodology is tested for a number of scenarios. We first validate our algorithms by reproducing simple canonical solutions, and then demonstrate our approach in more complex flow fields, including idealized, steady and unsteady double-gyre flows.

  11. Ensembler: Enabling High-Throughput Molecular Simulations at the Superfamily Scale.

    Directory of Open Access Journals (Sweden)

    Daniel L Parton

    2016-06-01

    tyrosine kinase family, using all available kinase catalytic domain structures from any organism as structural templates. Ensembler is free and open source software licensed under the GNU General Public License (GPL v2. It is compatible with Linux and OS X. The latest release can be installed via the conda package manager, and the latest source can be downloaded from https://github.com/choderalab/ensembler.

  12. Development of a regional ensemble prediction method for probabilistic weather prediction

    International Nuclear Information System (INIS)

    Nohara, Daisuke; Tamura, Hidetoshi; Hirakuchi, Hiromaru

    2015-01-01

    A regional ensemble prediction method has been developed to provide probabilistic weather prediction using a numerical weather prediction model. To obtain consistent perturbations with the synoptic weather pattern, both of initial and lateral boundary perturbations were given by differences between control and ensemble member of the Japan Meteorological Agency (JMA)'s operational one-week ensemble forecast. The method provides a multiple ensemble member with a horizontal resolution of 15 km for 48-hour based on a downscaling of the JMA's operational global forecast accompanied with the perturbations. The ensemble prediction was examined in the case of heavy snow fall event in Kanto area on January 14, 2013. The results showed that the predictions represent different features of high-resolution spatiotemporal distribution of precipitation affected by intensity and location of extra-tropical cyclone in each ensemble member. Although the ensemble prediction has model bias of mean values and variances in some variables such as wind speed and solar radiation, the ensemble prediction has a potential to append a probabilistic information to a deterministic prediction. (author)

  13. Concrete ensemble Kalman filters with rigorous catastrophic filter divergence.

    Science.gov (United States)

    Kelly, David; Majda, Andrew J; Tong, Xin T

    2015-08-25

    The ensemble Kalman filter and ensemble square root filters are data assimilation methods used to combine high-dimensional, nonlinear dynamical models with observed data. Ensemble methods are indispensable tools in science and engineering and have enjoyed great success in geophysical sciences, because they allow for computationally cheap low-ensemble-state approximation for extremely high-dimensional turbulent forecast models. From a theoretical perspective, the dynamical properties of these methods are poorly understood. One of the central mysteries is the numerical phenomenon known as catastrophic filter divergence, whereby ensemble-state estimates explode to machine infinity, despite the true state remaining in a bounded region. In this article we provide a breakthrough insight into the phenomenon, by introducing a simple and natural forecast model that transparently exhibits catastrophic filter divergence under all ensemble methods and a large set of initializations. For this model, catastrophic filter divergence is not an artifact of numerical instability, but rather a true dynamical property of the filter. The divergence is not only validated numerically but also proven rigorously. The model cleanly illustrates mechanisms that give rise to catastrophic divergence and confirms intuitive accounts of the phenomena given in past literature.

  14. An ensemble approach to simulate CO2 emissions from natural fires

    Science.gov (United States)

    Eliseev, A. V.; Mokhov, I. I.; Chernokulsky, A. V.

    2014-06-01

    This paper presents ensemble simulations with the global climate model developed at the A. M. Obukhov Institute of Atmospheric Physics, Russian Academy of Sciences (IAP RAS CM). These simulations are forced by historical reconstructions of concentrations of well-mixed greenhouse gases (CO2, CH4, and N2O), sulfate aerosols (both in the troposphere and stratosphere), extent of crops and pastures, and total solar irradiance for AD 850-2005 (hereafter all years are taken as being AD) and by the Representative Concentration Pathway (RCP) scenarios for the same forcing agents until the year 2300. Our model implements GlobFIRM (Global FIRe Model) as a scheme for calculating characteristics of natural fires. Comparing to the original GlobFIRM model, in our implementation, the scheme is extended by a module accounting for CO2 release from soil during fires. The novel approach of our paper is to simulate natural fires in an ensemble fashion. Different ensemble members in the present paper are constructed by varying the values of parameters of the natural fires module. These members are constrained by the GFED-3.1 data set for the burnt area and CO2 release from fires and further subjected to Bayesian averaging. Our simulations are the first coupled model assessment of future changes in gross characteristics of natural fires. In our model, the present-day (1998-2011) global area burnt due to natural fires is (2.1 ± 0.4) × 106 km2 yr-1 (ensemble mean and intra-ensemble standard deviation are presented), and the respective CO2 emissions to the atmosphere are (1.4 ± 0.2) Pg C yr-1. The latter value is in agreement with the corresponding GFED estimates. The area burnt by natural fires is generally larger than the GFED estimates except in boreal Eurasia, where it is realistic, and in Australia, where it is smaller than these estimates. Regionally, the modelled CO2 emissions are larger (smaller) than the GFED estimates in Europe (in the tropics and north-eastern Eurasia). From

  15. Quark ensembles with infinite correlation length

    OpenAIRE

    Molodtsov, S. V.; Zinovjev, G. M.

    2014-01-01

    By studying quark ensembles with infinite correlation length we formulate the quantum field theory model that, as we show, is exactly integrable and develops an instability of its standard vacuum ensemble (the Dirac sea). We argue such an instability is rooted in high ground state degeneracy (for 'realistic' space-time dimensions) featuring a fairly specific form of energy distribution, and with the cutoff parameter going to infinity this inherent energy distribution becomes infinitely narrow...

  16. A multi-model ensemble approach to seabed mapping

    Science.gov (United States)

    Diesing, Markus; Stephens, David

    2015-06-01

    Seabed habitat mapping based on swath acoustic data and ground-truth samples is an emergent and active marine science discipline. Significant progress could be achieved by transferring techniques and approaches that have been successfully developed and employed in such fields as terrestrial land cover mapping. One such promising approach is the multiple classifier system, which aims at improving classification performance by combining the outputs of several classifiers. Here we present results of a multi-model ensemble applied to multibeam acoustic data covering more than 5000 km2 of seabed in the North Sea with the aim to derive accurate spatial predictions of seabed substrate. A suite of six machine learning classifiers (k-Nearest Neighbour, Support Vector Machine, Classification Tree, Random Forest, Neural Network and Naïve Bayes) was trained with ground-truth sample data classified into seabed substrate classes and their prediction accuracy was assessed with an independent set of samples. The three and five best performing models were combined to classifier ensembles. Both ensembles led to increased prediction accuracy as compared to the best performing single classifier. The improvements were however not statistically significant at the 5% level. Although the three-model ensemble did not perform significantly better than its individual component models, we noticed that the five-model ensemble did perform significantly better than three of the five component models. A classifier ensemble might therefore be an effective strategy to improve classification performance. Another advantage is the fact that the agreement in predicted substrate class between the individual models of the ensemble could be used as a measure of confidence. We propose a simple and spatially explicit measure of confidence that is based on model agreement and prediction accuracy.

  17. A One-Step-Ahead Smoothing-Based Joint Ensemble Kalman Filter for State-Parameter Estimation of Hydrological Models

    KAUST Repository

    El Gharamti, Mohamad; Ait-El-Fquih, Boujemaa; Hoteit, Ibrahim

    2015-01-01

    The ensemble Kalman filter (EnKF) recursively integrates field data into simulation models to obtain a better characterization of the model’s state and parameters. These are generally estimated following a state-parameters joint augmentation

  18. Partitioning and mapping uncertainties in ensembles of forecasts of species turnover under climate change

    DEFF Research Database (Denmark)

    Diniz-Filho, José Alexandre F.; Bini, Luis Mauricio; Rangel, Thiago Fernando

    2009-01-01

    Forecasts of species range shifts under climate change are fraught with uncertainties and ensemble forecasting may provide a framework to deal with such uncertainties. Here, a novel approach to partition the variance among modeled attributes, such as richness or turnover, and map sources of uncer......Forecasts of species range shifts under climate change are fraught with uncertainties and ensemble forecasting may provide a framework to deal with such uncertainties. Here, a novel approach to partition the variance among modeled attributes, such as richness or turnover, and map sources...... of uncertainty in ensembles of forecasts is presented. We model the distributions of 3837 New World birds and project them into 2080. We then quantify and map the relative contribution of different sources of uncertainty from alternative methods for niche modeling, general circulation models (AOGCM......), and emission scenarios. The greatest source of uncertainty in forecasts of species range shifts arises from using alternative methods for niche modeling, followed by AOGCM, and their interaction. Our results concur with previous studies that discovered that projections from alternative models can be extremely...

  19. Critical Listening in the Ensemble Rehearsal: A Community of Learners

    Science.gov (United States)

    Bell, Cindy L.

    2018-01-01

    This article explores a strategy for engaging ensemble members in critical listening analysis of performances and presents opportunities for improving ensemble sound through rigorous dialogue, reflection, and attentive rehearsing. Critical listening asks ensemble members to draw on individual playing experience and knowledge to describe what they…

  20. SVM and SVM Ensembles in Breast Cancer Prediction.

    Science.gov (United States)

    Huang, Min-Wei; Chen, Chih-Wen; Lin, Wei-Chao; Ke, Shih-Wen; Tsai, Chih-Fong

    2017-01-01

    Breast cancer is an all too common disease in women, making how to effectively predict it an active research problem. A number of statistical and machine learning techniques have been employed to develop various breast cancer prediction models. Among them, support vector machines (SVM) have been shown to outperform many related techniques. To construct the SVM classifier, it is first necessary to decide the kernel function, and different kernel functions can result in different prediction performance. However, there have been very few studies focused on examining the prediction performances of SVM based on different kernel functions. Moreover, it is unknown whether SVM classifier ensembles which have been proposed to improve the performance of single classifiers can outperform single SVM classifiers in terms of breast cancer prediction. Therefore, the aim of this paper is to fully assess the prediction performance of SVM and SVM ensembles over small and large scale breast cancer datasets. The classification accuracy, ROC, F-measure, and computational times of training SVM and SVM ensembles are compared. The experimental results show that linear kernel based SVM ensembles based on the bagging method and RBF kernel based SVM ensembles with the boosting method can be the better choices for a small scale dataset, where feature selection should be performed in the data pre-processing stage. For a large scale dataset, RBF kernel based SVM ensembles based on boosting perform better than the other classifiers.

  1. SVM and SVM Ensembles in Breast Cancer Prediction.

    Directory of Open Access Journals (Sweden)

    Min-Wei Huang

    Full Text Available Breast cancer is an all too common disease in women, making how to effectively predict it an active research problem. A number of statistical and machine learning techniques have been employed to develop various breast cancer prediction models. Among them, support vector machines (SVM have been shown to outperform many related techniques. To construct the SVM classifier, it is first necessary to decide the kernel function, and different kernel functions can result in different prediction performance. However, there have been very few studies focused on examining the prediction performances of SVM based on different kernel functions. Moreover, it is unknown whether SVM classifier ensembles which have been proposed to improve the performance of single classifiers can outperform single SVM classifiers in terms of breast cancer prediction. Therefore, the aim of this paper is to fully assess the prediction performance of SVM and SVM ensembles over small and large scale breast cancer datasets. The classification accuracy, ROC, F-measure, and computational times of training SVM and SVM ensembles are compared. The experimental results show that linear kernel based SVM ensembles based on the bagging method and RBF kernel based SVM ensembles with the boosting method can be the better choices for a small scale dataset, where feature selection should be performed in the data pre-processing stage. For a large scale dataset, RBF kernel based SVM ensembles based on boosting perform better than the other classifiers.

  2. [Ammonia-oxidizing archaea and their important roles in nitrogen biogeochemical cycling: a review].

    Science.gov (United States)

    Liu, Jing-Jing; Wu, Wei-Xiang; Ding, Ying; Shi, De-Zhi; Chen, Ying-Xu

    2010-08-01

    As the first step of nitrification, ammonia oxidation is the key process in global nitrogen biogeochemical cycling. So far, the autotrophic ammonia-oxidizing bacteria (AOB) in the beta- and gamma-subgroups of proteobacteria have been considered as the most important contributors to ammonia oxidation, but the recent researches indicated that ammonia-oxidizing archaea (AOA) are widely distributed in various kinds of ecosystems and quantitatively predominant, playing important roles in the global nitrogen biogeochemical cycling. This paper reviewed the morphological, physiological, and ecological characteristics and the molecular phylogenies of AOA, and compared and analyzed the differences and similarities of the ammonia monooxygenase (AMO) and its encoding genes between AOA and AOB. In addition, the potential significant roles of AOA in nitrogen biogeochemical cycling in aquatic and terrestrial ecosystems were summarized, and the future research directions of AOA in applied ecology and environmental protection were put forward.

  3. Ensemble Data Mining Methods

    Data.gov (United States)

    National Aeronautics and Space Administration — Ensemble Data Mining Methods, also known as Committee Methods or Model Combiners, are machine learning methods that leverage the power of multiple models to achieve...

  4. Ensemble Network Architecture for Deep Reinforcement Learning

    Directory of Open Access Journals (Sweden)

    Xi-liang Chen

    2018-01-01

    Full Text Available The popular deep Q learning algorithm is known to be instability because of the Q-value’s shake and overestimation action values under certain conditions. These issues tend to adversely affect their performance. In this paper, we develop the ensemble network architecture for deep reinforcement learning which is based on value function approximation. The temporal ensemble stabilizes the training process by reducing the variance of target approximation error and the ensemble of target values reduces the overestimate and makes better performance by estimating more accurate Q-value. Our results show that this architecture leads to statistically significant better value evaluation and more stable and better performance on several classical control tasks at OpenAI Gym environment.

  5. Estimating predictive hydrological uncertainty by dressing deterministic and ensemble forecasts; a comparison, with application to Meuse and Rhine

    Science.gov (United States)

    Verkade, J. S.; Brown, J. D.; Davids, F.; Reggiani, P.; Weerts, A. H.

    2017-12-01

    Two statistical post-processing approaches for estimation of predictive hydrological uncertainty are compared: (i) 'dressing' of a deterministic forecast by adding a single, combined estimate of both hydrological and meteorological uncertainty and (ii) 'dressing' of an ensemble streamflow forecast by adding an estimate of hydrological uncertainty to each individual streamflow ensemble member. Both approaches aim to produce an estimate of the 'total uncertainty' that captures both the meteorological and hydrological uncertainties. They differ in the degree to which they make use of statistical post-processing techniques. In the 'lumped' approach, both sources of uncertainty are lumped by post-processing deterministic forecasts using their verifying observations. In the 'source-specific' approach, the meteorological uncertainties are estimated by an ensemble of weather forecasts. These ensemble members are routed through a hydrological model and a realization of the probability distribution of hydrological uncertainties (only) is then added to each ensemble member to arrive at an estimate of the total uncertainty. The techniques are applied to one location in the Meuse basin and three locations in the Rhine basin. Resulting forecasts are assessed for their reliability and sharpness, as well as compared in terms of multiple verification scores including the relative mean error, Brier Skill Score, Mean Continuous Ranked Probability Skill Score, Relative Operating Characteristic Score and Relative Economic Value. The dressed deterministic forecasts are generally more reliable than the dressed ensemble forecasts, but the latter are sharper. On balance, however, they show similar quality across a range of verification metrics, with the dressed ensembles coming out slightly better. Some additional analyses are suggested. Notably, these include statistical post-processing of the meteorological forecasts in order to increase their reliability, thus increasing the reliability

  6. Selecting a climate model subset to optimise key ensemble properties

    Directory of Open Access Journals (Sweden)

    N. Herger

    2018-02-01

    Full Text Available End users studying impacts and risks caused by human-induced climate change are often presented with large multi-model ensembles of climate projections whose composition and size are arbitrarily determined. An efficient and versatile method that finds a subset which maintains certain key properties from the full ensemble is needed, but very little work has been done in this area. Therefore, users typically make their own somewhat subjective subset choices and commonly use the equally weighted model mean as a best estimate. However, different climate model simulations cannot necessarily be regarded as independent estimates due to the presence of duplicated code and shared development history. Here, we present an efficient and flexible tool that makes better use of the ensemble as a whole by finding a subset with improved mean performance compared to the multi-model mean while at the same time maintaining the spread and addressing the problem of model interdependence. Out-of-sample skill and reliability are demonstrated using model-as-truth experiments. This approach is illustrated with one set of optimisation criteria but we also highlight the flexibility of cost functions, depending on the focus of different users. The technique is useful for a range of applications that, for example, minimise present-day bias to obtain an accurate ensemble mean, reduce dependence in ensemble spread, maximise future spread, ensure good performance of individual models in an ensemble, reduce the ensemble size while maintaining important ensemble characteristics, or optimise several of these at the same time. As in any calibration exercise, the final ensemble is sensitive to the metric, observational product, and pre-processing steps used.

  7. Selecting a climate model subset to optimise key ensemble properties

    Science.gov (United States)

    Herger, Nadja; Abramowitz, Gab; Knutti, Reto; Angélil, Oliver; Lehmann, Karsten; Sanderson, Benjamin M.

    2018-02-01

    End users studying impacts and risks caused by human-induced climate change are often presented with large multi-model ensembles of climate projections whose composition and size are arbitrarily determined. An efficient and versatile method that finds a subset which maintains certain key properties from the full ensemble is needed, but very little work has been done in this area. Therefore, users typically make their own somewhat subjective subset choices and commonly use the equally weighted model mean as a best estimate. However, different climate model simulations cannot necessarily be regarded as independent estimates due to the presence of duplicated code and shared development history. Here, we present an efficient and flexible tool that makes better use of the ensemble as a whole by finding a subset with improved mean performance compared to the multi-model mean while at the same time maintaining the spread and addressing the problem of model interdependence. Out-of-sample skill and reliability are demonstrated using model-as-truth experiments. This approach is illustrated with one set of optimisation criteria but we also highlight the flexibility of cost functions, depending on the focus of different users. The technique is useful for a range of applications that, for example, minimise present-day bias to obtain an accurate ensemble mean, reduce dependence in ensemble spread, maximise future spread, ensure good performance of individual models in an ensemble, reduce the ensemble size while maintaining important ensemble characteristics, or optimise several of these at the same time. As in any calibration exercise, the final ensemble is sensitive to the metric, observational product, and pre-processing steps used.

  8. Connections between physical, optical and biogeochemical processes in the Pacific Ocean

    Science.gov (United States)

    Xiu, Peng; Chai, Fei

    2014-03-01

    A new biogeochemical model has been developed and coupled to a three-dimensional physical model in the Pacific Ocean. With the explicitly represented dissolved organic pools, this new model is able to link key biogeochemical processes with optical processes. Model validation against satellite and in situ data indicates the model is robust in reproducing general biogeochemical and optical features. Colored dissolved organic matter (CDOM) has been suggested to play an important role in regulating underwater light field. With the coupled model, physical and biological regulations of CDOM in the euphotic zone are analyzed. Model results indicate seasonal variability of CDOM is mostly determined by biological processes, while the importance of physical regulation manifests in the annual mean terms. Without CDOM attenuating light, modeled depth-integrated primary production is about 10% higher than the control run when averaged over the entire basin, while this discrepancy is highly variable in space with magnitudes reaching higher than 100% in some locations. With CDOM dynamics integrated in physical-biological interactions, a new mechanism by which physical processes affect biological processes is suggested, namely, physical transport of CDOM changes water optical properties, which can further modify underwater light field and subsequently affect the distribution of phytoplankton chlorophyll. This mechanism tends to occur in the entire Pacific basin but with strong spatial variability, implying the importance of including optical processes in the coupled physical-biogeochemical model. If ammonium uptake is sufficient to permit utilization of DOM, that is, UB∗⩾-U{U}/{U}-{(1-r_b)}/{RB}, then bacteria uptake of DOM has the form of FB=(1-r_b){U}/{RB}, bacteria respiration, SB=r_b×U, remineralization by bacteria, EB=UC{UN}/{UC}-{(1-r_b)}/{RB}. If EB > 0, then UB = 0; otherwise, UB = -EB. If there is insufficient ammonium, that is, UB∗CO2 is calculated using the

  9. Ecohydrological Interfaces as Dynamic Hotspots of Biogeochemical Cycling

    Science.gov (United States)

    Krause, Stefan; Lewandowski, Joerg; Hannah, David; McDonald, Karlie; Folegot, Silvia; Baranov, Victor

    2016-04-01

    Ecohydrological interfaces, represent the boundaries between water-dependent ecosystems that can alter substantially the fluxes of energy and matter. There is still a critical gap of understanding the organisational principles of the drivers and controls of spatially and temporally variable ecohydrological interface functions. This knowledge gap limits our capacity to efficiently quantify, predict and manage the services provided by complex ecosystems. Many ecohydrological interfaces are characterized by step changes in microbial metabolic activity, steep redox gradients and often even thermodynamic phase shifts, for instance at the interfaces between atmosphere and water or soil matrix and macro-pores interfaces. This paper integrates investigations from point scale laboratory microcosm experiments with reach and subcatchment scale tracer experiments and numerical modeling studies to elaborate similarities in the drivers and controls that constitute the enhanced biogeochemical activity of different types of ecohydrologica interfaces across a range of spatial and temporal scales. We therefore combine smart metabolic activity tracers to quantify the impact of bioturbating benthic fauna onto ecosystem respiration and oxygen consumption and investigate at larger scale, how microbial metabolic activity and carbon turnover at the water-sediment interface are controlled by sediment physical and chemical properties as well as water temperatures. Numerical modeling confirmed that experimentally identified hotspots of streambed biogeochemical cycling were controlled by patterns of physical properties such as hydraulic conductivities or bioavailability of organic matter, impacting on residence time distributions and hence reaction times. In contrast to previous research, our investigations thus confirmed that small-scale variability of physical and chemical interface properties had a major impact on biogeochemical processing at the investigated ecohydrological interfaces

  10. Bayesian ensemble approach to error estimation of interatomic potentials

    DEFF Research Database (Denmark)

    Frederiksen, Søren Lund; Jacobsen, Karsten Wedel; Brown, K.S.

    2004-01-01

    Using a Bayesian approach a general method is developed to assess error bars on predictions made by models fitted to data. The error bars are estimated from fluctuations in ensembles of models sampling the model-parameter space with a probability density set by the minimum cost. The method...... is applied to the development of interatomic potentials for molybdenum using various potential forms and databases based on atomic forces. The calculated error bars on elastic constants, gamma-surface energies, structural energies, and dislocation properties are shown to provide realistic estimates...

  11. On evaluation of ensemble precipitation forecasts with observation-based ensembles

    Directory of Open Access Journals (Sweden)

    S. Jaun

    2007-04-01

    Full Text Available Spatial interpolation of precipitation data is uncertain. How important is this uncertainty and how can it be considered in evaluation of high-resolution probabilistic precipitation forecasts? These questions are discussed by experimental evaluation of the COSMO consortium's limited-area ensemble prediction system COSMO-LEPS. The applied performance measure is the often used Brier skill score (BSS. The observational references in the evaluation are (a analyzed rain gauge data by ordinary Kriging and (b ensembles of interpolated rain gauge data by stochastic simulation. This permits the consideration of either a deterministic reference (the event is observed or not with 100% certainty or a probabilistic reference that makes allowance for uncertainties in spatial averaging. The evaluation experiments show that the evaluation uncertainties are substantial even for the large area (41 300 km2 of Switzerland with a mean rain gauge distance as good as 7 km: the one- to three-day precipitation forecasts have skill decreasing with forecast lead time but the one- and two-day forecast performances differ not significantly.

  12. Polarized ensembles of random pure states

    Science.gov (United States)

    Deelan Cunden, Fabio; Facchi, Paolo; Florio, Giuseppe

    2013-08-01

    A new family of polarized ensembles of random pure states is presented. These ensembles are obtained by linear superposition of two random pure states with suitable distributions, and are quite manageable. We will use the obtained results for two purposes: on the one hand we will be able to derive an efficient strategy for sampling states from isopurity manifolds. On the other, we will characterize the deviation of a pure quantum state from separability under the influence of noise.

  13. Calculations of light scattering matrices for stochastic ensembles of nanosphere clusters

    International Nuclear Information System (INIS)

    Bunkin, N.F.; Shkirin, A.V.; Suyazov, N.V.; Starosvetskiy, A.V.

    2013-01-01

    Results of the calculation of the light scattering matrices for systems of stochastic nanosphere clusters are presented. A mathematical model of spherical particle clustering with allowance for cluster–cluster aggregation is used. The fractal properties of cluster structures are explored at different values of the model parameter that governs cluster–cluster interaction. General properties of the light scattering matrices of nanosphere-cluster ensembles as dependent on their mean fractal dimension have been found. The scattering-matrix calculations were performed for finite samples of 10 3 random clusters, made up of polydisperse spherical nanoparticles, having lognormal size distribution with the effective radius 50 nm and effective variance 0.02; the mean number of monomers in a cluster and its standard deviation were set to 500 and 70, respectively. The implemented computation environment, modeling the scattering matrices for overall sequences of clusters, is based upon T-matrix program code for a given single cluster of spheres, which was developed in [1]. The ensemble-averaged results have been compared with orientation-averaged ones calculated for individual clusters. -- Highlights: ► We suggested a hierarchical model of cluster growth allowing for cluster–cluster aggregation. ► We analyzed the light scattering by whole ensembles of nanosphere clusters. ► We studied the evolution of the light scattering matrix when changing the fractal dimension

  14. Mangrove forests: a potent nexus of coastal biogeochemical cycling

    Science.gov (United States)

    Barr, J. G.; Fuentes, J. D.; Shoemaker, B.; O'Halloran, T. L.; Lin, G., Sr.; Engel, V. C.

    2014-12-01

    Mangrove forests cover just 0.1% of the Earth's terrestrial surface, yet they provide a disproportionate source (~10 % globally) of terrestrially derived, refractory dissolved organic carbon to the oceans. Mangrove forests are biogeochemical reactors that convert biomass into dissolved organic and inorganic carbon at unusually high rates, and many studies recognize the value of mangrove ecosystems for the substantial amounts of soil carbon storage they produce. However, questions remain as to how mangrove forest ecosystem services should be valuated and quantified. Therefore, this study addresses several objectives. First, we demonstrate that seasonal and annual net ecosystem carbon exchange in three selected mangrove forests, derived from long-term eddy covariance measurements, represent key quantities in defining the magnitude of biogeochemical cycling and together with other information on carbon cycle parameters serves as a proxy to estimate ecosystem services. Second, we model ecosystem productivity across the mangrove forests of Everglades National Park and southern China by relating net ecosystem exchange values to remote sensing data. Finally, we develop a carbon budget for the mangrove forests in the Everglades National Park for the purposes of demonstrating that these forests and adjacent estuaries are sites of intense biogeochemical cycling. One conclusion from this study is that much of the carbon entering from the atmosphere as net ecosystem exchange (~1000 g C m-2 yr-1) is not retained in the net ecosystem carbon balance. Instead, a substantial fraction of the carbon entering the system as net ecosystem exchange is ultimately exported to the oceans or outgassed as reaction products within the adjacent estuary.

  15. Semi-Supervised Multi-View Ensemble Learning Based On Extracting Cross-View Correlation

    Directory of Open Access Journals (Sweden)

    ZALL, R.

    2016-05-01

    Full Text Available Correlated information between different views incorporate useful for learning in multi view data. Canonical correlation analysis (CCA plays important role to extract these information. However, CCA only extracts the correlated information between paired data and cannot preserve correlated information between within-class samples. In this paper, we propose a two-view semi-supervised learning method called semi-supervised random correlation ensemble base on spectral clustering (SS_RCE. SS_RCE uses a multi-view method based on spectral clustering which takes advantage of discriminative information in multiple views to estimate labeling information of unlabeled samples. In order to enhance discriminative power of CCA features, we incorporate the labeling information of both unlabeled and labeled samples into CCA. Then, we use random correlation between within-class samples from cross view to extract diverse correlated features for training component classifiers. Furthermore, we extend a general model namely SSMV_RCE to construct ensemble method to tackle semi-supervised learning in the presence of multiple views. Finally, we compare the proposed methods with existing multi-view feature extraction methods using multi-view semi-supervised ensembles. Experimental results on various multi-view data sets are presented to demonstrate the effectiveness of the proposed methods.

  16. Biogeochemical response to widespread anoxia in the past ocean

    NARCIS (Netherlands)

    Ruvalcaba Baroni, I.

    2015-01-01

    Oxygen is a key element for life on earth. Oxygen concentrations in the ocean vary greatly in space and time. These changes are regulated by various physical and biogeochemical processes, such as primary productivity, sea surface temperatures and ocean circulation. In the geological past, several

  17. Developing an Ensemble Prediction System based on COSMO-DE

    Science.gov (United States)

    Theis, S.; Gebhardt, C.; Buchhold, M.; Ben Bouallègue, Z.; Ohl, R.; Paulat, M.; Peralta, C.

    2010-09-01

    The numerical weather prediction model COSMO-DE is a configuration of the COSMO model with a horizontal grid size of 2.8 km. It has been running operationally at DWD since 2007, it covers the area of Germany and produces forecasts with a lead time of 0-21 hours. The model COSMO-DE is convection-permitting, which means that it does without a parametrisation of deep convection and simulates deep convection explicitly. One aim is an improved forecast of convective heavy rain events. Convection-permitting models are in operational use at several weather services, but currently not in ensemble mode. It is expected that an ensemble system could reveal the advantages of a convection-permitting model even better. The probabilistic approach is necessary, because the explicit simulation of convective processes for more than a few hours cannot be viewed as a deterministic forecast anymore. This is due to the chaotic behaviour and short life cycle of the processes which are simulated explicitly now. In the framework of the project COSMO-DE-EPS, DWD is developing and implementing an ensemble prediction system (EPS) for the model COSMO-DE. The project COSMO-DE-EPS comprises the generation of ensemble members, as well as the verification and visualization of the ensemble forecasts and also statistical postprocessing. A pre-operational mode of the EPS with 20 ensemble members is foreseen to start in 2010. Operational use is envisaged to start in 2012, after an upgrade to 40 members and inclusion of statistical postprocessing. The presentation introduces the project COSMO-DE-EPS and describes the design of the ensemble as it is planned for the pre-operational mode. In particular, the currently implemented method for the generation of ensemble members will be explained and discussed. The method includes variations of initial conditions, lateral boundary conditions, and model physics. At present, pragmatic methods are applied which resemble the basic ideas of a multi-model approach

  18. Hyporheic zone as a bioreactor: sediment heterogeneity influencing biogeochemical processes

    Science.gov (United States)

    Perujo, Nuria; Romani, Anna M.; Sanchez-Vila, Xavier

    2017-04-01

    Mediterranean fluvial systems are characterized by frequent periods of low flow or even drought. During low flow periods, water from wastewater treatment plants (WWTPs) is proportionally large in fluvial systems. River water might be vertically transported through the hyporheic zone, and then porous medium acts as a complementary treatment system since, as water infiltrates, a suite of biogeochemical processes occurs. Subsurface sediment heterogeneity plays an important role since it influences the interstitial fluxes of the medium and drives biomass growing, determining biogeochemical reactions. In this study, WWTP water was continuously infiltrated for 3 months through two porous medium tanks: one consisting of 40 cm of fine sediment (homogeneous); and another comprised of two layers of different grain size sediments (heterogeneous), 20 cm of coarse sediment in the upper part and 20 cm of fine one in the bottom. Several hydrological, physicochemical and biological parameters were measured periodically (weekly at the start of the experiment and biweekly at the end). Analysed parameters include dissolved nitrogen, phosphorus, organic carbon, and oxygen all measured at the surface, and at 5, 20 and 40 cm depth. Variations in hydraulic conductivity with time were evaluated. Sediment samples were also analysed at three depths (surface, 20 and 40 cm) to determine bacterial density, chlorophyll content, extracellular polymeric substances, and biofilm function (extracellular enzyme activities and carbon substrate utilization profiles). Preliminary results suggest hydraulic conductivity to be the main driver of the differences in the biogeochemical processes occurring in the subsurface. At the heterogeneous tank, a low nutrient reduction throughout the whole medium is measured. In this medium, high hydraulic conductivity allows for a large amount of infiltrating water, but with a small residence time. Since some biological processes are largely time-dependent, small water

  19. The Hydrologic Ensemble Prediction Experiment (HEPEX)

    Science.gov (United States)

    Wood, Andy; Wetterhall, Fredrik; Ramos, Maria-Helena

    2015-04-01

    The Hydrologic Ensemble Prediction Experiment was established in March, 2004, at a workshop hosted by the European Center for Medium Range Weather Forecasting (ECMWF), and co-sponsored by the US National Weather Service (NWS) and the European Commission (EC). The HEPEX goal was to bring the international hydrological and meteorological communities together to advance the understanding and adoption of hydrological ensemble forecasts for decision support. HEPEX pursues this goal through research efforts and practical implementations involving six core elements of a hydrologic ensemble prediction enterprise: input and pre-processing, ensemble techniques, data assimilation, post-processing, verification, and communication and use in decision making. HEPEX has grown through meetings that connect the user, forecast producer and research communities to exchange ideas, data and methods; the coordination of experiments to address specific challenges; and the formation of testbeds to facilitate shared experimentation. In the last decade, HEPEX has organized over a dozen international workshops, as well as sessions at scientific meetings (including AMS, AGU and EGU) and special issues of scientific journals where workshop results have been published. Through these interactions and an active online blog (www.hepex.org), HEPEX has built a strong and active community of nearly 400 researchers & practitioners around the world. This poster presents an overview of recent and planned HEPEX activities, highlighting case studies that exemplify the focus and objectives of HEPEX.

  20. Ensemble-based forecasting at Horns Rev: Ensemble conversion and kernel dressing

    DEFF Research Database (Denmark)

    Pinson, Pierre; Madsen, Henrik

    . The obtained ensemble forecasts of wind power are then converted into predictive distributions with an original adaptive kernel dressing method. The shape of the kernels is driven by a mean-variance model, the parameters of which are recursively estimated in order to maximize the overall skill of obtained...

  1. Lattice gauge theory in the microcanonical ensemble

    International Nuclear Information System (INIS)

    Callaway, D.J.E.; Rahman, A.

    1983-01-01

    The microcanonical-ensemble formulation of lattice gauge theory proposed recently is examined in detail. Expectation values in this new ensemble are determined by solving a large set of coupled ordinary differential equations, after the fashion of a molecular dynamics simulation. Following a brief review of the microcanonical ensemble, calculations are performed for the gauge groups U(1), SU(2), and SU(3). The results are compared and contrasted with standard methods of computation. Several advantages of the new formalism are noted. For example, no random numbers are required to update the system. Also, this update is performed in a simultaneous fashion. Thus the microcanonical method presumably adapts well to parallel processing techniques, especially when the p action is highly nonlocal (such as when fermions are included)

  2. An artificial neural network ensemble method for fault diagnosis of proton exchange membrane fuel cell system

    International Nuclear Information System (INIS)

    Shao, Meng; Zhu, Xin-Jian; Cao, Hong-Fei; Shen, Hai-Feng

    2014-01-01

    The commercial viability of PEMFC (proton exchange membrane fuel cell) systems depends on using effective fault diagnosis technologies in PEMFC systems. However, many researchers have experimentally studied PEMFC (proton exchange membrane fuel cell) systems without considering certain fault conditions. In this paper, an ANN (artificial neural network) ensemble method is presented that improves the stability and reliability of the PEMFC systems. In the first part, a transient model giving it flexibility in application to some exceptional conditions is built. The PEMFC dynamic model is built and simulated using MATLAB. In the second, using this model and experiments, the mechanisms of four different faults in PEMFC systems are analyzed in detail. Third, the ANN ensemble for the fault diagnosis is built and modeled. This model is trained and tested by the data. The test result shows that, compared with the previous method for fault diagnosis of PEMFC systems, the proposed fault diagnosis method has higher diagnostic rate and generalization ability. Moreover, the partial structure of this method can be altered easily, along with the change of the PEMFC systems. In general, this method for diagnosis of PEMFC has value for certain applications. - Highlights: • We analyze the principles and mechanisms of the four faults in PEMFC (proton exchange membrane fuel cell) system. • We design and model an ANN (artificial neural network) ensemble method for the fault diagnosis of PEMFC system. • This method has high diagnostic rate and strong generalization ability

  3. Polarized ensembles of random pure states

    International Nuclear Information System (INIS)

    Cunden, Fabio Deelan; Facchi, Paolo; Florio, Giuseppe

    2013-01-01

    A new family of polarized ensembles of random pure states is presented. These ensembles are obtained by linear superposition of two random pure states with suitable distributions, and are quite manageable. We will use the obtained results for two purposes: on the one hand we will be able to derive an efficient strategy for sampling states from isopurity manifolds. On the other, we will characterize the deviation of a pure quantum state from separability under the influence of noise. (paper)

  4. Probabilistic Determination of Native State Ensembles of Proteins

    DEFF Research Database (Denmark)

    Olsson, Simon; Vögeli, Beat Rolf; Cavalli, Andrea

    2014-01-01

    ensembles of proteins by the combination of physical force fields and experimental data through modern statistical methodology. As an example, we use NMR residual dipolar couplings to determine a native state ensemble of the extensively studied third immunoglobulin binding domain of protein G (GB3...

  5. Statistical hadronization and hadronic micro-canonical ensemble II

    International Nuclear Information System (INIS)

    Becattini, F.; Ferroni, L.

    2004-01-01

    We present a Monte Carlo calculation of the micro-canonical ensemble of the ideal hadron-resonance gas including all known states up to a mass of about 1.8 GeV and full quantum statistics. The micro-canonical average multiplicities of the various hadron species are found to converge to the canonical ones for moderately low values of the total energy, around 8 GeV, thus bearing out previous analyses of hadronic multiplicities in the canonical ensemble. The main numerical computing method is an importance sampling Monte Carlo algorithm using the product of Poisson distributions to generate multi-hadronic channels. It is shown that the use of this multi-Poisson distribution allows for an efficient and fast computation of averages, which can be further improved in the limit of very large clusters. We have also studied the fitness of a previously proposed computing method, based on the Metropolis Monte Carlo algorithm, for event generation in the statistical hadronization model. We find that the use of the multi-Poisson distribution as proposal matrix dramatically improves the computation performance. However, due to the correlation of subsequent samples, this method proves to be generally less robust and effective than the importance sampling method. (orig.)

  6. Ensembles of a small number of conformations with relative populations

    Energy Technology Data Exchange (ETDEWEB)

    Vammi, Vijay, E-mail: vsvammi@iastate.edu; Song, Guang, E-mail: gsong@iastate.edu [Iowa State University, Bioinformatics and Computational Biology Program, Department of Computer Science (United States)

    2015-12-15

    In our previous work, we proposed a new way to represent protein native states, using ensembles of a small number of conformations with relative Populations, or ESP in short. Using Ubiquitin as an example, we showed that using a small number of conformations could greatly reduce the potential of overfitting and assigning relative populations to protein ensembles could significantly improve their quality. To demonstrate that ESP indeed is an excellent alternative to represent protein native states, in this work we compare the quality of two ESP ensembles of Ubiquitin with several well-known regular ensembles or average structure representations. Extensive amount of significant experimental data are employed to achieve a thorough assessment. Our results demonstrate that ESP ensembles, though much smaller in size comparing to regular ensembles, perform equally or even better sometimes in all four different types of experimental data used in the assessment, namely, the residual dipolar couplings, residual chemical shift anisotropy, hydrogen exchange rates, and solution scattering profiles. This work further underlines the significance of having relative populations in describing the native states.

  7. On the proper use of Ensembles for Predictive Uncertainty assessment

    Science.gov (United States)

    Todini, Ezio; Coccia, Gabriele; Ortiz, Enrique

    2015-04-01

    Probabilistic forecasting has become popular in the last decades. Hydrological probabilistic forecasts have been based either on uncertainty processors (Krzysztofowic, 1999; Todini, 2004; Todini, 2008) or on ensembles, following meteorological traditional approaches and the establishment of the HEPEX program (http://hepex.irstea.fr. Unfortunately, the direct use of ensembles as a measure of the predictive density is an incorrect practice, because the ensemble measures the spread of the forecast instead of, following the definition of predictive uncertainty, the conditional probability of the future outcome conditional on the forecast. Only few correct approaches are reported in the literature, which correctly use the ensemble to estimate an expected conditional predictive density (Reggiani et al., 2009), similarly to what is done when several predictive models are available as in the BMA (Raftery et al., 2005) or MCP(Todini, 2008; Coccia and Todini, 2011) approaches. A major problem, limiting the correct use of ensembles, is in fact the difficulty of defining the time dependence of the ensemble members, due to the lack of a consistent ranking: in other words, when dealing with multiple models, the ith model remains the ith model regardless to the time of forecast, while this does not happen when dealing with ensemble members, since there is no definition for the ith member of an ensemble. Nonetheless, the MCP approach (Todini, 2008; Coccia and Todini, 2011), essentially based on a multiple regression in the Normal space, can be easily extended to use ensembles to represent the local (in time) smaller or larger conditional predictive uncertainty, as a function of the ensemble spread. This is done by modifying the classical linear regression equations, impliying perfectly observed predictors, to alternative regression equations similar to the Kalman filter ones, allowing for uncertain predictors. In this way, each prediction in time accounts for both the predictive

  8. Ensemble Kalman filtering with one-step-ahead smoothing

    KAUST Repository

    Raboudi, Naila F.

    2018-01-11

    The ensemble Kalman filter (EnKF) is widely used for sequential data assimilation. It operates as a succession of forecast and analysis steps. In realistic large-scale applications, EnKFs are implemented with small ensembles and poorly known model error statistics. This limits their representativeness of the background error covariances and, thus, their performance. This work explores the efficiency of the one-step-ahead (OSA) smoothing formulation of the Bayesian filtering problem to enhance the data assimilation performance of EnKFs. Filtering with OSA smoothing introduces an updated step with future observations, conditioning the ensemble sampling with more information. This should provide an improved background ensemble in the analysis step, which may help to mitigate the suboptimal character of EnKF-based methods. Here, the authors demonstrate the efficiency of a stochastic EnKF with OSA smoothing for state estimation. They then introduce a deterministic-like EnKF-OSA based on the singular evolutive interpolated ensemble Kalman (SEIK) filter. The authors show that the proposed SEIK-OSA outperforms both SEIK, as it efficiently exploits the data twice, and the stochastic EnKF-OSA, as it avoids observational error undersampling. They present extensive assimilation results from numerical experiments conducted with the Lorenz-96 model to demonstrate SEIK-OSA’s capabilities.

  9. Quadripartite cluster and Greenberger–Horne–Zeilinger entangled light via cascade interactions with separated atomic ensembles

    International Nuclear Information System (INIS)

    Li Xing; Hu Xiangming

    2012-01-01

    It has been known that two-mode entangled light can possibly be generated by employing near-resonant interaction with an ensemble of two-level atoms. The responsible mechanism is the absorption of two photons from the strong driving field and the emission of two new photons into the cavity field. Here, we generalize such a mechanism to three separated atomic ensembles and establish cascade interactions for four nondegenerate fields. It is shown that the quadripartite cluster and Greenberger–Horne–Zeilinger entangled states occur for continuous variables. The advantage of the present scheme for the multipartite entanglement lies in that the coupling strengths are much larger due to the near resonances than for far-off-resonance-based parametric processes. (paper)

  10. A novel hybrid ensemble learning paradigm for nuclear energy consumption forecasting

    International Nuclear Information System (INIS)

    Tang, Ling; Yu, Lean; Wang, Shuai; Li, Jianping; Wang, Shouyang

    2012-01-01

    Highlights: ► A hybrid ensemble learning paradigm integrating EEMD and LSSVR is proposed. ► The hybrid ensemble method is useful to predict time series with high volatility. ► The ensemble method can be used for both one-step and multi-step ahead forecasting. - Abstract: In this paper, a novel hybrid ensemble learning paradigm integrating ensemble empirical mode decomposition (EEMD) and least squares support vector regression (LSSVR) is proposed for nuclear energy consumption forecasting, based on the principle of “decomposition and ensemble”. This hybrid ensemble learning paradigm is formulated specifically to address difficulties in modeling nuclear energy consumption, which has inherently high volatility, complexity and irregularity. In the proposed hybrid ensemble learning paradigm, EEMD, as a competitive decomposition method, is first applied to decompose original data of nuclear energy consumption (i.e. a difficult task) into a number of independent intrinsic mode functions (IMFs) of original data (i.e. some relatively easy subtasks). Then LSSVR, as a powerful forecasting tool, is implemented to predict all extracted IMFs independently. Finally, these predicted IMFs are aggregated into an ensemble result as final prediction, using another LSSVR. For illustration and verification purposes, the proposed learning paradigm is used to predict nuclear energy consumption in China. Empirical results demonstrate that the novel hybrid ensemble learning paradigm can outperform some other popular forecasting models in both level prediction and directional forecasting, indicating that it is a promising tool to predict complex time series with high volatility and irregularity.

  11. Constructing Better Classifier Ensemble Based on Weighted Accuracy and Diversity Measure

    Directory of Open Access Journals (Sweden)

    Xiaodong Zeng

    2014-01-01

    Full Text Available A weighted accuracy and diversity (WAD method is presented, a novel measure used to evaluate the quality of the classifier ensemble, assisting in the ensemble selection task. The proposed measure is motivated by a commonly accepted hypothesis; that is, a robust classifier ensemble should not only be accurate but also different from every other member. In fact, accuracy and diversity are mutual restraint factors; that is, an ensemble with high accuracy may have low diversity, and an overly diverse ensemble may negatively affect accuracy. This study proposes a method to find the balance between accuracy and diversity that enhances the predictive ability of an ensemble for unknown data. The quality assessment for an ensemble is performed such that the final score is achieved by computing the harmonic mean of accuracy and diversity, where two weight parameters are used to balance them. The measure is compared to two representative measures, Kappa-Error and GenDiv, and two threshold measures that consider only accuracy or diversity, with two heuristic search algorithms, genetic algorithm, and forward hill-climbing algorithm, in ensemble selection tasks performed on 15 UCI benchmark datasets. The empirical results demonstrate that the WAD measure is superior to others in most cases.

  12. Visualization and classification of physiological failure modes in ensemble hemorrhage simulation

    Science.gov (United States)

    Zhang, Song; Pruett, William Andrew; Hester, Robert

    2015-01-01

    In an emergency situation such as hemorrhage, doctors need to predict which patients need immediate treatment and care. This task is difficult because of the diverse response to hemorrhage in human population. Ensemble physiological simulations provide a means to sample a diverse range of subjects and may have a better chance of containing the correct solution. However, to reveal the patterns and trends from the ensemble simulation is a challenging task. We have developed a visualization framework for ensemble physiological simulations. The visualization helps users identify trends among ensemble members, classify ensemble member into subpopulations for analysis, and provide prediction to future events by matching a new patient's data to existing ensembles. We demonstrated the effectiveness of the visualization on simulated physiological data. The lessons learned here can be applied to clinically-collected physiological data in the future.

  13. Relation between native ensembles and experimental structures of proteins

    DEFF Research Database (Denmark)

    Best, R. B.; Lindorff-Larsen, Kresten; DePristo, M. A.

    2006-01-01

    Different experimental structures of the same protein or of proteins with high sequence similarity contain many small variations. Here we construct ensembles of "high-sequence similarity Protein Data Bank" (HSP) structures and consider the extent to which such ensembles represent the structural...... Data Bank ensembles; moreover, we show that the effects of uncertainties in structure determination are insufficient to explain the results. These results highlight the importance of accounting for native-state protein dynamics in making comparisons with ensemble-averaged experimental data and suggest...... heterogeneity of the native state in solution. We find that different NMR measurements probing structure and dynamics of given proteins in solution, including order parameters, scalar couplings, and residual dipolar couplings, are remarkably well reproduced by their respective high-sequence similarity Protein...

  14. On the structure and phase transitions of power-law Poissonian ensembles

    Science.gov (United States)

    Eliazar, Iddo; Oshanin, Gleb

    2012-10-01

    Power-law Poissonian ensembles are Poisson processes that are defined on the positive half-line, and that are governed by power-law intensities. Power-law Poissonian ensembles are stochastic objects of fundamental significance; they uniquely display an array of fractal features and they uniquely generate a span of important applications. In this paper we apply three different methods—oligarchic analysis, Lorenzian analysis and heterogeneity analysis—to explore power-law Poissonian ensembles. The amalgamation of these analyses, combined with the topology of power-law Poissonian ensembles, establishes a detailed and multi-faceted picture of the statistical structure and the statistical phase transitions of these elemental ensembles.

  15. Dissipation induced asymmetric steering of distant atomic ensembles

    Science.gov (United States)

    Cheng, Guangling; Tan, Huatang; Chen, Aixi

    2018-04-01

    The asymmetric steering effects of separated atomic ensembles denoted by the effective bosonic modes have been explored by the means of quantum reservoir engineering in the setting of the cascaded cavities, in each of which an atomic ensemble is involved. It is shown that the steady-state asymmetric steering of the mesoscopic objects is unconditionally achieved via the dissipation of the cavities, by which the nonlocal interaction occurs between two atomic ensembles, and the direction of steering could be easily controlled through variation of certain tunable system parameters. One advantage of the present scheme is that it could be rather robust against parameter fluctuations, and does not require the accurate control of evolution time and the original state of the system. Furthermore, the double-channel Raman transitions between the long-lived atomic ground states are used and the atomic ensembles act as the quantum network nodes, which makes our scheme insensitive to the collective spontaneous emission of atoms.

  16. Skill forecasting from different wind power ensemble prediction methods

    International Nuclear Information System (INIS)

    Pinson, Pierre; Nielsen, Henrik A; Madsen, Henrik; Kariniotakis, George

    2007-01-01

    This paper presents an investigation on alternative approaches to the providing of uncertainty estimates associated to point predictions of wind generation. Focus is given to skill forecasts in the form of prediction risk indices, aiming at giving a comprehensive signal on the expected level of forecast uncertainty. Ensemble predictions of wind generation are used as input. A proposal for the definition of prediction risk indices is given. Such skill forecasts are based on the dispersion of ensemble members for a single prediction horizon, or over a set of successive look-ahead times. It is shown on the test case of a Danish offshore wind farm how prediction risk indices may be related to several levels of forecast uncertainty (and energy imbalances). Wind power ensemble predictions are derived from the transformation of ECMWF and NCEP ensembles of meteorological variables to power, as well as by a lagged average approach alternative. The ability of risk indices calculated from the various types of ensembles forecasts to resolve among situations with different levels of uncertainty is discussed

  17. Using ensemble forecasting for wind power

    Energy Technology Data Exchange (ETDEWEB)

    Giebel, G.; Landberg, L.; Badger, J. [Risoe National Lab., Roskilde (Denmark); Sattler, K.

    2003-07-01

    Short-term prediction of wind power has a long tradition in Denmark. It is an essential tool for the operators to keep the grid from becoming unstable in a region like Jutland, where more than 27% of the electricity consumption comes from wind power. This means that the minimum load is already lower than the maximum production from wind energy alone. Danish utilities have therefore used short-term prediction of wind energy since the mid-90ies. However, the accuracy is still far from being sufficient in the eyes of the utilities (used to have load forecasts accurate to within 5% on a one-week horizon). The Ensemble project tries to alleviate the dependency of the forecast quality on one model by using multiple models, and also will investigate the possibilities of using the model spread of multiple models or of dedicated ensemble runs for a prediction of the uncertainty of the forecast. Usually, short-term forecasting works (especially for the horizon beyond 6 hours) by gathering input from a Numerical Weather Prediction (NWP) model. This input data is used together with online data in statistical models (this is the case eg in Zephyr/WPPT) to yield the output of the wind farms or of a whole region for the next 48 hours (only limited by the NWP model horizon). For the accuracy of the final production forecast, the accuracy of the NWP prediction is paramount. While many efforts are underway to increase the accuracy of the NWP forecasts themselves (which ultimately are limited by the amount of computing power available, the lack of a tight observational network on the Atlantic and limited physics modelling), another approach is to use ensembles of different models or different model runs. This can be either an ensemble of different models output for the same area, using different data assimilation schemes and different model physics, or a dedicated ensemble run by a large institution, where the same model is run with slight variations in initial conditions and

  18. Operational hydrological forecasting in Bavaria. Part II: Ensemble forecasting

    Science.gov (United States)

    Ehret, U.; Vogelbacher, A.; Moritz, K.; Laurent, S.; Meyer, I.; Haag, I.

    2009-04-01

    In part I of this study, the operational flood forecasting system in Bavaria and an approach to identify and quantify forecast uncertainty was introduced. The approach is split into the calculation of an empirical 'overall error' from archived forecasts and the calculation of an empirical 'model error' based on hydrometeorological forecast tests, where rainfall observations were used instead of forecasts. The 'model error' can especially in upstream catchments where forecast uncertainty is strongly dependent on the current predictability of the atrmosphere be superimposed on the spread of a hydrometeorological ensemble forecast. In Bavaria, two meteorological ensemble prediction systems are currently tested for operational use: the 16-member COSMO-LEPS forecast and a poor man's ensemble composed of DWD GME, DWD Cosmo-EU, NCEP GFS, Aladin-Austria, MeteoSwiss Cosmo-7. The determination of the overall forecast uncertainty is dependent on the catchment characteristics: 1. Upstream catchment with high influence of weather forecast a) A hydrological ensemble forecast is calculated using each of the meteorological forecast members as forcing. b) Corresponding to the characteristics of the meteorological ensemble forecast, each resulting forecast hydrograph can be regarded as equally likely. c) The 'model error' distribution, with parameters dependent on hydrological case and lead time, is added to each forecast timestep of each ensemble member d) For each forecast timestep, the overall (i.e. over all 'model error' distribution of each ensemble member) error distribution is calculated e) From this distribution, the uncertainty range on a desired level (here: the 10% and 90% percentile) is extracted and drawn as forecast envelope. f) As the mean or median of an ensemble forecast does not necessarily exhibit meteorologically sound temporal evolution, a single hydrological forecast termed 'lead forecast' is chosen and shown in addition to the uncertainty bounds. This can be

  19. Exploring diversity in ensemble classification: Applications in large area land cover mapping

    Science.gov (United States)

    Mellor, Andrew; Boukir, Samia

    2017-07-01

    Ensemble classifiers, such as random forests, are now commonly applied in the field of remote sensing, and have been shown to perform better than single classifier systems, resulting in reduced generalisation error. Diversity across the members of ensemble classifiers is known to have a strong influence on classification performance - whereby classifier errors are uncorrelated and more uniformly distributed across ensemble members. The relationship between ensemble diversity and classification performance has not yet been fully explored in the fields of information science and machine learning and has never been examined in the field of remote sensing. This study is a novel exploration of ensemble diversity and its link to classification performance, applied to a multi-class canopy cover classification problem using random forests and multisource remote sensing and ancillary GIS data, across seven million hectares of diverse dry-sclerophyll dominated public forests in Victoria Australia. A particular emphasis is placed on analysing the relationship between ensemble diversity and ensemble margin - two key concepts in ensemble learning. The main novelty of our work is on boosting diversity by emphasizing the contribution of lower margin instances used in the learning process. Exploring the influence of tree pruning on diversity is also a new empirical analysis that contributes to a better understanding of ensemble performance. Results reveal insights into the trade-off between ensemble classification accuracy and diversity, and through the ensemble margin, demonstrate how inducing diversity by targeting lower margin training samples is a means of achieving better classifier performance for more difficult or rarer classes and reducing information redundancy in classification problems. Our findings inform strategies for collecting training data and designing and parameterising ensemble classifiers, such as random forests. This is particularly important in large area

  20. On the forecast skill of a convection-permitting ensemble

    Science.gov (United States)

    Schellander-Gorgas, Theresa; Wang, Yong; Meier, Florian; Weidle, Florian; Wittmann, Christoph; Kann, Alexander

    2017-01-01

    The 2.5 km convection-permitting (CP) ensemble AROME-EPS (Applications of Research to Operations at Mesoscale - Ensemble Prediction System) is evaluated by comparison with the regional 11 km ensemble ALADIN-LAEF (Aire Limitée Adaption dynamique Développement InterNational - Limited Area Ensemble Forecasting) to show whether a benefit is provided by a CP EPS. The evaluation focuses on the abilities of the ensembles to quantitatively predict precipitation during a 3-month convective summer period over areas consisting of mountains and lowlands. The statistical verification uses surface observations and 1 km × 1 km precipitation analyses, and the verification scores involve state-of-the-art statistical measures for deterministic and probabilistic forecasts as well as novel spatial verification methods. The results show that the convection-permitting ensemble with higher-resolution AROME-EPS outperforms its mesoscale counterpart ALADIN-LAEF for precipitation forecasts. The positive impact is larger for the mountainous areas than for the lowlands. In particular, the diurnal precipitation cycle is improved in AROME-EPS, which leads to a significant improvement of scores at the concerned times of day (up to approximately one-third of the scored verification measure). Moreover, there are advantages for higher precipitation thresholds at small spatial scales, which are due to the improved simulation of the spatial structure of precipitation.

  1. Ensemble computing for the petroleum industry

    International Nuclear Information System (INIS)

    Annaratone, M.; Dossa, D.

    1995-01-01

    Computer downsizing is one of the most often used buzzwords in today's competitive business, and the petroleum industry is at the forefront of this revolution. Ensemble computing provides the key for computer downsizing with its first incarnation, i.e., workstation farms. This paper concerns the importance of increasing the productivity cycle and not just the execution time of a job. The authors introduce the concept of ensemble computing and workstation farms. The they discuss how different computing paradigms can be addressed by workstation farms

  2. Ocean Predictability and Uncertainty Forecasts Using Local Ensemble Transfer Kalman Filter (LETKF)

    Science.gov (United States)

    Wei, M.; Hogan, P. J.; Rowley, C. D.; Smedstad, O. M.; Wallcraft, A. J.; Penny, S. G.

    2017-12-01

    Ocean predictability and uncertainty are studied with an ensemble system that has been developed based on the US Navy's operational HYCOM using the Local Ensemble Transfer Kalman Filter (LETKF) technology. One of the advantages of this method is that the best possible initial analysis states for the HYCOM forecasts are provided by the LETKF which assimilates operational observations using ensemble method. The background covariance during this assimilation process is implicitly supplied with the ensemble avoiding the difficult task of developing tangent linear and adjoint models out of HYCOM with the complicated hybrid isopycnal vertical coordinate for 4D-VAR. The flow-dependent background covariance from the ensemble will be an indispensable part in the next generation hybrid 4D-Var/ensemble data assimilation system. The predictability and uncertainty for the ocean forecasts are studied initially for the Gulf of Mexico. The results are compared with another ensemble system using Ensemble Transfer (ET) method which has been used in the Navy's operational center. The advantages and disadvantages are discussed.

  3. REAL - Ensemble radar precipitation estimation for hydrology in a mountainous region

    OpenAIRE

    Germann, Urs; Berenguer Ferrer, Marc; Sempere Torres, Daniel; Zappa, Massimiliano

    2009-01-01

    An elegant solution to characterise the residual errors in radar precipitation estimates is to generate an ensemble of precipitation fields. The paper proposes a radar ensemble generator designed for usage in the Alps using LU decomposition (REAL), and presents first results from a real-time implementation coupling the radar ensemble with a semi-distributed rainfall–runoff model for flash flood modelling in a steep Alpine catchment. Each member of the radar ensemble is a possible realisati...

  4. Non-Boltzmann Ensembles and Monte Carlo Simulations

    International Nuclear Information System (INIS)

    Murthy, K. P. N.

    2016-01-01

    Boltzmann sampling based on Metropolis algorithm has been extensively used for simulating a canonical ensemble and for calculating macroscopic properties of a closed system at desired temperatures. An estimate of a mechanical property, like energy, of an equilibrium system, is made by averaging over a large number microstates generated by Boltzmann Monte Carlo methods. This is possible because we can assign a numerical value for energy to each microstate. However, a thermal property like entropy, is not easily accessible to these methods. The reason is simple. We can not assign a numerical value for entropy, to a microstate. Entropy is not a property associated with any single microstate. It is a collective property of all the microstates. Toward calculating entropy and other thermal properties, a non-Boltzmann Monte Carlo technique called Umbrella sampling was proposed some forty years ago. Umbrella sampling has since undergone several metamorphoses and we have now, multi-canonical Monte Carlo, entropic sampling, flat histogram methods, Wang-Landau algorithm etc . This class of methods generates non-Boltzmann ensembles which are un-physical. However, physical quantities can be calculated as follows. First un-weight a microstates of the entropic ensemble; then re-weight it to the desired physical ensemble. Carry out weighted average over the entropic ensemble to estimate physical quantities. In this talk I shall tell you of the most recent non- Boltzmann Monte Carlo method and show how to calculate free energy for a few systems. We first consider estimation of free energy as a function of energy at different temperatures to characterize phase transition in an hairpin DNA in the presence of an unzipping force. Next we consider free energy as a function of order parameter and to this end we estimate density of states g ( E , M ), as a function of both energy E , and order parameter M . This is carried out in two stages. We estimate g ( E ) in the first stage

  5. Greenland's glacial fjords and their role in regional biogeochemical dynamics.

    Science.gov (United States)

    Crosby, J.; Arndt, S.

    2017-12-01

    Greenland's coastal fjords serve as important pathways that connect the Greenland Ice Sheet (GrIS) and the surrounding oceans. They export seasonal glacial meltwater whilst being significant sites of primary production. These fjords are home to some of the most productive ecosystems in the world and possess high socio-economic value via fisheries. A growing number of studies have proposed the GrIS as an underappreciated yet significant source of nutrients to surrounding oceans. Acting as both transfer routes and sinks for glacial nutrient export, fjords have the potential to act as significant biogeochemical processors, yet remain underexplored. Critically, an understanding of the quantitative contribution of fjords to carbon and nutrient budgets is lacking, with large uncertainties associated with limited availability of field data and the lack of robust upscaling approaches. To close this knowledge gap we developed a coupled 2D physical-biogeochemical model of the Godthåbsfjord system, a sub-Arctic sill fjord in southwest Greenland, to quantitatively assess the impact of nutrients exported from the GrIS on fjord primary productivity and biogeochemical dynamics. Glacial meltwater is found to be a key driver of fjord-scale circulation patterns, whilst tracer simulations reveal the relative nutrient contributions from meltwater-driven upwelling and meltwater export from the GrIS. Hydrodynamic circulation patterns and freshwater transit times are explored to provide a first understanding of the glacier-fjord-ocean continuum, demonstrating the complex pattern of carbon and nutrient cycling at this critical land-ocean interface.

  6. Noodles: a tool for visualization of numerical weather model ensemble uncertainty.

    Science.gov (United States)

    Sanyal, Jibonananda; Zhang, Song; Dyer, Jamie; Mercer, Andrew; Amburn, Philip; Moorhead, Robert J

    2010-01-01

    Numerical weather prediction ensembles are routinely used for operational weather forecasting. The members of these ensembles are individual simulations with either slightly perturbed initial conditions or different model parameterizations, or occasionally both. Multi-member ensemble output is usually large, multivariate, and challenging to interpret interactively. Forecast meteorologists are interested in understanding the uncertainties associated with numerical weather prediction; specifically variability between the ensemble members. Currently, visualization of ensemble members is mostly accomplished through spaghetti plots of a single mid-troposphere pressure surface height contour. In order to explore new uncertainty visualization methods, the Weather Research and Forecasting (WRF) model was used to create a 48-hour, 18 member parameterization ensemble of the 13 March 1993 "Superstorm". A tool was designed to interactively explore the ensemble uncertainty of three important weather variables: water-vapor mixing ratio, perturbation potential temperature, and perturbation pressure. Uncertainty was quantified using individual ensemble member standard deviation, inter-quartile range, and the width of the 95% confidence interval. Bootstrapping was employed to overcome the dependence on normality in the uncertainty metrics. A coordinated view of ribbon and glyph-based uncertainty visualization, spaghetti plots, iso-pressure colormaps, and data transect plots was provided to two meteorologists for expert evaluation. They found it useful in assessing uncertainty in the data, especially in finding outliers in the ensemble run and therefore avoiding the WRF parameterizations that lead to these outliers. Additionally, the meteorologists could identify spatial regions where the uncertainty was significantly high, allowing for identification of poorly simulated storm environments and physical interpretation of these model issues.

  7. Device and Method for Gathering Ensemble Data Sets

    Science.gov (United States)

    Racette, Paul E. (Inventor)

    2014-01-01

    An ensemble detector uses calibrated noise references to produce ensemble sets of data from which properties of non-stationary processes may be extracted. The ensemble detector comprising: a receiver; a switching device coupled to the receiver, the switching device configured to selectively connect each of a plurality of reference noise signals to the receiver; and a gain modulation circuit coupled to the receiver and configured to vary a gain of the receiver based on a forcing signal; whereby the switching device selectively connects each of the plurality of reference noise signals to the receiver to produce an output signal derived from the plurality of reference noise signals and the forcing signal.

  8. Combining 2-m temperature nowcasting and short range ensemble forecasting

    Directory of Open Access Journals (Sweden)

    A. Kann

    2011-12-01

    Full Text Available During recent years, numerical ensemble prediction systems have become an important tool for estimating the uncertainties of dynamical and physical processes as represented in numerical weather models. The latest generation of limited area ensemble prediction systems (LAM-EPSs allows for probabilistic forecasts at high resolution in both space and time. However, these systems still suffer from systematic deficiencies. Especially for nowcasting (0–6 h applications the ensemble spread is smaller than the actual forecast error. This paper tries to generate probabilistic short range 2-m temperature forecasts by combining a state-of-the-art nowcasting method and a limited area ensemble system, and compares the results with statistical methods. The Integrated Nowcasting Through Comprehensive Analysis (INCA system, which has been in operation at the Central Institute for Meteorology and Geodynamics (ZAMG since 2006 (Haiden et al., 2011, provides short range deterministic forecasts at high temporal (15 min–60 min and spatial (1 km resolution. An INCA Ensemble (INCA-EPS of 2-m temperature forecasts is constructed by applying a dynamical approach, a statistical approach, and a combined dynamic-statistical method. The dynamical method takes uncertainty information (i.e. ensemble variance from the operational limited area ensemble system ALADIN-LAEF (Aire Limitée Adaptation Dynamique Développement InterNational Limited Area Ensemble Forecasting which is running operationally at ZAMG (Wang et al., 2011. The purely statistical method assumes a well-calibrated spread-skill relation and applies ensemble spread according to the skill of the INCA forecast of the most recent past. The combined dynamic-statistical approach adapts the ensemble variance gained from ALADIN-LAEF with non-homogeneous Gaussian regression (NGR which yields a statistical mbox{correction} of the first and second moment (mean bias and dispersion for Gaussian distributed continuous

  9. Modeling Dynamic Systems with Efficient Ensembles of Process-Based Models.

    Directory of Open Access Journals (Sweden)

    Nikola Simidjievski

    Full Text Available Ensembles are a well established machine learning paradigm, leading to accurate and robust models, predominantly applied to predictive modeling tasks. Ensemble models comprise a finite set of diverse predictive models whose combined output is expected to yield an improved predictive performance as compared to an individual model. In this paper, we propose a new method for learning ensembles of process-based models of dynamic systems. The process-based modeling paradigm employs domain-specific knowledge to automatically learn models of dynamic systems from time-series observational data. Previous work has shown that ensembles based on sampling observational data (i.e., bagging and boosting, significantly improve predictive performance of process-based models. However, this improvement comes at the cost of a substantial increase of the computational time needed for learning. To address this problem, the paper proposes a method that aims at efficiently learning ensembles of process-based models, while maintaining their accurate long-term predictive performance. This is achieved by constructing ensembles with sampling domain-specific knowledge instead of sampling data. We apply the proposed method to and evaluate its performance on a set of problems of automated predictive modeling in three lake ecosystems using a library of process-based knowledge for modeling population dynamics. The experimental results identify the optimal design decisions regarding the learning algorithm. The results also show that the proposed ensembles yield significantly more accurate predictions of population dynamics as compared to individual process-based models. Finally, while their predictive performance is comparable to the one of ensembles obtained with the state-of-the-art methods of bagging and boosting, they are substantially more efficient.

  10. JEnsembl: a version-aware Java API to Ensembl data systems.

    Science.gov (United States)

    Paterson, Trevor; Law, Andy

    2012-11-01

    The Ensembl Project provides release-specific Perl APIs for efficient high-level programmatic access to data stored in various Ensembl database schema. Although Perl scripts are perfectly suited for processing large volumes of text-based data, Perl is not ideal for developing large-scale software applications nor embedding in graphical interfaces. The provision of a novel Java API would facilitate type-safe, modular, object-orientated development of new Bioinformatics tools with which to access, analyse and visualize Ensembl data. The JEnsembl API implementation provides basic data retrieval and manipulation functionality from the Core, Compara and Variation databases for all species in Ensembl and EnsemblGenomes and is a platform for the development of a richer API to Ensembl datasources. The JEnsembl architecture uses a text-based configuration module to provide evolving, versioned mappings from database schema to code objects. A single installation of the JEnsembl API can therefore simultaneously and transparently connect to current and previous database instances (such as those in the public archive) thus facilitating better analysis repeatability and allowing 'through time' comparative analyses to be performed. Project development, released code libraries, Maven repository and documentation are hosted at SourceForge (http://jensembl.sourceforge.net).

  11. Dispersion of aerosol particles in the free atmosphere using ensemble forecasts

    Directory of Open Access Journals (Sweden)

    T. Haszpra

    2013-10-01

    Full Text Available The dispersion of aerosol particle pollutants is studied using 50 members of an ensemble forecast in the example of a hypothetical free atmospheric emission above Fukushima over a period of 2.5 days. Considerable differences are found among the dispersion predictions of the different ensemble members, as well as between the ensemble mean and the deterministic result at the end of the observation period. The variance is found to decrease with the particle size. The geographical area where a threshold concentration is exceeded in at least one ensemble member expands to a 5–10 times larger region than the area from the deterministic forecast, both for air column "concentration" and in the "deposition" field. We demonstrate that the root-mean-square distance of any particle from its own clones in the ensemble members can reach values on the order of one thousand kilometers. Even the centers of mass of the particle cloud of the ensemble members deviate considerably from that obtained by the deterministic forecast. All these indicate that an investigation of the dispersion of aerosol particles in the spirit of ensemble forecast contains useful hints for the improvement of risk assessment.

  12. Biogeochemical Reactions Under Simulated Europa Ocean Conditions

    Science.gov (United States)

    Amashukeli, X.; Connon, S. A.; Gleeson, D. F.; Kowalczyk, R. S.; Pappalardo, R. T.

    2007-12-01

    Galileo data have demonstrated the probable presence of a liquid water ocean on Europa, and existence of salts and carbon dioxide in the satellite's surface ice (e.g., Carr et al., 1998; McCord et al., 1999, Pappalardo et al., 1999; Kivelson et al., 2000). Subsequently, the discovery of chemical signatures of extinct or extant life in Europa's ocean and on its surface became a distinct possibility. Moreover, understanding of Europa's potential habitability is now one of the major goals of the Europa Orbiter Flagship mission. It is likely, that in the early stages of Europa's ocean formation, moderately alkaline oceanic sulfate-carbonate species and a magnetite-silicate mantel could have participated in low-temperature biogeochemical sulfur, iron and carbon cycles facilitated by primitive organisms (Zolotov and Shock, 2004). If periodic supplies of fresh rock and sulfate-carbonate ions are available in Europa's ocean, then an exciting prospect exists that life may be present in Europa's ocean today. In our laboratory, we began the study of the plausible biogeochemical reactions under conditions appropriate to Europa's ocean using barophilic psychrophilic organisms that thrive under anaerobic conditions. In the near absence of abiotic synthetic pathways due to low Europa's temperatures, the biotic synthesis may present a viable opportunity for the formation of the organic and inorganic compounds under these extreme conditions. This work is independent of assumptions regarding hydrothermal vents at Europa's ocean floor or surface-derived oxidant sources. For our studies, we have fabricated a high-pressure (5,000 psi) reaction vessel that simulates aqueous conditions on Europa. We were also successful at reviving barophilic psychrophilic strains of Shewanella bacterium, which serve as test organisms in this investigation. Currently, facultative barophilic psychrophilic stains of Shewanella are grown in the presence of ferric food source; the strains exhibiting iron

  13. Orbital magnetism in ensembles of ballistic billiards

    International Nuclear Information System (INIS)

    Ullmo, D.; Richter, K.; Jalabert, R.A.

    1993-01-01

    The magnetic response of ensembles of small two-dimensional structures at finite temperatures is calculated. Using semiclassical methods and numerical calculation it is demonstrated that only short classical trajectories are relevant. The magnetic susceptibility is enhanced in regular systems, where these trajectories appear in families. For ensembles of squares large paramagnetic susceptibility is obtained, in good agreement with recent measurements in the ballistic regime. (authors). 20 refs., 2 figs

  14. Targeted quantification of functional enzyme dynamics in environmental samples for microbially mediated biogeochemical processes: Targeted quantification of functional enzyme dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Li, Minjing [School of Environmental Studies, China University of Geosciences, Wuhan 430074 People' s Republic of China; Gao, Yuqian [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Qian, Wei-Jun [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Shi, Liang [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Liu, Yuanyuan [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Nelson, William C. [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Nicora, Carrie D. [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Resch, Charles T. [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Thompson, Christopher [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Yan, Sen [School of Environmental Studies, China University of Geosciences, Wuhan 430074 People' s Republic of China; Fredrickson, James K. [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Zachara, John M. [Pacific Northwest National Laboratory, Richland, WA 99354 USA; Liu, Chongxuan [Pacific Northwest National Laboratory, Richland, WA 99354 USA; School of Environmental Science and Engineering, Southern University of Science and Technology, Shenzhen 518055 People' s Republic of China

    2017-07-13

    Microbially mediated biogeochemical processes are catalyzed by enzymes that control the transformation of carbon, nitrogen, and other elements in environment. The dynamic linkage between enzymes and biogeochemical species transformation has, however, rarely been investigated because of the lack of analytical approaches to efficiently and reliably quantify enzymes and their dynamics in soils and sediments. Herein, we developed a signature peptide-based technique for sensitively quantifying dissimilatory and assimilatory enzymes using nitrate-reducing enzymes in a hyporheic zone sediment as an example. Moreover, the measured changes in enzyme concentration were found to correlate with the nitrate reduction rate in a way different from that inferred from biogeochemical models based on biomass or functional genes as surrogates for functional enzymes. This phenomenon has important implications for understanding and modeling the dynamics of microbial community functions and biogeochemical processes in environments. Our results also demonstrate the importance of enzyme quantification for the identification and interrogation of those biogeochemical processes with low metabolite concentrations as a result of faster enzyme-catalyzed consumption of metabolites than their production. The dynamic enzyme behaviors provide a basis for the development of enzyme-based models to describe the relationship between the microbial community and biogeochemical processes.

  15. Ensemble dispersion forecasting - Part 2. Application and evaluation

    DEFF Research Database (Denmark)

    Galmarini, S.; Bianconi, R.; Addis, R.

    2004-01-01

    of the dispersion of ETEX release 1 and the model ensemble is compared with the monitoring data. The scope of the comparison is to estimate to what extent the ensemble analysis is an improvement with respect to the single model results and represents a superior analysis of the process evolution. (C) 2004 Elsevier...

  16. A variational ensemble scheme for noisy image data assimilation

    Science.gov (United States)

    Yang, Yin; Robinson, Cordelia; Heitz, Dominique; Mémin, Etienne

    2014-05-01

    Data assimilation techniques aim at recovering a system state variables trajectory denoted as X, along time from partially observed noisy measurements of the system denoted as Y. These procedures, which couple dynamics and noisy measurements of the system, fulfill indeed a twofold objective. On one hand, they provide a denoising - or reconstruction - procedure of the data through a given model framework and on the other hand, they provide estimation procedures for unknown parameters of the dynamics. A standard variational data assimilation problem can be formulated as the minimization of the following objective function with respect to the initial discrepancy, η, from the background initial guess: δ« J(η(x)) = 1∥Xb (x) - X (t ,x)∥2 + 1 tf∥H(X (t,x ))- Y (t,x)∥2dt. 2 0 0 B 2 t0 R (1) where the observation operator H links the state variable and the measurements. The cost function can be interpreted as the log likelihood function associated to the a posteriori distribution of the state given the past history of measurements and the background. In this work, we aim at studying ensemble based optimal control strategies for data assimilation. Such formulation nicely combines the ingredients of ensemble Kalman filters and variational data assimilation (4DVar). It is also formulated as the minimization of the objective function (1), but similarly to ensemble filter, it introduces in its objective function an empirical ensemble-based background-error covariance defined as: B ≡ )(Xb - )T>. (2) Thus, it works in an off-line smoothing mode rather than on the fly like sequential filters. Such resulting ensemble variational data assimilation technique corresponds to a relatively new family of methods [1,2,3]. It presents two main advantages: first, it does not require anymore to construct the adjoint of the dynamics tangent linear operator, which is a considerable advantage with respect to the method's implementation, and second, it enables the handling of a flow

  17. Ensemble atmospheric dispersion modeling for emergency response consequence assessments

    International Nuclear Information System (INIS)

    Addis, R.P.; Buckley, R.L.

    2003-01-01

    Full text: Prognostic atmospheric dispersion models are used to generate consequence assessments, which assist decision-makers in the event of a release from a nuclear facility. Differences in the forecast wind fields generated by various meteorological agencies, differences in the transport and diffusion models themselves, as well as differences in the way these models treat the release source term, all may result in differences in the simulated plumes. This talk will address the U.S. participation in the European ENSEMBLE project, and present a perspective an how ensemble techniques may be used to enable atmospheric modelers to provide decision-makers with a more realistic understanding of how both the atmosphere and the models behave. Meteorological forecasts generated by numerical models from national and multinational meteorological agencies provide individual realizations of three-dimensional, time dependent atmospheric wind fields. These wind fields may be used to drive atmospheric dispersion (transport and diffusion) models, or they may be used to initiate other, finer resolution meteorological models, which in turn drive dispersion models. Many modeling agencies now utilize ensemble-modeling techniques to determine how sensitive the prognostic fields are to minor perturbations in the model parameters. However, the European Union programs RTMOD and ENSEMBLE are the first projects to utilize a WEB based ensemble approach to interpret the output from atmospheric dispersion models. The ensembles produced are different from those generated by meteorological forecasting centers in that they are ensembles of dispersion model outputs from many different atmospheric transport and diffusion models utilizing prognostic atmospheric fields from several different forecast centers. As such, they enable a decision-maker to consider the uncertainty in the plume transport and growth as a result of the differences in the forecast wind fields as well as the differences in the

  18. River Flow Prediction Using the Nearest Neighbor Probabilistic Ensemble Method

    Directory of Open Access Journals (Sweden)

    H. Sanikhani

    2016-02-01

    Full Text Available Introduction: In the recent years, researchers interested on probabilistic forecasting of hydrologic variables such river flow.A probabilistic approach aims at quantifying the prediction reliability through a probability distribution function or a prediction interval for the unknown future value. The evaluation of the uncertainty associated to the forecast is seen as a fundamental information, not only to correctly assess the prediction, but also to compare forecasts from different methods and to evaluate actions and decisions conditionally on the expected values. Several probabilistic approaches have been proposed in the literature, including (1 methods that use resampling techniques to assess parameter and model uncertainty, such as the Metropolis algorithm or the Generalized Likelihood Uncertainty Estimation (GLUE methodology for an application to runoff prediction, (2 methods based on processing the forecast errors of past data to produce the probability distributions of future values and (3 methods that evaluate how the uncertainty propagates from the rainfall forecast to the river discharge prediction, as the Bayesian forecasting system. Materials and Methods: In this study, two different probabilistic methods are used for river flow prediction.Then the uncertainty related to the forecast is quantified. One approach is based on linear predictors and in the other, nearest neighbor was used. The nonlinear probabilistic ensemble can be used for nonlinear time series analysis using locally linear predictors, while NNPE utilize a method adapted for one step ahead nearest neighbor methods. In this regard, daily river discharge (twelve years of Dizaj and Mashin Stations on Baranduz-Chay basin in west Azerbijan and Zard-River basin in Khouzestan provinces were used, respectively. The first six years of data was applied for fitting the model. The next three years was used to calibration and the remained three yeas utilized for testing the models

  19. Benchmarking Commercial Conformer Ensemble Generators.

    Science.gov (United States)

    Friedrich, Nils-Ole; de Bruyn Kops, Christina; Flachsenberg, Florian; Sommer, Kai; Rarey, Matthias; Kirchmair, Johannes

    2017-11-27

    We assess and compare the performance of eight commercial conformer ensemble generators (ConfGen, ConfGenX, cxcalc, iCon, MOE LowModeMD, MOE Stochastic, MOE Conformation Import, and OMEGA) and one leading free algorithm, the distance geometry algorithm implemented in RDKit. The comparative study is based on a new version of the Platinum Diverse Dataset, a high-quality benchmarking dataset of 2859 protein-bound ligand conformations extracted from the PDB. Differences in the performance of commercial algorithms are much smaller than those observed for free algorithms in our previous study (J. Chem. Inf. 2017, 57, 529-539). For commercial algorithms, the median minimum root-mean-square deviations measured between protein-bound ligand conformations and ensembles of a maximum of 250 conformers are between 0.46 and 0.61 Å. Commercial conformer ensemble generators are characterized by their high robustness, with at least 99% of all input molecules successfully processed and few or even no substantial geometrical errors detectable in their output conformations. The RDKit distance geometry algorithm (with minimization enabled) appears to be a good free alternative since its performance is comparable to that of the midranked commercial algorithms. Based on a statistical analysis, we elaborate on which algorithms to use and how to parametrize them for best performance in different application scenarios.

  20. Mass Conservation and Positivity Preservation with Ensemble-type Kalman Filter Algorithms

    Science.gov (United States)

    Janjic, Tijana; McLaughlin, Dennis B.; Cohn, Stephen E.; Verlaan, Martin

    2013-01-01

    Maintaining conservative physical laws numerically has long been recognized as being important in the development of numerical weather prediction (NWP) models. In the broader context of data assimilation, concerted efforts to maintain conservation laws numerically and to understand the significance of doing so have begun only recently. In order to enforce physically based conservation laws of total mass and positivity in the ensemble Kalman filter, we incorporate constraints to ensure that the filter ensemble members and the ensemble mean conserve mass and remain nonnegative through measurement updates. We show that the analysis steps of ensemble transform Kalman filter (ETKF) algorithm and ensemble Kalman filter algorithm (EnKF) can conserve the mass integral, but do not preserve positivity. Further, if localization is applied or if negative values are simply set to zero, then the total mass is not conserved either. In order to ensure mass conservation, a projection matrix that corrects for localization effects is constructed. In order to maintain both mass conservation and positivity preservation through the analysis step, we construct a data assimilation algorithms based on quadratic programming and ensemble Kalman filtering. Mass and positivity are both preserved by formulating the filter update as a set of quadratic programming problems that incorporate constraints. Some simple numerical experiments indicate that this approach can have a significant positive impact on the posterior ensemble distribution, giving results that are more physically plausible both for individual ensemble members and for the ensemble mean. The results show clear improvements in both analyses and forecasts, particularly in the presence of localized features. Behavior of the algorithm is also tested in presence of model error.

  1. Ensembl Genomes 2016: more genomes, more complexity.

    Science.gov (United States)

    Kersey, Paul Julian; Allen, James E; Armean, Irina; Boddu, Sanjay; Bolt, Bruce J; Carvalho-Silva, Denise; Christensen, Mikkel; Davis, Paul; Falin, Lee J; Grabmueller, Christoph; Humphrey, Jay; Kerhornou, Arnaud; Khobova, Julia; Aranganathan, Naveen K; Langridge, Nicholas; Lowy, Ernesto; McDowall, Mark D; Maheswari, Uma; Nuhn, Michael; Ong, Chuang Kee; Overduin, Bert; Paulini, Michael; Pedro, Helder; Perry, Emily; Spudich, Giulietta; Tapanari, Electra; Walts, Brandon; Williams, Gareth; Tello-Ruiz, Marcela; Stein, Joshua; Wei, Sharon; Ware, Doreen; Bolser, Daniel M; Howe, Kevin L; Kulesha, Eugene; Lawson, Daniel; Maslen, Gareth; Staines, Daniel M

    2016-01-04

    Ensembl Genomes (http://www.ensemblgenomes.org) is an integrating resource for genome-scale data from non-vertebrate species, complementing the resources for vertebrate genomics developed in the context of the Ensembl project (http://www.ensembl.org). Together, the two resources provide a consistent set of programmatic and interactive interfaces to a rich range of data including reference sequence, gene models, transcriptional data, genetic variation and comparative analysis. This paper provides an update to the previous publications about the resource, with a focus on recent developments. These include the development of new analyses and views to represent polyploid genomes (of which bread wheat is the primary exemplar); and the continued up-scaling of the resource, which now includes over 23 000 bacterial genomes, 400 fungal genomes and 100 protist genomes, in addition to 55 genomes from invertebrate metazoa and 39 genomes from plants. This dramatic increase in the number of included genomes is one part of a broader effort to automate the integration of archival data (genome sequence, but also associated RNA sequence data and variant calls) within the context of reference genomes and make it available through the Ensembl user interfaces. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.

  2. Conductor gestures influence evaluations of ensemble performance

    Directory of Open Access Journals (Sweden)

    Steven eMorrison

    2014-07-01

    Full Text Available Previous research has found that listener evaluations of ensemble performances vary depending on the expressivity of the conductor’s gestures, even when performances are otherwise identical. It was the purpose of the present study to test whether this effect of visual information was evident in the evaluation of specific aspects of ensemble performance, articulation and dynamics. We constructed a set of 32 music performances that combined auditory and visual information and were designed to feature a high degree of contrast along one of two target characteristics: articulation and dynamics. We paired each of four music excerpts recorded by a chamber ensemble in both a high- and low-contrast condition with video of four conductors demonstrating high- and low-contrast gesture specifically appropriate to either articulation or dynamics. Using one of two equivalent test forms, college music majors and nonmajors (N = 285 viewed sixteen 30-second performances and evaluated the quality of the ensemble’s articulation, dynamics, technique and tempo along with overall expressivity. Results showed significantly higher evaluations for performances featuring high rather than low conducting expressivity regardless of the ensemble’s performance quality. Evaluations for both articulation and dynamics were strongly and positively correlated with evaluations of overall ensemble expressivity.

  3. Conservation of Mass and Preservation of Positivity with Ensemble-Type Kalman Filter Algorithms

    Science.gov (United States)

    Janjic, Tijana; Mclaughlin, Dennis; Cohn, Stephen E.; Verlaan, Martin

    2014-01-01

    This paper considers the incorporation of constraints to enforce physically based conservation laws in the ensemble Kalman filter. In particular, constraints are used to ensure that the ensemble members and the ensemble mean conserve mass and remain nonnegative through measurement updates. In certain situations filtering algorithms such as the ensemble Kalman filter (EnKF) and ensemble transform Kalman filter (ETKF) yield updated ensembles that conserve mass but are negative, even though the actual states must be nonnegative. In such situations if negative values are set to zero, or a log transform is introduced, the total mass will not be conserved. In this study, mass and positivity are both preserved by formulating the filter update as a set of quadratic programming problems that incorporate non-negativity constraints. Simple numerical experiments indicate that this approach can have a significant positive impact on the posterior ensemble distribution, giving results that are more physically plausible both for individual ensemble members and for the ensemble mean. In two examples, an update that includes a non-negativity constraint is able to properly describe the transport of a sharp feature (e.g., a triangle or cone). A number of implementation questions still need to be addressed, particularly the need to develop a computationally efficient quadratic programming update for large ensemble.

  4. Momentum distribution functions in ensembles: the inequivalence of microcannonical and canonical ensembles in a finite ultracold system.

    Science.gov (United States)

    Wang, Pei; Xianlong, Gao; Li, Haibin

    2013-08-01

    It is demonstrated in many thermodynamic textbooks that the equivalence of the different ensembles is achieved in the thermodynamic limit. In this present work we discuss the inequivalence of microcanonical and canonical ensembles in a finite ultracold system at low energies. We calculate the microcanonical momentum distribution function (MDF) in a system of identical fermions (bosons). We find that the microcanonical MDF deviates from the canonical one, which is the Fermi-Dirac (Bose-Einstein) function, in a finite system at low energies where the single-particle density of states and its inverse are finite.

  5. Biogeochemical processes on tree islands in the greater everglades: Initiating a new paradigm

    Science.gov (United States)

    Wetzel, P.R.; Sklar, Fred H.; Coronado, C.A.; Troxler, T.G.; Krupa, S.L.; Sullivan, P.L.; Ewe, S.; Price, R.M.; Newman, S.; Orem, W.H.

    2011-01-01

    Scientists' understanding of the role of tree islands in the Everglades has evolved from a plant community of minor biogeochemical importance to a plant community recognized as the driving force for localized phosphorus accumulation within the landscape. Results from this review suggest that tree transpiration, nutrient infiltration from the soil surface, and groundwater flow create a soil zone of confluence where nutrients and salts accumulate under the head of a tree island during dry periods. Results also suggest accumulated salts and nutrients are flushed downstream by regional water flows during wet periods. That trees modulate their environment to create biogeochemical hot spots and strong nutrient gradients is a significant ecological paradigm shift in the understanding of the biogeochemical processes in the Everglades. In terms of island sustainability, this new paradigm suggests the need for distinct dry-wet cycles as well as a hydrologic regime that supports tree survival. Restoration of historic tree islands needs further investigation but the creation of functional tree islands is promising. Copyright ?? 2011 Taylor & Francis Group, LLC.

  6. An educational model for ensemble streamflow simulation and uncertainty analysis

    Directory of Open Access Journals (Sweden)

    A. AghaKouchak

    2013-02-01

    Full Text Available This paper presents the hands-on modeling toolbox, HBV-Ensemble, designed as a complement to theoretical hydrology lectures, to teach hydrological processes and their uncertainties. The HBV-Ensemble can be used for in-class lab practices and homework assignments, and assessment of students' understanding of hydrological processes. Using this modeling toolbox, students can gain more insights into how hydrological processes (e.g., precipitation, snowmelt and snow accumulation, soil moisture, evapotranspiration and runoff generation are interconnected. The educational toolbox includes a MATLAB Graphical User Interface (GUI and an ensemble simulation scheme that can be used for teaching uncertainty analysis, parameter estimation, ensemble simulation and model sensitivity. HBV-Ensemble was administered in a class for both in-class instruction and a final project, and students submitted their feedback about the toolbox. The results indicate that this educational software had a positive impact on students understanding and knowledge of uncertainty in hydrological modeling.

  7. A Hyper-Heuristic Ensemble Method for Static Job-Shop Scheduling.

    Science.gov (United States)

    Hart, Emma; Sim, Kevin

    2016-01-01

    We describe a new hyper-heuristic method NELLI-GP for solving job-shop scheduling problems (JSSP) that evolves an ensemble of heuristics. The ensemble adopts a divide-and-conquer approach in which each heuristic solves a unique subset of the instance set considered. NELLI-GP extends an existing ensemble method called NELLI by introducing a novel heuristic generator that evolves heuristics composed of linear sequences of dispatching rules: each rule is represented using a tree structure and is itself evolved. Following a training period, the ensemble is shown to outperform both existing dispatching rules and a standard genetic programming algorithm on a large set of new test instances. In addition, it obtains superior results on a set of 210 benchmark problems from the literature when compared to two state-of-the-art hyper-heuristic approaches. Further analysis of the relationship between heuristics in the evolved ensemble and the instances each solves provides new insights into features that might describe similar instances.

  8. Inhomogeneous ensembles of radical pairs in chemical compasses

    Science.gov (United States)

    Procopio, Maria; Ritz, Thorsten

    2016-11-01

    The biophysical basis for the ability of animals to detect the geomagnetic field and to use it for finding directions remains a mystery of sensory biology. One much debated hypothesis suggests that an ensemble of specialized light-induced radical pair reactions can provide the primary signal for a magnetic compass sensor. The question arises what features of such a radical pair ensemble could be optimized by evolution so as to improve the detection of the direction of weak magnetic fields. Here, we focus on the overlooked aspect of the noise arising from inhomogeneity of copies of biomolecules in a realistic biological environment. Such inhomogeneity leads to variations of the radical pair parameters, thereby deteriorating the signal arising from an ensemble and providing a source of noise. We investigate the effect of variations in hyperfine interactions between different copies of simple radical pairs on the directional response of a compass system. We find that the choice of radical pair parameters greatly influences how strongly the directional response of an ensemble is affected by inhomogeneity.

  9. Enhancing COSMO-DE ensemble forecasts by inexpensive techniques

    Directory of Open Access Journals (Sweden)

    Zied Ben Bouallègue

    2013-02-01

    Full Text Available COSMO-DE-EPS, a convection-permitting ensemble prediction system based on the high-resolution numerical weather prediction model COSMO-DE, is pre-operational since December 2010, providing probabilistic forecasts which cover Germany. This ensemble system comprises 20 members based on variations of the lateral boundary conditions, the physics parameterizations and the initial conditions. In order to increase the sample size in a computationally inexpensive way, COSMO-DE-EPS is combined with alternative ensemble techniques: the neighborhood method and the time-lagged approach. Their impact on the quality of the resulting probabilistic forecasts is assessed. Objective verification is performed over a six months period, scores based on the Brier score and its decomposition are shown for June 2011. The combination of the ensemble system with the alternative approaches improves probabilistic forecasts of precipitation in particular for high precipitation thresholds. Moreover, combining COSMO-DE-EPS with only the time-lagged approach improves the skill of area probabilities for precipitation and does not deteriorate the skill of 2 m-temperature and wind gusts forecasts.

  10. Realization of Deutsch-like algorithm using ensemble computing

    International Nuclear Information System (INIS)

    Wei Daxiu; Luo Jun; Sun Xianping; Zeng Xizhi

    2003-01-01

    The Deutsch-like algorithm [Phys. Rev. A. 63 (2001) 034101] distinguishes between even and odd query functions using fewer function calls than its possible classical counterpart in a two-qubit system. But the similar method cannot be applied to a multi-qubit system. We propose a new approach for solving Deutsch-like problem using ensemble computing. The proposed algorithm needs an ancillary qubit and can be easily extended to multi-qubit system with one query. Our ensemble algorithm beginning with a easily-prepared initial state has three main steps. The classifications of the functions can be obtained directly from the spectra of the ancilla qubit. We also demonstrate the new algorithm in a four-qubit molecular system using nuclear magnetic resonance (NMR). One hydrogen and three carbons are selected as the four qubits, and one of carbons is ancilla qubit. We choice two unitary transformations, corresponding to two functions (one odd function and one even function), to validate the ensemble algorithm. The results show that our experiment is successfully and our ensemble algorithm for solving the Deutsch-like problem is virtual

  11. Spectral statistics in semiclassical random-matrix ensembles

    International Nuclear Information System (INIS)

    Feingold, M.; Leitner, D.M.; Wilkinson, M.

    1991-01-01

    A novel random-matrix ensemble is introduced which mimics the global structure inherent in the Hamiltonian matrices of autonomous, ergodic systems. Changes in its parameters induce a transition between a Poisson and a Wigner distribution for the level spacings, P(s). The intermediate distributions are uniquely determined by a single scaling variable. Semiclassical constraints force the ensemble to be in a regime with Wigner P(s) for systems with more than two freedoms

  12. Multivariate localization methods for ensemble Kalman filtering

    OpenAIRE

    S. Roh; M. Jun; I. Szunyogh; M. G. Genton

    2015-01-01

    In ensemble Kalman filtering (EnKF), the small number of ensemble members that is feasible to use in a practical data assimilation application leads to sampling variability of the estimates of the background error covariances. The standard approach to reducing the effects of this sampling variability, which has also been found to be highly efficient in improving the performance of EnKF, is the localization of the estimates of the covariances. One family of ...

  13. Deviations from Wick's theorem in the canonical ensemble

    Science.gov (United States)

    Schönhammer, K.

    2017-07-01

    Wick's theorem for the expectation values of products of field operators for a system of noninteracting fermions or bosons plays an important role in the perturbative approach to the quantum many-body problem. A finite-temperature version holds in the framework of the grand canonical ensemble, but not for the canonical ensemble appropriate for systems with fixed particle number such as ultracold quantum gases in optical lattices. Here we present formulas for expectation values of products of field operators in the canonical ensemble using a method in the spirit of Gaudin's proof of Wick's theorem for the grand canonical case. The deviations from Wick's theorem are examined quantitatively for two simple models of noninteracting fermions.

  14. Preferences of and Attitudes toward Treble Choral Ensembles

    Science.gov (United States)

    Wilson, Jill M.

    2012-01-01

    In choral ensembles, a pursuit where females far outnumber males, concern exists that females are being devalued. Attitudes of female choral singers may be negatively affected by the gender imbalance that exists in mixed choirs and by the placement of the mixed choir as the most select ensemble in a program. The purpose of this research was to…

  15. A note on the multi model super ensemble technique for reducing forecast errors

    International Nuclear Information System (INIS)

    Kantha, L.; Carniel, S.; Sclavo, M.

    2008-01-01

    The multi model super ensemble (S E) technique has been used with considerable success to improve meteorological forecasts and is now being applied to ocean models. Although the technique has been shown to produce deterministic forecasts that can be superior to the individual models in the ensemble or a simple multi model ensemble forecast, there is a clear need to understand its strengths and limitations. This paper is an attempt to do so in simple, easily understood contexts. The results demonstrate that the S E forecast is almost always better than the simple ensemble forecast, the degree of improvement depending on the properties of the models in the ensemble. However, the skill of the S E forecast with respect to the true forecast depends on a number of factors, principal among which is the skill of the models in the ensemble. As can be expected, if the ensemble consists of models with poor skill, the S E forecast will also be poor, although better than the ensemble forecast. On the other hand, the inclusion of even a single skillful model in the ensemble increases the forecast skill significantly.

  16. Diel biogeochemical processes in terrestrial waters

    Science.gov (United States)

    Nimick, David A.; Gammons, Christopher H.

    2011-01-01

    Many biogeochemical processes in rivers and lakes respond to the solar photocycle and produce persistent patterns of measureable phenomena that exhibit a day–night, or 24-h, cycle. Despite a large body of recent literature, the mechanisms responsible for these diel fluctuations are widely debated, with a growing consensus that combinations of physical, chemical, and biological processes are involved. These processes include streamflow variation, photosynthesis and respiration, plant assimilation, and reactions involving photochemistry, adsorption and desorption, and mineral precipitation and dissolution. Diel changes in streamflow and water properties such as temperature, pH, and dissolved oxygen concentration have been widely recognized, and recently, diel studies have focused more widely by considering other constituents such as dissolved and particulate trace metals, metalloids, rare earth elements, mercury, organic matter, dissolved inorganic carbon (DIC), and nutrients. The details of many diel processes are being studied using stable isotopes, which also can exhibit diel cycles in response to microbial metabolism, photosynthesis and respiration, or changes in phase, speciation, or redox state. In addition, secondary effects that diel cycles might have, for example, on biota or in the hyporheic zone are beginning to be considered.This special issue is composed primarily of papers presented at the topical session “Diurnal Biogeochemical Processes in Rivers, Lakes, and Shallow Groundwater” held at the annual meeting of the Geological Society of America in October 2009 in Portland, Oregon. This session was organized because many of the growing number of diel studies have addressed just a small part of the full range of diel cycling phenomena found in rivers and lakes. This limited focus is understandable because (1) fundamental aspects of many diel processes are poorly understood and require detailed study, (2) the interests and expertise of individual

  17. Predictive Understanding of Mountainous Watershed Hydro-Biogeochemical Function and Response to Perturbations

    Science.gov (United States)

    Hubbard, S. S.; Williams, K. H.; Agarwal, D.; Banfield, J. F.; Beller, H. R.; Bouskill, N.; Brodie, E.; Maxwell, R. M.; Nico, P. S.; Steefel, C. I.; Steltzer, H.; Tokunaga, T. K.; Wainwright, H. M.; Dwivedi, D.; Newcomer, M. E.

    2017-12-01

    Recognizing the societal importance, vulnerability and complexity of mountainous watersheds, the `Watershed Function' project is developing a predictive understanding of how mountainous watersheds retain and release downgradient water, nutrients, carbon, and metals. In particular, the project is exploring how early snowmelt, drought, floods and other disturbances will influence mountainous watershed dynamics at seasonal to decadal timescales. Located in the 300km2 East River headwater catchment of the Upper Colorado River Basin, the project is guided by several constructs. First, the project considers the integrated role of surface and subsurface flow and biogeochemical reactions - from bedrock to the top of the vegetative canopy, from terrestrial through aquatic compartments, and from summit to receiving waters. The project takes a system-of-systems perspective, focused on developing new methods to quantify the cumulative watershed hydrobiogeochemical response to perturbations based on information from select subsystems within the watershed, each having distinct vegetation-subsurface biogeochemical-hydrological characteristics. A `scale-adaptive' modeling capability, in development using adaptive mesh refinement methods, serves as the organizing framework for the SFA. The scale-adaptive approach is intended to permit simulation of system-within-systems behavior - and aggregation of that behavior - from genome through watershed scales. This presentation will describe several early project discoveries and advances made using experimental, observational and numerical approaches. Among others, examples may include:quantiying how seasonal hydrological perturbations drive biogeochemical responses across critical zone compartments, with a focus on N and C transformations; metagenomic documentation of the spatial variability in floodplain meander microbial ecology; 3D reactive transport simulations of couped hydrological-biogeochemical behavior in the hyporheic zone; and

  18. Ensemble prediction of floods – catchment non-linearity and forecast probabilities

    Directory of Open Access Journals (Sweden)

    C. Reszler

    2007-07-01

    Full Text Available Quantifying the uncertainty of flood forecasts by ensemble methods is becoming increasingly important for operational purposes. The aim of this paper is to examine how the ensemble distribution of precipitation forecasts propagates in the catchment system, and to interpret the flood forecast probabilities relative to the forecast errors. We use the 622 km2 Kamp catchment in Austria as an example where a comprehensive data set, including a 500 yr and a 1000 yr flood, is available. A spatially-distributed continuous rainfall-runoff model is used along with ensemble and deterministic precipitation forecasts that combine rain gauge data, radar data and the forecast fields of the ALADIN and ECMWF numerical weather prediction models. The analyses indicate that, for long lead times, the variability of the precipitation ensemble is amplified as it propagates through the catchment system as a result of non-linear catchment response. In contrast, for lead times shorter than the catchment lag time (e.g. 12 h and less, the variability of the precipitation ensemble is decreased as the forecasts are mainly controlled by observed upstream runoff and observed precipitation. Assuming that all ensemble members are equally likely, the statistical analyses for five flood events at the Kamp showed that the ensemble spread of the flood forecasts is always narrower than the distribution of the forecast errors. This is because the ensemble forecasts focus on the uncertainty in forecast precipitation as the dominant source of uncertainty, and other sources of uncertainty are not accounted for. However, a number of analyses, including Relative Operating Characteristic diagrams, indicate that the ensemble spread is a useful indicator to assess potential forecast errors for lead times larger than 12 h.

  19. Multimillennium changes in dissolved oxygen under global warming: results from an AOGCM and offline ocean biogeochemical model

    Science.gov (United States)

    Yamamoto, A.; Abe-Ouchi, A.; Shigemitsu, M.; Oka, A.; Takahashi, K.; Ohgaito, R.; Yamanaka, Y.

    2016-12-01

    Long-term oceanic oxygen change due to global warming is still unclear; most future projections (such as CMIP5) are only performed until 2100. Indeed, few previous studies using conceptual models project oxygen change in the next thousands of years, showing persistent global oxygen reduction by about 30% in the next 2000 years, even after atmospheric carbon dioxide stops rising. Yet, these models cannot sufficiently represent the ocean circulation change: the key driver of oxygen change. Moreover, considering serious effect oxygen reduction has on marine life and biogeochemical cycling, long-term oxygen change should be projected for higher validity. Therefore, we used a coupled atmosphere-ocean general circulation model (AOGCM) and an offline ocean biogeochemical model, investigating realistic long-term changes in oceanic oxygen concentration and ocean circulation. We integrated these models for 2000 years under atmospheric CO2 doubling and quadrupling. After global oxygen reduction in the first 500 years, oxygen concentration in deep ocean globally recovers and overshoots, despite surface oxygen decrease and weaker Atlantic Meridional Overturning Circulation. Deep ocean convection in the Weddell Sea recovers and overshoots, after initial cessation. Thus, enhanced deep convection and associated Antarctic Bottom Water supply oxygen-rich surface waters to deep ocean, resulting global deep ocean oxygenation. We conclude that the change in ocean circulation in the Southern Ocean potentially drives millennial-scale oxygenation in the deep ocean; contrary to past reported long-term oxygen reduction and general expectation. In presentation, we will discuss the mechanism of response of deep ocean convection in the Weddell Sea and show the volume changes of hypoxic waters.

  20. Rainfall downscaling of weekly ensemble forecasts using self-organising maps

    Directory of Open Access Journals (Sweden)

    Masamichi Ohba

    2016-03-01

    Full Text Available This study presents an application of self-organising maps (SOMs to downscaling medium-range ensemble forecasts and probabilistic prediction of local precipitation in Japan. SOM was applied to analyse and connect the relationship between atmospheric patterns over Japan and local high-resolution precipitation data. Multiple SOM was simultaneously employed on four variables derived from the JRA-55 reanalysis over the area of study (south-western Japan, and a two-dimensional lattice of weather patterns (WPs was obtained. Weekly ensemble forecasts can be downscaled to local precipitation using the obtained multiple SOM. The downscaled precipitation is derived by the five SOM lattices based on the WPs of the global model ensemble forecasts for a particular day in 2009–2011. Because this method effectively handles the stochastic uncertainties from the large number of ensemble members, a probabilistic local precipitation is easily and quickly obtained from the ensemble forecasts. This downscaling of ensemble forecasts provides results better than those from a 20-km global spectral model (i.e. capturing the relatively detailed precipitation distribution over the region. To capture the effect of the detailed pattern differences in each SOM node, a statistical model is additionally concreted for each SOM node. The predictability skill of the ensemble forecasts is significantly improved under the neural network-statistics hybrid-downscaling technique, which then brings a much better skill score than the traditional method. It is expected that the results of this study will provide better guidance to the user community and contribute to the future development of dam-management models.

  1. High resolution modelling of the biogeochemical processes in the eutrophic Loire River (France)

    Science.gov (United States)

    Minaudo, Camille; Moatar, Florentina; Curie, Florence; Gassama, Nathalie; Billen, Gilles

    2016-04-01

    A biogeochemical model was developed, coupling a physically based water temperature model (T-NET) with a semi-mechanistic biogeochemical model (RIVE, used in ProSe and Riverstrahler models) in order to assess at a fine temporal and spatial resolution the biogeochemical processes in the eutrophic Middle Loire hydrosystem (≈10 000 km², 3361 river segments). The code itself allows parallelized computing, which decreased greatly the calculation time (5 hours for simulating 3 years hourly). We conducted a daily survey during the period 2012-2014 at 2 sampling stations located in the Middle Loire of nutrients, chlorophyll pigments, phytoplankton and physic-chemical variables. This database was used as both input data (upstream Loire boundary) and validation data of the model (basin outlet). Diffuse and non-point sources were assessed based on a land cover analysis and WWTP datasets. The results appeared very sensible to the coefficients governing the dynamic of suspended solids and of phosphorus (sorption/desorption processes) within the model and some parameters needed to be estimated numerically. Both the Lagrangian point of view and fluxes budgets at the seasonal and event-based scale evidenced the biogeochemical functioning of the Loire River. Low discharge levels set up favorable physical conditions for phytoplankton growth (long water travel time, limited water depth, suspended particles sedimentation). Conversely, higher discharge levels highly limited the phytoplankton biomass (dilution of the colony, washing-out, limited travel time, remobilization of suspended sediments increasing turbidity), and most biogeochemical species were basically transferred downstream. When hydrological conditions remained favorable for phytoplankton development, P-availability was the critical factor. However, the model evidenced that most of the P in summer was recycled within the water body: on one hand it was assimilated by the algae biomass, and on the other hand it was

  2. Quantum Ensemble Classification: A Sampling-Based Learning Control Approach.

    Science.gov (United States)

    Chen, Chunlin; Dong, Daoyi; Qi, Bo; Petersen, Ian R; Rabitz, Herschel

    2017-06-01

    Quantum ensemble classification (QEC) has significant applications in discrimination of atoms (or molecules), separation of isotopes, and quantum information extraction. However, quantum mechanics forbids deterministic discrimination among nonorthogonal states. The classification of inhomogeneous quantum ensembles is very challenging, since there exist variations in the parameters characterizing the members within different classes. In this paper, we recast QEC as a supervised quantum learning problem. A systematic classification methodology is presented by using a sampling-based learning control (SLC) approach for quantum discrimination. The classification task is accomplished via simultaneously steering members belonging to different classes to their corresponding target states (e.g., mutually orthogonal states). First, a new discrimination method is proposed for two similar quantum systems. Then, an SLC method is presented for QEC. Numerical results demonstrate the effectiveness of the proposed approach for the binary classification of two-level quantum ensembles and the multiclass classification of multilevel quantum ensembles.

  3. Cluster ensembles, quantization and the dilogarithm

    DEFF Research Database (Denmark)

    Fock, Vladimir; Goncharov, Alexander B.

    2009-01-01

    A cluster ensemble is a pair of positive spaces (i.e. varieties equipped with positive atlases), coming with an action of a symmetry group . The space is closely related to the spectrum of a cluster algebra [ 12 ]. The two spaces are related by a morphism . The space is equipped with a closed -form......, possibly degenerate, and the space has a Poisson structure. The map is compatible with these structures. The dilogarithm together with its motivic and quantum avatars plays a central role in the cluster ensemble structure. We define a non-commutative -deformation of the -space. When is a root of unity...

  4. Nonequilibrium statistical mechanics ensemble method

    CERN Document Server

    Eu, Byung Chan

    1998-01-01

    In this monograph, nonequilibrium statistical mechanics is developed by means of ensemble methods on the basis of the Boltzmann equation, the generic Boltzmann equations for classical and quantum dilute gases, and a generalised Boltzmann equation for dense simple fluids The theories are developed in forms parallel with the equilibrium Gibbs ensemble theory in a way fully consistent with the laws of thermodynamics The generalised hydrodynamics equations are the integral part of the theory and describe the evolution of macroscopic processes in accordance with the laws of thermodynamics of systems far removed from equilibrium Audience This book will be of interest to researchers in the fields of statistical mechanics, condensed matter physics, gas dynamics, fluid dynamics, rheology, irreversible thermodynamics and nonequilibrium phenomena

  5. Dorsal-CA1 Hippocampal Neuronal Ensembles Encode Nicotine-Reward Contextual Associations.

    Science.gov (United States)

    Xia, Li; Nygard, Stephanie K; Sobczak, Gabe G; Hourguettes, Nicholas J; Bruchas, Michael R

    2017-06-06

    Natural and drug rewards increase the motivational valence of stimuli in the environment that, through Pavlovian learning mechanisms, become conditioned stimuli that directly motivate behavior in the absence of the original unconditioned stimulus. While the hippocampus has received extensive attention for its role in learning and memory processes, less is known regarding its role in drug-reward associations. We used in vivo Ca 2+ imaging in freely moving mice during the formation of nicotine preference behavior to examine the role of the dorsal-CA1 region of the hippocampus in encoding contextual reward-seeking behavior. We show the development of specific neuronal ensembles whose activity encodes nicotine-reward contextual memories and that are necessary for the expression of place preference. Our findings increase our understanding of CA1 hippocampal function in general and as it relates to reward processing by identifying a critical role for CA1 neuronal ensembles in nicotine place preference. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.

  6. Quantum canonical ensemble: A projection operator approach

    Science.gov (United States)

    Magnus, Wim; Lemmens, Lucien; Brosens, Fons

    2017-09-01

    Knowing the exact number of particles N, and taking this knowledge into account, the quantum canonical ensemble imposes a constraint on the occupation number operators. The constraint particularly hampers the systematic calculation of the partition function and any relevant thermodynamic expectation value for arbitrary but fixed N. On the other hand, fixing only the average number of particles, one may remove the above constraint and simply factorize the traces in Fock space into traces over single-particle states. As is well known, that would be the strategy of the grand-canonical ensemble which, however, comes with an additional Lagrange multiplier to impose the average number of particles. The appearance of this multiplier can be avoided by invoking a projection operator that enables a constraint-free computation of the partition function and its derived quantities in the canonical ensemble, at the price of an angular or contour integration. Introduced in the recent past to handle various issues related to particle-number projected statistics, the projection operator approach proves beneficial to a wide variety of problems in condensed matter physics for which the canonical ensemble offers a natural and appropriate environment. In this light, we present a systematic treatment of the canonical ensemble that embeds the projection operator into the formalism of second quantization while explicitly fixing N, the very number of particles rather than the average. Being applicable to both bosonic and fermionic systems in arbitrary dimensions, transparent integral representations are provided for the partition function ZN and the Helmholtz free energy FN as well as for two- and four-point correlation functions. The chemical potential is not a Lagrange multiplier regulating the average particle number but can be extracted from FN+1 -FN, as illustrated for a two-dimensional fermion gas.

  7. An iterative ensemble Kalman filter for reservoir engineering applications

    NARCIS (Netherlands)

    Krymskaya, M.V.; Hanea, R.G.; Verlaan, M.

    2009-01-01

    The study has been focused on examining the usage and the applicability of ensemble Kalman filtering techniques to the history matching procedures. The ensemble Kalman filter (EnKF) is often applied nowadays to solving such a problem. Meanwhile, traditional EnKF requires assumption of the

  8. Improving the ensemble-optimization method through covariance-matrix adaptation

    NARCIS (Netherlands)

    Fonseca, R.M.; Leeuwenburgh, O.; Hof, P.M.J. van den; Jansen, J.D.

    2015-01-01

    Ensemble optimization (referred to throughout the remainder of the paper as EnOpt) is a rapidly emerging method for reservoirmodel-based production optimization. EnOpt uses an ensemble of controls to approximate the gradient of the objective function with respect to the controls. Current

  9. Parallel quantum computing in a single ensemble quantum computer

    International Nuclear Information System (INIS)

    Long Guilu; Xiao, L.

    2004-01-01

    We propose a parallel quantum computing mode for ensemble quantum computer. In this mode, some qubits are in pure states while other qubits are in mixed states. It enables a single ensemble quantum computer to perform 'single-instruction-multidata' type of parallel computation. Parallel quantum computing can provide additional speedup in Grover's algorithm and Shor's algorithm. In addition, it also makes a fuller use of qubit resources in an ensemble quantum computer. As a result, some qubits discarded in the preparation of an effective pure state in the Schulman-Varizani and the Cleve-DiVincenzo algorithms can be reutilized

  10. The chiral Gaussian two-matrix ensemble of real asymmetric matrices

    International Nuclear Information System (INIS)

    Akemann, G; Phillips, M J; Sommers, H-J

    2010-01-01

    We solve a family of Gaussian two-matrix models with rectangular N x (N + ν) matrices, having real asymmetric matrix elements and depending on a non-Hermiticity parameter μ. Our model can be thought of as the chiral extension of the real Ginibre ensemble, relevant for Dirac operators in the same symmetry class. It has the property that its eigenvalues are either real, purely imaginary or come in complex conjugate eigenvalue pairs. The eigenvalue joint probability distribution for our model is explicitly computed, leading to a non-Gaussian distribution including K-Bessel functions. All n-point density correlation functions are expressed for finite N in terms of a Pfaffian form. This contains a kernel involving Laguerre polynomials in the complex plane as a building block which was previously computed by the authors. This kernel can be expressed in terms of the kernel for complex non-Hermitian matrices, generalizing the known relation among ensembles of Hermitian random matrices. Compact expressions are given for the density at finite N as an example, as well as its microscopic large-N limits at the origin for fixed ν at strong and weak non-Hermiticity.

  11. An efficient method to generate a perturbed parameter ensemble of a fully coupled AOGCM without flux-adjustment

    Directory of Open Access Journals (Sweden)

    P. J. Irvine

    2013-09-01

    Full Text Available We present a simple method to generate a perturbed parameter ensemble (PPE of a fully-coupled atmosphere-ocean general circulation model (AOGCM, HadCM3, without requiring flux-adjustment. The aim was to produce an ensemble that samples parametric uncertainty in some key variables and gives a plausible representation of the climate. Six atmospheric parameters, a sea-ice parameter and an ocean parameter were jointly perturbed within a reasonable range to generate an initial group of 200 members. To screen out implausible ensemble members, 20 yr pre-industrial control simulations were run and members whose temperature responses to the parameter perturbations were projected to be outside the range of 13.6 ± 2 °C, i.e. near to the observed pre-industrial global mean, were discarded. Twenty-one members, including the standard unperturbed model, were accepted, covering almost the entire span of the eight parameters, challenging the argument that without flux-adjustment parameter ranges would be unduly restricted. This ensemble was used in 2 experiments; an 800 yr pre-industrial and a 150 yr quadrupled CO2 simulation. The behaviour of the PPE for the pre-industrial control compared well to ERA-40 reanalysis data and the CMIP3 ensemble for a number of surface and atmospheric column variables with the exception of a few members in the Tropics. However, we find that members of the PPE with low values of the entrainment rate coefficient show very large increases in upper tropospheric and stratospheric water vapour concentrations in response to elevated CO2 and one member showed an implausible nonlinear climate response, and as such will be excluded from future experiments with this ensemble. The outcome of this study is a PPE of a fully-coupled AOGCM which samples parametric uncertainty and a simple methodology which would be applicable to other GCMs.

  12. Fluctuation, stationarity, and ergodic properties of random-matrix ensembles

    International Nuclear Information System (INIS)

    Pandey, A.

    1979-01-01

    The properties of random-matrix ensembles and the application of such ensembles to energy-level fluctuations and strength fluctuations are discussed. The two-point correlation function for complex spectra described by the three standard Gaussian ensembles is calculated, and its essential simplicity, displayed by an elementary procedure that derives from the dominance of binary correlations. The resultant function is exact for the unitary case and a very good approximation to the orthogonal and symplectic cases. The same procedure yields the spectrum for a Gaussian orthogonal ensemble (GOE) deformed by a pairing interaction. Several extensions are given and relationships to other problems of current interest are discussed. The standard fluctuation measures are rederived for the GOE, and their extensions to the unitary and symplectic cases are given. The measures are shown to derive, for the most part, from the two-point function, and new relationships between them are established, answering some long-standing questions. Some comparisons with experimental values are also made. All the cluster functions, and therefore the fluctuation measures, are shown to be stationary and strongly ergodic, thus justifying the use of random matrices for individual spectra. Strength fluctuations in the orthogonal ensemble are also considered. The Porter-Thomas distribution in its various forms is rederived and its ergodicity is established

  13. Convergence of the Square Root Ensemble Kalman Filter in the Large Ensemble Limit

    Czech Academy of Sciences Publication Activity Database

    Kwiatkowski, E.; Mandel, Jan

    2015-01-01

    Roč. 3, č. 1 (2015), s. 1-17 ISSN 2166-2525 R&D Projects: GA ČR GA13-34856S Institutional support: RVO:67985807 Keywords : data assimilation * Lp laws of large numbers * Hilbert space * ensemble Kalman filter Subject RIV: IN - Informatics, Computer Science

  14. Spatio-temporal behaviour of medium-range ensemble forecasts

    Science.gov (United States)

    Kipling, Zak; Primo, Cristina; Charlton-Perez, Andrew

    2010-05-01

    Using the recently-developed mean-variance of logarithms (MVL) diagram, together with the TIGGE archive of medium-range ensemble forecasts from nine different centres, we present an analysis of the spatio-temporal dynamics of their perturbations, and show how the differences between models and perturbation techniques can explain the shape of their characteristic MVL curves. We also consider the use of the MVL diagram to compare the growth of perturbations within the ensemble with the growth of the forecast error, showing that there is a much closer correspondence for some models than others. We conclude by looking at how the MVL technique might assist in selecting models for inclusion in a multi-model ensemble, and suggest an experiment to test its potential in this context.

  15. Forecasting European cold waves based on subsampling strategies of CMIP5 and Euro-CORDEX ensembles

    Science.gov (United States)

    Cordero-Llana, Laura; Braconnot, Pascale; Vautard, Robert; Vrac, Mathieu; Jezequel, Aglae

    2016-04-01

    Forecasting future extreme events under the present changing climate represents a difficult task. Currently there are a large number of ensembles of simulations for climate projections that take in account different models and scenarios. However, there is a need for reducing the size of the ensemble to make the interpretation of these simulations more manageable for impact studies or climate risk assessment. This can be achieved by developing subsampling strategies to identify a limited number of simulations that best represent the ensemble. In this study, cold waves are chosen to test different approaches for subsampling available simulations. The definition of cold waves depends on the criteria used, but they are generally defined using a minimum temperature threshold, the duration of the cold spell as well as their geographical extend. These climate indicators are not universal, highlighting the difficulty of directly comparing different studies. As part of the of the CLIPC European project, we use daily surface temperature data obtained from CMIP5 outputs as well as Euro-CORDEX simulations to predict future cold waves events in Europe. From these simulations a clustering method is applied to minimise the number of ensembles required. Furthermore, we analyse the different uncertainties that arise from the different model characteristics and definitions of climate indicators. Finally, we will test if the same subsampling strategy can be used for different climate indicators. This will facilitate the use of the subsampling results for a wide number of impact assessment studies.

  16. A state-space Bayesian framework for estimating biogeochemical transformations using time-lapse geophysical data

    Energy Technology Data Exchange (ETDEWEB)

    Chen, J.; Hubbard, S.; Williams, K.; Pride, S.; Li, L.; Steefel, C.; Slater, L.

    2009-04-15

    We develop a state-space Bayesian framework to combine time-lapse geophysical data with other types of information for quantitative estimation of biogeochemical parameters during bioremediation. We consider characteristics of end-products of biogeochemical transformations as state vectors, which evolve under constraints of local environments through evolution equations, and consider time-lapse geophysical data as available observations, which could be linked to the state vectors through petrophysical models. We estimate the state vectors and their associated unknown parameters over time using Markov chain Monte Carlo sampling methods. To demonstrate the use of the state-space approach, we apply it to complex resistivity data collected during laboratory column biostimulation experiments that were poised to precipitate iron and zinc sulfides during sulfate reduction. We develop a petrophysical model based on sphere-shaped cells to link the sulfide precipitate properties to the time-lapse geophysical attributes and estimate volume fraction of the sulfide precipitates, fraction of the dispersed, sulfide-encrusted cells, mean radius of the aggregated clusters, and permeability over the course of the experiments. Results of the case study suggest that the developed state-space approach permits the use of geophysical datasets for providing quantitative estimates of end-product characteristics and hydrological feedbacks associated with biogeochemical transformations. Although tested here on laboratory column experiment datasets, the developed framework provides the foundation needed for quantitative field-scale estimation of biogeochemical parameters over space and time using direct, but often sparse wellbore data with indirect, but more spatially extensive geophysical datasets.

  17. Ensemble modeling of E. coli in the Charles River, Boston, Massachusetts, USA.

    Science.gov (United States)

    Hellweger, F L

    2007-01-01

    A case study of ensemble modeling of Escherichia coli (E. coli) densities in surface waters in the context of public health risk prediction is presented. The output of two different models, mechanistic and empirical, are combined and compared to data. The mechanistic model is a high-resolution, time-variable, three-dimensional coupled hydrodynamic and water quality model. It generally reproduces the mechanisms of E. coli fate and transport in the river, including the presence and absence of a plume in the study area under similar input, but different hydrodynamic conditions caused by the operation of a downstream dam and wind. At the time series station, the model has a root mean square error (RMSE) of 370 CFU/100mL, a total error rate (with respect to the EPA-recommended single sample criteria value of 235 CFU/100mL) (TER) of 15% and negative error rate (NER) of 30%. The empirical model is based on multiple linear regression using the forcing functions of the mechanistic model as independent variables. It has better overall performance (at the time series station), due to a strong correlation of E. coli density with upstream inflow for this time period (RMSE =200 CFU/100mL, TER =13%, NER =1.6%). However, the model is mechanistically incorrect in that it predicts decreasing densities with increasing Combined Sewer Overflow (CSO) input. The two models are fundamentally different and their errors are uncorrelated (R(2) =0.02), which motivates their combination in an ensemble. Two combination approaches, a geometric mean ensemble (GME) and an "either exceeds" ensemble (EEE), are explored. The GME model outperforms the mechanistic and empirical models in terms of RMSE (190 CFU/100mL) and TER (11%), but has a higher NER (23%). The EEE has relatively high TER (16%), but low NER (0.8%) and may be the best method for a conservative prediction. The study demonstrates the potential utility of ensemble modeling for pathogen indicators, but significant further research is

  18. Modeling of steam generator in nuclear power plant using neural network ensemble

    International Nuclear Information System (INIS)

    Lee, S. K.; Lee, E. C.; Jang, J. W.

    2003-01-01

    Neural network is now being used in modeling the steam generator is known to be difficult due to the reverse dynamics. However, Neural network is prone to the problem of overfitting. This paper investigates the use of neural network combining methods to model steam generator water level and compares with single neural network. The results show that neural network ensemble is effective tool which can offer improved generalization, lower dependence of the training set and reduced training time

  19. One-day-ahead streamflow forecasting via super-ensembles of several neural network architectures based on the Multi-Level Diversity Model

    Science.gov (United States)

    Brochero, Darwin; Hajji, Islem; Pina, Jasson; Plana, Queralt; Sylvain, Jean-Daniel; Vergeynst, Jenna; Anctil, Francois

    2015-04-01

    Theories about generalization error with ensembles are mainly based on the diversity concept, which promotes resorting to many members of different properties to support mutually agreeable decisions. Kuncheva (2004) proposed the Multi Level Diversity Model (MLDM) to promote diversity in model ensembles, combining different data subsets, input subsets, models, parameters, and including a combiner level in order to optimize the final ensemble. This work tests the hypothesis about the minimisation of the generalization error with ensembles of Neural Network (NN) structures. We used the MLDM to evaluate two different scenarios: (i) ensembles from a same NN architecture, and (ii) a super-ensemble built by a combination of sub-ensembles of many NN architectures. The time series used correspond to the 12 basins of the MOdel Parameter Estimation eXperiment (MOPEX) project that were used by Duan et al. (2006) and Vos (2013) as benchmark. Six architectures are evaluated: FeedForward NN (FFNN) trained with the Levenberg Marquardt algorithm (Hagan et al., 1996), FFNN trained with SCE (Duan et al., 1993), Recurrent NN trained with a complex method (Weins et al., 2008), Dynamic NARX NN (Leontaritis and Billings, 1985), Echo State Network (ESN), and leak integrator neuron (L-ESN) (Lukosevicius and Jaeger, 2009). Each architecture performs separately an Input Variable Selection (IVS) according to a forward stepwise selection (Anctil et al., 2009) using mean square error as objective function. Post-processing by Predictor Stepwise Selection (PSS) of the super-ensemble has been done following the method proposed by Brochero et al. (2011). IVS results showed that the lagged stream flow, lagged precipitation, and Standardized Precipitation Index (SPI) (McKee et al., 1993) were the most relevant variables. They were respectively selected as one of the firsts three selected variables in 66, 45, and 28 of the 72 scenarios. A relationship between aridity index (Arora, 2002) and NN

  20. Competitive Learning Neural Network Ensemble Weighted by Predicted Performance

    Science.gov (United States)

    Ye, Qiang

    2010-01-01

    Ensemble approaches have been shown to enhance classification by combining the outputs from a set of voting classifiers. Diversity in error patterns among base classifiers promotes ensemble performance. Multi-task learning is an important characteristic for Neural Network classifiers. Introducing a secondary output unit that receives different…

  1. Ensemble system for Part-of-Speech tagging

    OpenAIRE

    Dell'Orletta, Felice

    2009-01-01

    The paper contains a description of the Felice-POS-Tagger and of its performance in Evalita 2009. Felice-POS-Tagger is an ensemble system that combines six different POS taggers. When evaluated on the official test set, the ensemble system outperforms each of the single tagger components and achieves the highest accuracy score in Evalita 2009 POS Closed Task. It is shown rst that the errors made from the dierent taggers are complementary, and then how to use this complementary behavior to the...

  2. Learning to Run with Actor-Critic Ensemble

    OpenAIRE

    Huang, Zhewei; Zhou, Shuchang; Zhuang, BoEr; Zhou, Xinyu

    2017-01-01

    We introduce an Actor-Critic Ensemble(ACE) method for improving the performance of Deep Deterministic Policy Gradient(DDPG) algorithm. At inference time, our method uses a critic ensemble to select the best action from proposals of multiple actors running in parallel. By having a larger candidate set, our method can avoid actions that have fatal consequences, while staying deterministic. Using ACE, we have won the 2nd place in NIPS'17 Learning to Run competition, under the name of "Megvii-hzw...

  3. Robust Ensemble Filtering and Its Relation to Covariance Inflation in the Ensemble Kalman Filter

    KAUST Repository

    Luo, Xiaodong; Hoteit, Ibrahim

    2011-01-01

    A robust ensemble filtering scheme based on the H∞ filtering theory is proposed. The optimal H∞ filter is derived by minimizing the supremum (or maximum) of a predefined cost function, a criterion different from the minimum variance used

  4. Ensemble Nonlinear Autoregressive Exogenous Artificial Neural Networks for Short-Term Wind Speed and Power Forecasting.

    Science.gov (United States)

    Men, Zhongxian; Yee, Eugene; Lien, Fue-Sang; Yang, Zhiling; Liu, Yongqian

    2014-01-01

    Short-term wind speed and wind power forecasts (for a 72 h period) are obtained using a nonlinear autoregressive exogenous artificial neural network (ANN) methodology which incorporates either numerical weather prediction or high-resolution computational fluid dynamics wind field information as an exogenous input. An ensemble approach is used to combine the predictions from many candidate ANNs in order to provide improved forecasts for wind speed and power, along with the associated uncertainties in these forecasts. More specifically, the ensemble ANN is used to quantify the uncertainties arising from the network weight initialization and from the unknown structure of the ANN. All members forming the ensemble of neural networks were trained using an efficient particle swarm optimization algorithm. The results of the proposed methodology are validated using wind speed and wind power data obtained from an operational wind farm located in Northern China. The assessment demonstrates that this methodology for wind speed and power forecasting generally provides an improvement in predictive skills when compared to the practice of using an "optimal" weight vector from a single ANN while providing additional information in the form of prediction uncertainty bounds.

  5. Machinery fault diagnosis using joint global and local/nonlocal discriminant analysis with selective ensemble learning

    Science.gov (United States)

    Yu, Jianbo

    2016-11-01

    The vibration signals of faulty machine are generally non-stationary and nonlinear under those complicated working conditions. Thus, it is a big challenge to extract and select the effective features from vibration signals for machinery fault diagnosis. This paper proposes a new manifold learning algorithm, joint global and local/nonlocal discriminant analysis (GLNDA), which aims to extract effective intrinsic geometrical information from the given vibration data. Comparisons with other regular methods, principal component analysis (PCA), local preserving projection (LPP), linear discriminant analysis (LDA) and local LDA (LLDA), illustrate the superiority of GLNDA in machinery fault diagnosis. Based on the extracted information by GLNDA, a GLNDA-based Fisher discriminant rule (FDR) is put forward and applied to machinery fault diagnosis without additional recognizer construction procedure. By importing Bagging into GLNDA score-based feature selection and FDR, a novel manifold ensemble method (selective GLNDA ensemble, SE-GLNDA) is investigated for machinery fault diagnosis. The motivation for developing ensemble of manifold learning components is that it can achieve higher accuracy and applicability than single component in machinery fault diagnosis. The effectiveness of the SE-GLNDA-based fault diagnosis method has been verified by experimental results from bearing full life testers.

  6. Marine and estuarine natural microbial biofilms: ecological and biogeochemical dimensions

    Directory of Open Access Journals (Sweden)

    O. Roger Anderson

    2016-08-01

    Full Text Available Marine and estuarine microbial biofilms are ubiquitously distributed worldwide and are increasingly of interest in basic and applied sciences because of their unique structural and functional features that make them remarkably different from the biota in the plankton. This is a review of some current scientific knowledge of naturally occurring microbial marine and estuarine biofilms including prokaryotic and microeukaryotic biota, but excluding research specifically on engineering and applied aspects of biofilms such as biofouling. Because the microbial communities including bacteria and protists are integral to the fundamental ecological and biogeochemical processes that support biofilm communities, particular attention is given to the structural and ecological aspects of microbial biofilm formation, succession, and maturation, as well as the dynamics of the interactions of the microbiota in biofilms. The intent is to highlight current state of scientific knowledge and possible avenues of future productive research, especially focusing on the ecological and biogeochemical dimensions.

  7. Development of Gridded Ensemble Precipitation and Temperature Datasets for the Contiguous United States Plus Hawai'i and Alaska

    Science.gov (United States)

    Newman, A. J.; Clark, M. P.; Nijssen, B.; Wood, A.; Gutmann, E. D.; Mizukami, N.; Longman, R. J.; Giambelluca, T. W.; Cherry, J.; Nowak, K.; Arnold, J.; Prein, A. F.

    2016-12-01

    Gridded precipitation and temperature products are inherently uncertain due to myriad factors. These include interpolation from a sparse observation network, measurement representativeness, and measurement errors. Despite this inherent uncertainty, uncertainty is typically not included, or is a specific addition to each dataset without much general applicability across different datasets. A lack of quantitative uncertainty estimates for hydrometeorological forcing fields limits their utility to support land surface and hydrologic modeling techniques such as data assimilation, probabilistic forecasting and verification. To address this gap, we have developed a first of its kind gridded, observation-based ensemble of precipitation and temperature at a daily increment for the period 1980-2012 over the United States (including Alaska and Hawaii). A longer, higher resolution version (1970-present, 1/16th degree) has also been implemented to support real-time hydrologic- monitoring and prediction in several regional US domains. We will present the development and evaluation of the dataset, along with initial applications of the dataset for ensemble data assimilation and probabilistic evaluation of high resolution regional climate model simulations. We will also present results on the new high resolution products for Alaska and Hawaii (2 km and 250 m respectively), to complete the first ensemble observation based product suite for the entire 50 states. Finally, we will present plans to improve the ensemble dataset, focusing on efforts to improve the methods used for station interpolation and ensemble generation, as well as methods to fuse station data with numerical weather prediction model output.

  8. Precision bounds for gradient magnetometry with atomic ensembles

    Science.gov (United States)

    Apellaniz, Iagoba; Urizar-Lanz, Iñigo; Zimborás, Zoltán; Hyllus, Philipp; Tóth, Géza

    2018-05-01

    We study gradient magnetometry with an ensemble of atoms with arbitrary spin. We calculate precision bounds for estimating the gradient of the magnetic field based on the quantum Fisher information. For quantum states that are invariant under homogeneous magnetic fields, we need to measure a single observable to estimate the gradient. On the other hand, for states that are sensitive to homogeneous fields, a simultaneous measurement is needed, as the homogeneous field must also be estimated. We prove that for the cases studied in this paper, such a measurement is feasible. We present a method to calculate precision bounds for gradient estimation with a chain of atoms or with two spatially separated atomic ensembles. We also consider a single atomic ensemble with an arbitrary density profile, where the atoms cannot be addressed individually, and which is a very relevant case for experiments. Our model can take into account even correlations between particle positions. While in most of the discussion we consider an ensemble of localized particles that are classical with respect to their spatial degree of freedom, we also discuss the case of gradient metrology with a single Bose-Einstein condensate.

  9. Evaluation of the NMC regional ensemble prediction system during the Beijing 2008 Olympic Games

    Science.gov (United States)

    Li, Xiaoli; Tian, Hua; Deng, Guo

    2011-10-01

    Based on the B08RDP (Beijing 2008 Olympic Games Mesoscale Ensemble Prediction Research and Development Project) that was launched by the World Weather Research Programme (WWRP) in 2004, a regional ensemble prediction system (REPS) at a 15-km horizontal resolution was developed at the National Meteorological Center (NMC) of the China Meteorological Administration (CMA). Supplementing to the forecasters' subjective affirmation on the promising performance of the REPS during the 2008 Beijing Olympic Games (BOG), this paper focuses on the objective verification of the REPS for precipitation forecasts during the BOG period. By use of a set of advanced probabilistic verification scores, the value of the REPS compared to the quasi-operational global ensemble prediction system (GEPS) is assessed for a 36-day period (21 July-24 August 2008). The evaluation here involves different aspects of the REPS and GEPS, including their general forecast skills, specific attributes (reliability and resolution), and related economic values. The results indicate that the REPS generally performs significantly better for the short-range precipitation forecasts than the GEPS, and for light to heavy rainfall events, the REPS provides more skillful forecasts for accumulated 6- and 24-h precipitation. By further identifying the performance of the REPS through the attribute-focused measures, it is found that the advantages of the REPS over the GEPS come from better reliability (smaller biases and better dispersion) and increased resolution. Also, evaluation of a decision-making score reveals that a much larger group of users benefits from using the REPS forecasts than using the single model (the control run) forecasts, especially for the heavy rainfall events.

  10. A global water resources ensemble of hydrological models: the eartH2Observe Tier-1 dataset

    Science.gov (United States)

    Schellekens, Jaap; Dutra, Emanuel; Martínez-de la Torre, Alberto; Balsamo, Gianpaolo; van Dijk, Albert; Sperna Weiland, Frederiek; Minvielle, Marie; Calvet, Jean-Christophe; Decharme, Bertrand; Eisner, Stephanie; Fink, Gabriel; Flörke, Martina; Peßenteiner, Stefanie; van Beek, Rens; Polcher, Jan; Beck, Hylke; Orth, René; Calton, Ben; Burke, Sophia; Dorigo, Wouter; Weedon, Graham P.

    2017-07-01

    The dataset presented here consists of an ensemble of 10 global hydrological and land surface models for the period 1979-2012 using a reanalysis-based meteorological forcing dataset (0.5° resolution). The current dataset serves as a state of the art in current global hydrological modelling and as a benchmark for further improvements in the coming years. A signal-to-noise ratio analysis revealed low inter-model agreement over (i) snow-dominated regions and (ii) tropical rainforest and monsoon areas. The large uncertainty of precipitation in the tropics is not reflected in the ensemble runoff. Verification of the results against benchmark datasets for evapotranspiration, snow cover, snow water equivalent, soil moisture anomaly and total water storage anomaly using the tools from The International Land Model Benchmarking Project (ILAMB) showed overall useful model performance, while the ensemble mean generally outperformed the single model estimates. The results also show that there is currently no single best model for all variables and that model performance is spatially variable. In our unconstrained model runs the ensemble mean of total runoff into the ocean was 46 268 km3 yr-1 (334 kg m-2 yr-1), while the ensemble mean of total evaporation was 537 kg m-2 yr-1. All data are made available openly through a Water Cycle Integrator portal (WCI, wci.earth2observe.eu), and via a direct http and ftp download. The portal follows the protocols of the open geospatial consortium such as OPeNDAP, WCS and WMS. The DOI for the data is https://doi.org/10.1016/10.5281/zenodo.167070.

  11. A global water resources ensemble of hydrological models: the eartH2Observe Tier-1 dataset

    Directory of Open Access Journals (Sweden)

    J. Schellekens

    2017-07-01

    Full Text Available The dataset presented here consists of an ensemble of 10 global hydrological and land surface models for the period 1979–2012 using a reanalysis-based meteorological forcing dataset (0.5° resolution. The current dataset serves as a state of the art in current global hydrological modelling and as a benchmark for further improvements in the coming years. A signal-to-noise ratio analysis revealed low inter-model agreement over (i snow-dominated regions and (ii tropical rainforest and monsoon areas. The large uncertainty of precipitation in the tropics is not reflected in the ensemble runoff. Verification of the results against benchmark datasets for evapotranspiration, snow cover, snow water equivalent, soil moisture anomaly and total water storage anomaly using the tools from The International Land Model Benchmarking Project (ILAMB showed overall useful model performance, while the ensemble mean generally outperformed the single model estimates. The results also show that there is currently no single best model for all variables and that model performance is spatially variable. In our unconstrained model runs the ensemble mean of total runoff into the ocean was 46 268 km3 yr−1 (334 kg m−2 yr−1, while the ensemble mean of total evaporation was 537 kg m−2 yr−1. All data are made available openly through a Water Cycle Integrator portal (WCI, wci.earth2observe.eu, and via a direct http and ftp download. The portal follows the protocols of the open geospatial consortium such as OPeNDAP, WCS and WMS. The DOI for the data is https://doi.org/10.1016/10.5281/zenodo.167070.

  12. Understanding ensemble protein folding at atomic detail

    International Nuclear Information System (INIS)

    Wallin, Stefan; Shakhnovich, Eugene I

    2008-01-01

    Although far from routine, simulating the folding of specific short protein chains on the computer, at a detailed atomic level, is starting to become a reality. This remarkable progress, which has been made over the last decade or so, allows a fundamental aspect of the protein folding process to be addressed, namely its statistical nature. In order to make quantitative comparisons with experimental kinetic data a complete ensemble view of folding must be achieved, with key observables averaged over the large number of microscopically different folding trajectories available to a protein chain. Here we review recent advances in atomic-level protein folding simulations and the new insight provided by them into the protein folding process. An important element in understanding ensemble folding kinetics are methods for analyzing many separate folding trajectories, and we discuss techniques developed to condense the large amount of information contained in an ensemble of trajectories into a manageable picture of the folding process. (topical review)

  13. Projecting the long-term biogeochemical impacts of a diverse agroforestry system in the Midwest

    Science.gov (United States)

    Wolz, K. J.; DeLucia, E. H.; Paul, R. F.

    2014-12-01

    Annual, monoculture cropping systems have become the standard agricultural model in the Midwestern US. Unintended consequences of these systems include surface and groundwater pollution, greenhouse gas emissions, loss of biodiversity, and soil erosion. Diverse agroforestry (DA) systems dominated by fruit and nut trees/shrubs have been proposed as an agricultural model for the Midwestern US that can restore ecosystem services while simultaneously providing economically viable and industrially relevant staple food crops. A DA system including six species of fruit and nut crops was established on long-time conventional agricultural land at the University of Illinois at Urbana-Champaign in 2012, with the conventional corn-soybean rotation (CSR) as a control. Initial field measurements of the nitrogen and water cycles during the first two years of transition have indicated a significant decrease in N losses and modification of the seasonal evapotranspiration (ET) pattern. While these early results suggest that the land use transition from CSR to DA can have positive biogeochemical consequences, models must be utilized to make long-term biogeochemical projections in agroforestry systems. Initial field measurements of plant phenology, net N2O flux, nitrate leaching, soil respiration, and soil moisture were used to parameterize the DA system within the DayCENT biogeochemical model as the "savanna" ecosystem type. The model was validated with an independent subset of field measurements and then run to project biogeochemical cycling in the DA system for 25 years past establishment. Model results show that N losses via N2O emission or nitrate leaching reach a minimum within the first 5 years and then maintain this tight cycle into the future. While early ET field measurements revealed similar magnitudes between the DA and CSR systems, modeled ET continued to increase for the DA system throughout the projected time since the trees would continue to grow larger. These modeling

  14. Online probabilistic learning with an ensemble of forecasts

    Science.gov (United States)

    Thorey, Jean; Mallet, Vivien; Chaussin, Christophe

    2016-04-01

    Our objective is to produce a calibrated weighted ensemble to forecast a univariate time series. In addition to a meteorological ensemble of forecasts, we rely on observations or analyses of the target variable. The celebrated Continuous Ranked Probability Score (CRPS) is used to evaluate the probabilistic forecasts. However applying the CRPS on weighted empirical distribution functions (deriving from the weighted ensemble) may introduce a bias because of which minimizing the CRPS does not produce the optimal weights. Thus we propose an unbiased version of the CRPS which relies on clusters of members and is strictly proper. We adapt online learning methods for the minimization of the CRPS. These methods generate the weights associated to the members in the forecasted empirical distribution function. The weights are updated before each forecast step using only past observations and forecasts. Our learning algorithms provide the theoretical guarantee that, in the long run, the CRPS of the weighted forecasts is at least as good as the CRPS of any weighted ensemble with weights constant in time. In particular, the performance of our forecast is better than that of any subset ensemble with uniform weights. A noteworthy advantage of our algorithm is that it does not require any assumption on the distributions of the observations and forecasts, both for the application and for the theoretical guarantee to hold. As application example on meteorological forecasts for photovoltaic production integration, we show that our algorithm generates a calibrated probabilistic forecast, with significant performance improvements on probabilistic diagnostic tools (the CRPS, the reliability diagram and the rank histogram).

  15. Improving the ensemble optimization method through covariance matrix adaptation (CMA-EnOpt)

    NARCIS (Netherlands)

    Fonseca, R.M.; Leeuwenburgh, O.; Hof, P.M.J. van den; Jansen, J.D.

    2013-01-01

    Ensemble Optimization (EnOpt) is a rapidly emerging method for reservoir model based production optimization. EnOpt uses an ensemble of controls to approximate the gradient of the objective function with respect to the controls. Current implementations of EnOpt use a Gaussian ensemble with a

  16. Scalable quantum information processing with atomic ensembles and flying photons

    International Nuclear Information System (INIS)

    Mei Feng; Yu Yafei; Feng Mang; Zhang Zhiming

    2009-01-01

    We present a scheme for scalable quantum information processing with atomic ensembles and flying photons. Using the Rydberg blockade, we encode the qubits in the collective atomic states, which could be manipulated fast and easily due to the enhanced interaction in comparison to the single-atom case. We demonstrate that our proposed gating could be applied to generation of two-dimensional cluster states for measurement-based quantum computation. Moreover, the atomic ensembles also function as quantum repeaters useful for long-distance quantum state transfer. We show the possibility of our scheme to work in bad cavity or in weak coupling regime, which could much relax the experimental requirement. The efficient coherent operations on the ensemble qubits enable our scheme to be switchable between quantum computation and quantum communication using atomic ensembles.

  17. diceR: an R package for class discovery using an ensemble driven approach.

    Science.gov (United States)

    Chiu, Derek S; Talhouk, Aline

    2018-01-15

    Given a set of features, researchers are often interested in partitioning objects into homogeneous clusters. In health research, cancer research in particular, high-throughput data is collected with the aim of segmenting patients into sub-populations to aid in disease diagnosis, prognosis or response to therapy. Cluster analysis, a class of unsupervised learning techniques, is often used for class discovery. Cluster analysis suffers from some limitations, including the need to select up-front the algorithm to be used as well as the number of clusters to generate, in addition, there may exist several groupings consistent with the data, making it very difficult to validate a final solution. Ensemble clustering is a technique used to mitigate these limitations and facilitate the generalization and reproducibility of findings in new cohorts of patients. We introduce diceR (diverse cluster ensemble in R), a software package available on CRAN: https://CRAN.R-project.org/package=diceR CONCLUSIONS: diceR is designed to provide a set of tools to guide researchers through a general cluster analysis process that relies on minimizing subjective decision-making. Although developed in a biological context, the tools in diceR are data-agnostic and thus can be applied in different contexts.

  18. Statistical Analysis of Protein Ensembles

    Science.gov (United States)

    Máté, Gabriell; Heermann, Dieter

    2014-04-01

    As 3D protein-configuration data is piling up, there is an ever-increasing need for well-defined, mathematically rigorous analysis approaches, especially that the vast majority of the currently available methods rely heavily on heuristics. We propose an analysis framework which stems from topology, the field of mathematics which studies properties preserved under continuous deformations. First, we calculate a barcode representation of the molecules employing computational topology algorithms. Bars in this barcode represent different topological features. Molecules are compared through their barcodes by statistically determining the difference in the set of their topological features. As a proof-of-principle application, we analyze a dataset compiled of ensembles of different proteins, obtained from the Ensemble Protein Database. We demonstrate that our approach correctly detects the different protein groupings.

  19. Ensemble Forecasts with Useful Skill-Spread Relationships for African meningitis and Asia Streamflow Forecasting

    Science.gov (United States)

    Hopson, T. M.

    2014-12-01

    One potential benefit of an ensemble prediction system (EPS) is its capacity to forecast its own forecast error through the ensemble spread-error relationship. In practice, an EPS is often quite limited in its ability to represent the variable expectation of forecast error through the variable dispersion of the ensemble, and perhaps more fundamentally, in its ability to provide enough variability in the ensembles dispersion to make the skill-spread relationship even potentially useful (irrespective of whether the EPS is well-calibrated or not). In this paper we examine the ensemble skill-spread relationship of an ensemble constructed from the TIGGE (THORPEX Interactive Grand Global Ensemble) dataset of global forecasts and a combination of multi-model and post-processing approaches. Both of the multi-model and post-processing techniques are based on quantile regression (QR) under a step-wise forward selection framework leading to ensemble forecasts with both good reliability and sharpness. The methodology utilizes the ensemble's ability to self-diagnose forecast instability to produce calibrated forecasts with informative skill-spread relationships. A context for these concepts is provided by assessing the constructed ensemble in forecasting district-level humidity impacting the incidence of meningitis in the meningitis belt of Africa, and in forecasting flooding events in the Brahmaputra and Ganges basins of South Asia.

  20. Adiabatic passage and ensemble control of quantum systems

    International Nuclear Information System (INIS)

    Leghtas, Z; Sarlette, A; Rouchon, P

    2011-01-01

    This paper considers population transfer between eigenstates of a finite quantum ladder controlled by a classical electric field. Using an appropriate change of variables, we show that this setting can be set in the framework of adiabatic passage, which is known to facilitate ensemble control of quantum systems. Building on this insight, we present a mathematical proof of robustness for a control protocol-chirped pulse-practised by experimentalists to drive an ensemble of quantum systems from the ground state to the most excited state. We then propose new adiabatic control protocols using a single chirped and amplitude-shaped pulse, to robustly perform any permutation of eigenstate populations, on an ensemble of systems with unknown coupling strengths. These adiabatic control protocols are illustrated by simulations on a four-level ladder.

  1. Conductor and Ensemble Performance Expressivity and State Festival Ratings

    Science.gov (United States)

    Price, Harry E.; Chang, E. Christina

    2005-01-01

    This study is the second in a series examining the relationship between conducting and ensemble performance. The purpose was to further examine the associations among conductor, ensemble performance expressivity, and festival ratings. Participants were asked to rate the expressivity of video-only conducting and parallel audio-only excerpts from a…

  2. Constructing Support Vector Machine Ensembles for Cancer Classification Based on Proteomic Profiling

    Institute of Scientific and Technical Information of China (English)

    Yong Mao; Xiao-Bo Zhou; Dao-Ying Pi; You-Xian Sun

    2005-01-01

    In this study, we present a constructive algorithm for training cooperative support vector machine ensembles (CSVMEs). CSVME combines ensemble architecture design with cooperative training for individual SVMs in ensembles. Unlike most previous studies on training ensembles, CSVME puts emphasis on both accuracy and collaboration among individual SVMs in an ensemble. A group of SVMs selected on the basis of recursive classifier elimination is used in CSVME, and the number of the individual SVMs selected to construct CSVME is determined by 10-fold cross-validation. This kind of SVME has been tested on two ovarian cancer datasets previously obtained by proteomic mass spectrometry. By combining several individual SVMs, the proposed method achieves better performance than the SVME of all base SVMs.

  3. Sequential Ensembles Tolerant to Synthetic Aperture Radar (SAR Soil Moisture Retrieval Errors

    Directory of Open Access Journals (Sweden)

    Ju Hyoung Lee

    2016-04-01

    Full Text Available Due to complicated and undefined systematic errors in satellite observation, data assimilation integrating model states with satellite observations is more complicated than field measurements-based data assimilation at a local scale. In the case of Synthetic Aperture Radar (SAR soil moisture, the systematic errors arising from uncertainties in roughness conditions are significant and unavoidable, but current satellite bias correction methods do not resolve the problems very well. Thus, apart from the bias correction process of satellite observation, it is important to assess the inherent capability of satellite data assimilation in such sub-optimal but more realistic observational error conditions. To this end, time-evolving sequential ensembles of the Ensemble Kalman Filter (EnKF is compared with stationary ensemble of the Ensemble Optimal Interpolation (EnOI scheme that does not evolve the ensembles over time. As the sensitivity analysis demonstrated that the surface roughness is more sensitive to the SAR retrievals than measurement errors, it is a scope of this study to monitor how data assimilation alters the effects of roughness on SAR soil moisture retrievals. In results, two data assimilation schemes all provided intermediate values between SAR overestimation, and model underestimation. However, under the same SAR observational error conditions, the sequential ensembles approached a calibrated model showing the lowest Root Mean Square Error (RMSE, while the stationary ensemble converged towards the SAR observations exhibiting the highest RMSE. As compared to stationary ensembles, sequential ensembles have a better tolerance to SAR retrieval errors. Such inherent nature of EnKF suggests an operational merit as a satellite data assimilation system, due to the limitation of bias correction methods currently available.

  4. Application of a hybrid multiscale approach to simulate hydrologic and biogeochemical processes in the river-groundwater interaction zone.

    Energy Technology Data Exchange (ETDEWEB)

    Hammond, Glenn Edward; Yang, Xiaofan; Song, Xuehang; Song, Hyun-Seob; Hou, Zhangshuan; Chen, Xingyuan; Liu, Yuanyuan; Scheibe, Tim

    2017-03-01

    The groundwater-surface water interaction zone (GSIZ) plays an important role in riverine and watershed ecosystems as the exchange of waters of variable composition and temperature (hydrologic exchange flows) stimulate microbial activity and associated biogeochemical reactions. Variable temporal and spatial scales of hydrologic exchange flows, heterogeneity of the subsurface environment, and complexity of biogeochemical reaction networks in the GSIZ present challenges to incorporation of fundamental process representations and model parameterization across a range of spatial scales (e.g. from pore-scale to field scale). This paper presents a novel hybrid multiscale simulation approach that couples hydrologic-biogeochemical (HBGC) processes between two distinct length scales of interest.

  5. Benchmarking ensemble streamflow prediction skill in the UK

    Science.gov (United States)

    Harrigan, Shaun; Prudhomme, Christel; Parry, Simon; Smith, Katie; Tanguy, Maliko

    2018-03-01

    ; correlation between catchment base flow index (BFI) and ESP skill was very strong (Spearman's rank correlation coefficient = 0.90 at 1-month lead time). This was in contrast to the more highly responsive catchments in the north and west which were generally not skilful at seasonal lead times. Overall, this work provides scientific justification for when and where use of such a relatively simple forecasting approach is appropriate in the UK. This study, furthermore, creates a low cost benchmark against which potential skill improvements from more sophisticated hydro-meteorological ensemble prediction systems can be judged.

  6. Visualizing Confidence in Cluster-Based Ensemble Weather Forecast Analyses.

    Science.gov (United States)

    Kumpf, Alexander; Tost, Bianca; Baumgart, Marlene; Riemer, Michael; Westermann, Rudiger; Rautenhaus, Marc

    2018-01-01

    In meteorology, cluster analysis is frequently used to determine representative trends in ensemble weather predictions in a selected spatio-temporal region, e.g., to reduce a set of ensemble members to simplify and improve their analysis. Identified clusters (i.e., groups of similar members), however, can be very sensitive to small changes of the selected region, so that clustering results can be misleading and bias subsequent analyses. In this article, we - a team of visualization scientists and meteorologists-deliver visual analytics solutions to analyze the sensitivity of clustering results with respect to changes of a selected region. We propose an interactive visual interface that enables simultaneous visualization of a) the variation in composition of identified clusters (i.e., their robustness), b) the variability in cluster membership for individual ensemble members, and c) the uncertainty in the spatial locations of identified trends. We demonstrate that our solution shows meteorologists how representative a clustering result is, and with respect to which changes in the selected region it becomes unstable. Furthermore, our solution helps to identify those ensemble members which stably belong to a given cluster and can thus be considered similar. In a real-world application case we show how our approach is used to analyze the clustering behavior of different regions in a forecast of "Tropical Cyclone Karl", guiding the user towards the cluster robustness information required for subsequent ensemble analysis.

  7. Comparison of standard resampling methods for performance estimation of artificial neural network ensembles

    OpenAIRE

    Green, Michael; Ohlsson, Mattias

    2007-01-01

    Estimation of the generalization performance for classification within the medical applications domain is always an important task. In this study we focus on artificial neural network ensembles as the machine learning technique. We present a numerical comparison between five common resampling techniques: k-fold cross validation (CV), holdout, using three cutoffs, and bootstrap using five different data sets. The results show that CV together with holdout $0.25$ and $0.50$ are the best resampl...

  8. Assessing an ensemble Kalman filter inference of Manning's n coefficient of an idealized tidal inlet against a polynomial chaos-based MCMC

    Science.gov (United States)

    Siripatana, Adil; Mayo, Talea; Sraj, Ihab; Knio, Omar; Dawson, Clint; Le Maitre, Olivier; Hoteit, Ibrahim

    2017-08-01

    Bayesian estimation/inversion is commonly used to quantify and reduce modeling uncertainties in coastal ocean model, especially in the framework of parameter estimation. Based on Bayes rule, the posterior probability distribution function (pdf) of the estimated quantities is obtained conditioned on available data. It can be computed either directly, using a Markov chain Monte Carlo (MCMC) approach, or by sequentially processing the data following a data assimilation approach, which is heavily exploited in large dimensional state estimation problems. The advantage of data assimilation schemes over MCMC-type methods arises from the ability to algorithmically accommodate a large number of uncertain quantities without significant increase in the computational requirements. However, only approximate estimates are generally obtained by this approach due to the restricted Gaussian prior and noise assumptions that are generally imposed in these methods. This contribution aims at evaluating the effectiveness of utilizing an ensemble Kalman-based data assimilation method for parameter estimation of a coastal ocean model against an MCMC polynomial chaos (PC)-based scheme. We focus on quantifying the uncertainties of a coastal ocean ADvanced CIRCulation (ADCIRC) model with respect to the Manning's n coefficients. Based on a realistic framework of observation system simulation experiments (OSSEs), we apply an ensemble Kalman filter and the MCMC method employing a surrogate of ADCIRC constructed by a non-intrusive PC expansion for evaluating the likelihood, and test both approaches under identical scenarios. We study the sensitivity of the estimated posteriors with respect to the parameters of the inference methods, including ensemble size, inflation factor, and PC order. A full analysis of both methods, in the context of coastal ocean model, suggests that an ensemble Kalman filter with appropriate ensemble size and well-tuned inflation provides reliable mean estimates and

  9. Analyzing the impact of changing size and composition of a crop model ensemble

    Science.gov (United States)

    Rodríguez, Alfredo

    2017-04-01

    The use of an ensemble of crop growth simulation models is a practice recently adopted in order to quantify aspects of uncertainties in model simulations. Yet, while the climate modelling community has extensively investigated the properties of model ensembles and their implications, this has hardly been investigated for crop model ensembles (Wallach et al., 2016). In their ensemble of 27 wheat models, Martre et al. (2015) found that the accuracy of the multi-model ensemble-average only increases up to an ensemble size of ca. 10, but does not improve when including more models in the analysis. However, even when this number of members is reached, questions about the impact of the addition or removal of a member to/from the ensemble arise. When selecting ensemble members, identifying members with poor performance or giving implausible results can make a large difference on the outcome. The objective of this study is to set up a methodology that defines indicators to show the effects of changing the ensemble composition and size on simulation results, when a selection procedure of ensemble members is applied. Ensemble mean or median, and variance are measures used to depict ensemble results among other indicators. We are utilizing simulations from an ensemble of wheat models that have been used to construct impact response surfaces (Pirttioja et al., 2015) (IRSs). These show the response of an impact variable (e.g., crop yield) to systematic changes in two explanatory variables (e.g., precipitation and temperature). Using these, we compare different sub-ensembles in terms of the mean, median and spread, and also by comparing IRSs. The methodology developed here allows comparing an ensemble before and after applying any procedure that changes the ensemble composition and size by measuring the impact of this decision on the ensemble central tendency measures. The methodology could also be further developed to compare the effect of changing ensemble composition and size

  10. Ensemble ANNs-PSO-GA Approach for Day-ahead Stock E-exchange Prices Forecasting

    Directory of Open Access Journals (Sweden)

    Yi Xiao

    2013-02-01

    Full Text Available Stock e-exchange prices forecasting is an important financial problem that is receiving increasing attention. This study proposes a novel three-stage nonlinear ensemble model. In the proposed model, three different types of neural-network based models, i.e. Elman network, generalized regression neural network (GRNN and wavelet neural network (WNN are constructed by three non-overlapping training sets and are further optimized by improved particle swarm optimization (IPSO. Finally, a neural-network-based nonlinear meta-model is generated by learning three neural-network based models through support vector machines (SVM neural network. The superiority of the proposed approach lies in its flexibility to account for potentially complex nonlinear relationships. Three daily stock indices time series are used for validating the forecasting model. Empirical results suggest the ensemble ANNs-PSO-GA approach can significantly improve the prediction performance over other individual models and linear combination models listed in this study.

  11. An Improved Ensemble Learning Method for Classifying High-Dimensional and Imbalanced Biomedicine Data.

    Science.gov (United States)

    Yu, Hualong; Ni, Jun

    2014-01-01

    Training classifiers on skewed data can be technically challenging tasks, especially if the data is high-dimensional simultaneously, the tasks can become more difficult. In biomedicine field, skewed data type often appears. In this study, we try to deal with this problem by combining asymmetric bagging ensemble classifier (asBagging) that has been presented in previous work and an improved random subspace (RS) generation strategy that is called feature subspace (FSS). Specifically, FSS is a novel method to promote the balance level between accuracy and diversity of base classifiers in asBagging. In view of the strong generalization capability of support vector machine (SVM), we adopt it to be base classifier. Extensive experiments on four benchmark biomedicine data sets indicate that the proposed ensemble learning method outperforms many baseline approaches in terms of Accuracy, F-measure, G-mean and AUC evaluation criterions, thus it can be regarded as an effective and efficient tool to deal with high-dimensional and imbalanced biomedical data.

  12. Impact of ensemble learning in the assessment of skeletal maturity.

    Science.gov (United States)

    Cunha, Pedro; Moura, Daniel C; Guevara López, Miguel Angel; Guerra, Conceição; Pinto, Daniela; Ramos, Isabel

    2014-09-01

    The assessment of the bone age, or skeletal maturity, is an important task in pediatrics that measures the degree of maturation of children's bones. Nowadays, there is no standard clinical procedure for assessing bone age and the most widely used approaches are the Greulich and Pyle and the Tanner and Whitehouse methods. Computer methods have been proposed to automatize the process; however, there is a lack of exploration about how to combine the features of the different parts of the hand, and how to take advantage of ensemble techniques for this purpose. This paper presents a study where the use of ensemble techniques for improving bone age assessment is evaluated. A new computer method was developed that extracts descriptors for each joint of each finger, which are then combined using different ensemble schemes for obtaining a final bone age value. Three popular ensemble schemes are explored in this study: bagging, stacking and voting. Best results were achieved by bagging with a rule-based regression (M5P), scoring a mean absolute error of 10.16 months. Results show that ensemble techniques improve the prediction performance of most of the evaluated regression algorithms, always achieving best or comparable to best results. Therefore, the success of the ensemble methods allow us to conclude that their use may improve computer-based bone age assessment, offering a scalable option for utilizing multiple regions of interest and combining their output.

  13. Ecosystem services and biogeochemical cycles on a global scale: valuation of water, carbon and nitrogen processes

    International Nuclear Information System (INIS)

    Watanabe, Marcos D.B.; Ortega, Enrique

    2011-01-01

    Ecosystem services (ES) are provided by healthy ecosystems and are fundamental to support human life. However, natural systems have been degraded all over the world and the process of degradation is partially attributed to the lack of knowledge regarding the economic benefits associated with ES, which usually are not captured in the market. To valuate ES without using conventional approaches, such as the human's willingness-to-pay for ecosystem goods and services, this paper uses a different method based on Energy Systems Theory to estimate prices for biogeochemical flows that affect ecosystem services by considering their emergy content converted to equivalent monetary terms. Ecosystem services related to water, carbon and nitrogen biogeochemical flows were assessed since they are connected to a range of final ecosystem services including climate regulation, hydrological regulation, food production, soil formation and others. Results in this paper indicate that aquifer recharge, groundwater flow, carbon dioxide sequestration, methane emission, biological nitrogen fixation, nitrous oxide emission and nitrogen leaching/runoff are the most critical biogeochemical flows in terrestrial systems. Moreover, monetary values related to biogeochemical flows on a global scale could provide important information for policymakers concerned with payment mechanisms for ecosystem services and costs of greenhouse gas emissions.

  14. Limitations of the equivalence between spatial and ensemble estimators in the case of a single-tone excitation.

    Science.gov (United States)

    Monsef, Florian; Cozza, Andrea

    2011-10-01

    The ensemble-average value of the mean-square pressure is often assessed by using the spatial-average technique, underlying an equivalence principle between spatial and ensemble estimators. Using the ideal-diffuse-field model, the accuracy of the spatial-average method has been studied theoretically forty years ago in the case of a single-tone excitation. This study is revisited in the present work on the basis of a more realistic description of the sound field accounting for a finite number of plane waves. The analysis of the spatial-average estimator is based on the study of its convergence rate. Using experimental data from practical examples, it is shown that the classical expression underestimates the estimator uncertainty even for frequencies greater than Schroeder's frequency, and that the number of plane waves may act as lower bound on the spatial-average estimator accuracy. The comparison of the convergence rate with an ensemble-estimator shows that the two statistics cannot be regarded as equivalent in a general case. © 2011 Acoustical Society of America

  15. CMS: Simulated Physical-Biogeochemical Data, SABGOM Model, Gulf of Mexico, 2005-2010

    Data.gov (United States)

    National Aeronautics and Space Administration — This dataset contains monthly mean ocean surface physical and biogeochemical data for the Gulf of Mexico simulated by the South Atlantic Bight and Gulf of Mexico...

  16. Skill assessment of the coupled physical-biogeochemical operational Mediterranean Forecasting System

    Science.gov (United States)

    Cossarini, Gianpiero; Clementi, Emanuela; Salon, Stefano; Grandi, Alessandro; Bolzon, Giorgio; Solidoro, Cosimo

    2016-04-01

    The Mediterranean Monitoring and Forecasting Centre (Med-MFC) is one of the regional production centres of the European Marine Environment Monitoring Service (CMEMS-Copernicus). Med-MFC operatively manages a suite of numerical model systems (3DVAR-NEMO-WW3 and 3DVAR-OGSTM-BFM) that provides gridded datasets of physical and biogeochemical variables for the Mediterranean marine environment with a horizontal resolution of about 6.5 km. At the present stage, the operational Med-MFC produces ten-day forecast: daily for physical parameters and bi-weekly for biogeochemical variables. The validation of the coupled model system and the estimate of the accuracy of model products are key issues to ensure reliable information to the users and the downstream services. Product quality activities at Med-MFC consist of two levels of validation and skill analysis procedures. Pre-operational qualification activities focus on testing the improvement of the quality of a new release of the model system and relays on past simulation and historical data. Then, near real time (NRT) validation activities aim at the routinely and on-line skill assessment of the model forecast and relays on the NRT available observations. Med-MFC validation framework uses both independent (i.e. Bio-Argo float data, in-situ mooring and vessel data of oxygen, nutrients and chlorophyll, moored buoys, tide-gauges and ADCP of temperature, salinity, sea level and velocity) and semi-independent data (i.e. data already used for assimilation, such as satellite chlorophyll, Satellite SLA and SST and in situ vertical profiles of temperature and salinity from XBT, Argo and Gliders) We give evidence that different variables (e.g. CMEMS-products) can be validated at different levels (i.e. at the forecast level or at the level of model consistency) and at different spatial and temporal scales. The fundamental physical parameters temperature, salinity and sea level are routinely validated on daily, weekly and quarterly base

  17. Ensemble of classifiers based network intrusion detection system performance bound

    CSIR Research Space (South Africa)

    Mkuzangwe, Nenekazi NP

    2017-11-01

    Full Text Available This paper provides a performance bound of a network intrusion detection system (NIDS) that uses an ensemble of classifiers. Currently researchers rely on implementing the ensemble of classifiers based NIDS before they can determine the performance...

  18. A Ruby API to query the Ensembl database for genomic features.

    Science.gov (United States)

    Strozzi, Francesco; Aerts, Jan

    2011-04-01

    The Ensembl database makes genomic features available via its Genome Browser. It is also possible to access the underlying data through a Perl API for advanced querying. We have developed a full-featured Ruby API to the Ensembl databases, providing the same functionality as the Perl interface with additional features. A single Ruby API is used to access different releases of the Ensembl databases and is also able to query multi-species databases. Most functionality of the API is provided using the ActiveRecord pattern. The library depends on introspection to make it release independent. The API is available through the Rubygem system and can be installed with the command gem install ruby-ensembl-api.

  19. An Adaptive Approach to Mitigate Background Covariance Limitations in the Ensemble Kalman Filter

    KAUST Repository

    Song, Hajoon

    2010-07-01

    A new approach is proposed to address the background covariance limitations arising from undersampled ensembles and unaccounted model errors in the ensemble Kalman filter (EnKF). The method enhances the representativeness of the EnKF ensemble by augmenting it with new members chosen adaptively to add missing information that prevents the EnKF from fully fitting the data to the ensemble. The vectors to be added are obtained by back projecting the residuals of the observation misfits from the EnKF analysis step onto the state space. The back projection is done using an optimal interpolation (OI) scheme based on an estimated covariance of the subspace missing from the ensemble. In the experiments reported here, the OI uses a preselected stationary background covariance matrix, as in the hybrid EnKF–three-dimensional variational data assimilation (3DVAR) approach, but the resulting correction is included as a new ensemble member instead of being added to all existing ensemble members. The adaptive approach is tested with the Lorenz-96 model. The hybrid EnKF–3DVAR is used as a benchmark to evaluate the performance of the adaptive approach. Assimilation experiments suggest that the new adaptive scheme significantly improves the EnKF behavior when it suffers from small size ensembles and neglected model errors. It was further found to be competitive with the hybrid EnKF–3DVAR approach, depending on ensemble size and data coverage.

  20. Subsurface Biogeochemical Research FY11 Second Quarter Performance Measure

    Energy Technology Data Exchange (ETDEWEB)

    Scheibe, Timothy D.

    2011-03-31

    The Subsurface Biogeochemical Research (SBR) Long Term Measure for 2011 under the Performance Assessment Rating Tool (PART) measure is to "Refine subsurface transport models by developing computational methods to link important processes impacting contaminant transport at smaller scales to the field scale." The second quarter performance measure is to "Provide a report on computational methods linking genome-enabled understanding of microbial metabolism with reactive transport models to describe processes impacting contaminant transport in the subsurface." Microorganisms such as bacteria are by definition small (typically on the order of a micron in size), and their behavior is controlled by their local biogeochemical environment (typically within a single pore or a biofilm on a grain surface, on the order of tens of microns in size). However, their metabolic activity exerts strong influence on the transport and fate of groundwater contaminants of significant concern at DOE sites, in contaminant plumes with spatial extents of meters to kilometers. This report describes progress and key findings from research aimed at integrating models of microbial metabolism based on genomic information (small scale) with models of contaminant fate and transport in aquifers (field scale).

  1. Spam comments prediction using stacking with ensemble learning

    Science.gov (United States)

    Mehmood, Arif; On, Byung-Won; Lee, Ingyu; Ashraf, Imran; Choi, Gyu Sang

    2018-01-01

    Illusive comments of product or services are misleading for people in decision making. The current methodologies to predict deceptive comments are concerned for feature designing with single training model. Indigenous features have ability to show some linguistic phenomena but are hard to reveal the latent semantic meaning of the comments. We propose a prediction model on general features of documents using stacking with ensemble learning. Term Frequency/Inverse Document Frequency (TF/IDF) features are inputs to stacking of Random Forest and Gradient Boosted Trees and the outputs of the base learners are encapsulated with decision tree to make final training of the model. The results exhibits that our approach gives the accuracy of 92.19% which outperform the state-of-the-art method.

  2. Using geochemical indicators to distinguish high biogeochemical activity in floodplain soils and sediments

    Energy Technology Data Exchange (ETDEWEB)

    Kenwell, Amy [Hydrologic Sciences and Engineering Program, Colorado School of Mines, 1500 Illinois St., Golden, CO 80401 (United States); Navarre-Sitchler, Alexis, E-mail: asitchle@mines.edu [Hydrologic Sciences and Engineering Program, Colorado School of Mines, 1500 Illinois St., Golden, CO 80401 (United States); Prugue, Rodrigo [Hydrologic Sciences and Engineering Program, Colorado School of Mines, 1500 Illinois St., Golden, CO 80401 (United States); Spear, John R. [Department of Civil and Environmental Engineering, Colorado School of Mines, 1500 Illinois St., Golden, CO 80401 (United States); Hering, Amanda S. [Department of Applied Mathematics and Statistics, Colorado School of Mines, 1500 Illinois St., Golden, CO 80401 (United States); Maxwell, Reed M. [Hydrologic Sciences and Engineering Program, Colorado School of Mines, 1500 Illinois St., Golden, CO 80401 (United States); Carroll, Rosemary W.H. [Desert Research Institute, Division of Hydrologic Sciences, 2215 Raggio Parkway, Reno, NV 89512 (United States); Williams, Kenneth H. [Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States)

    2016-09-01

    A better understanding of how microbial communities interact with their surroundings in physically and chemically heterogeneous subsurface environments will lead to improved quantification of biogeochemical reactions and associated nutrient cycling. This study develops a methodology to predict potential elevated rates of biogeochemical activity (microbial “hotspots”) in subsurface environments by correlating microbial DNA and aspects of the community structure with the spatial distribution of geochemical indicators in subsurface sediments. Multiple linear regression models of simulated precipitation leachate, HCl and hydroxylamine extractable iron and manganese, total organic carbon (TOC), and microbial community structure were used to identify sample characteristics indicative of biogeochemical hotspots within fluvially-derived aquifer sediments and overlying soils. The method has been applied to (a) alluvial materials collected at a former uranium mill site near Rifle, Colorado and (b) relatively undisturbed floodplain deposits (soils and sediments) collected along the East River near Crested Butte, Colorado. At Rifle, 16 alluvial samples were taken from 8 sediment cores, and at the East River, 46 soil/sediment samples were collected across and perpendicular to 3 active meanders and an oxbow meander. Regression models using TOC and TOC combined with extractable iron and manganese results were determined to be the best fitting statistical models of microbial DNA (via 16S rRNA gene analysis). Fitting these models to observations in both contaminated and natural floodplain deposits, and their associated alluvial aquifers, demonstrates the broad applicability of the geochemical indicator based approach. - Highlights: • Biogeochemical characterization of alluvial floodplain soils and sediments was performed to investigate parameters that may indicate microbial hot spot formation. • A correlation between geochemical parameters (total organic carbon and

  3. Using geochemical indicators to distinguish high biogeochemical activity in floodplain soils and sediments

    International Nuclear Information System (INIS)

    Kenwell, Amy; Navarre-Sitchler, Alexis; Prugue, Rodrigo; Spear, John R.; Hering, Amanda S.; Maxwell, Reed M.; Carroll, Rosemary W.H.; Williams, Kenneth H.

    2016-01-01

    A better understanding of how microbial communities interact with their surroundings in physically and chemically heterogeneous subsurface environments will lead to improved quantification of biogeochemical reactions and associated nutrient cycling. This study develops a methodology to predict potential elevated rates of biogeochemical activity (microbial “hotspots”) in subsurface environments by correlating microbial DNA and aspects of the community structure with the spatial distribution of geochemical indicators in subsurface sediments. Multiple linear regression models of simulated precipitation leachate, HCl and hydroxylamine extractable iron and manganese, total organic carbon (TOC), and microbial community structure were used to identify sample characteristics indicative of biogeochemical hotspots within fluvially-derived aquifer sediments and overlying soils. The method has been applied to (a) alluvial materials collected at a former uranium mill site near Rifle, Colorado and (b) relatively undisturbed floodplain deposits (soils and sediments) collected along the East River near Crested Butte, Colorado. At Rifle, 16 alluvial samples were taken from 8 sediment cores, and at the East River, 46 soil/sediment samples were collected across and perpendicular to 3 active meanders and an oxbow meander. Regression models using TOC and TOC combined with extractable iron and manganese results were determined to be the best fitting statistical models of microbial DNA (via 16S rRNA gene analysis). Fitting these models to observations in both contaminated and natural floodplain deposits, and their associated alluvial aquifers, demonstrates the broad applicability of the geochemical indicator based approach. - Highlights: • Biogeochemical characterization of alluvial floodplain soils and sediments was performed to investigate parameters that may indicate microbial hot spot formation. • A correlation between geochemical parameters (total organic carbon and

  4. Reconstruction of the 1997/1998 El Nino from TOPEX/POSEIDON and TOGA/TAO Data Using a Massively Parallel Pacific-Ocean Model and Ensemble Kalman Filter

    Science.gov (United States)

    Keppenne, C. L.; Rienecker, M.; Borovikov, A. Y.

    1999-01-01

    Two massively parallel data assimilation systems in which the model forecast-error covariances are estimated from the distribution of an ensemble of model integrations are applied to the assimilation of 97-98 TOPEX/POSEIDON altimetry and TOGA/TAO temperature data into a Pacific basin version the NASA Seasonal to Interannual Prediction Project (NSIPP)ls quasi-isopycnal ocean general circulation model. in the first system, ensemble of model runs forced by an ensemble of atmospheric model simulations is used to calculate asymptotic error statistics. The data assimilation then occurs in the reduced phase space spanned by the corresponding leading empirical orthogonal functions. The second system is an ensemble Kalman filter in which new error statistics are computed during each assimilation cycle from the time-dependent ensemble distribution. The data assimilation experiments are conducted on NSIPP's 512-processor CRAY T3E. The two data assimilation systems are validated by withholding part of the data and quantifying the extent to which the withheld information can be inferred from the assimilation of the remaining data. The pros and cons of each system are discussed.

  5. Substrate-specific reorganization of the conformational ensemble of CSK implicates novel modes of kinase function.

    Directory of Open Access Journals (Sweden)

    Michael A Jamros

    Full Text Available Protein kinases use ATP as a phosphoryl donor for the posttranslational modification of signaling targets. It is generally thought that the binding of this nucleotide induces conformational changes leading to closed, more compact forms of the kinase domain that ideally orient active-site residues for efficient catalysis. The kinase domain is oftentimes flanked by additional ligand binding domains that up- or down-regulate catalytic function. C-terminal Src kinase (Csk is a multidomain tyrosine kinase that is up-regulated by N-terminal SH2 and SH3 domains. Although the X-ray structure of Csk suggests the enzyme is compact, X-ray scattering studies indicate that the enzyme possesses both compact and open conformational forms in solution. Here, we investigated whether interactions with the ATP analog AMP-PNP and ADP can shift the conformational ensemble of Csk in solution using a combination of small angle x-ray scattering and molecular dynamics simulations. We find that binding of AMP-PNP shifts the ensemble towards more extended rather than more compact conformations. Binding of ADP further shifts the ensemble towards extended conformations, including highly extended conformations not adopted by the apo protein, nor by the AMP-PNP bound protein. These ensembles indicate that any compaction of the kinase domain induced by nucleotide binding does not extend to the overall multi-domain architecture. Instead, assembly of an ATP-bound kinase domain generates further extended forms of Csk that may have relevance for kinase scaffolding and Src regulation in the cell.

  6. Ensemble prediction of air quality using the WRF/CMAQ model system for health effect studies in China

    Science.gov (United States)

    Hu, Jianlin; Li, Xun; Huang, Lin; Ying, Qi; Zhang, Qiang; Zhao, Bin; Wang, Shuxiao; Zhang, Hongliang

    2017-11-01

    Accurate exposure estimates are required for health effect analyses of severe air pollution in China. Chemical transport models (CTMs) are widely used to provide spatial distribution, chemical composition, particle size fractions, and source origins of air pollutants. The accuracy of air quality predictions in China is greatly affected by the uncertainties of emission inventories. The Community Multiscale Air Quality (CMAQ) model with meteorological inputs from the Weather Research and Forecasting (WRF) model were used in this study to simulate air pollutants in China in 2013. Four simulations were conducted with four different anthropogenic emission inventories, including the Multi-resolution Emission Inventory for China (MEIC), the Emission Inventory for China by School of Environment at Tsinghua University (SOE), the Emissions Database for Global Atmospheric Research (EDGAR), and the Regional Emission inventory in Asia version 2 (REAS2). Model performance of each simulation was evaluated against available observation data from 422 sites in 60 cities across China. Model predictions of O3 and PM2.5 generally meet the model performance criteria, but performance differences exist in different regions, for different pollutants, and among inventories. Ensemble predictions were calculated by linearly combining the results from different inventories to minimize the sum of the squared errors between the ensemble results and the observations in all cities. The ensemble concentrations show improved agreement with observations in most cities. The mean fractional bias (MFB) and mean fractional errors (MFEs) of the ensemble annual PM2.5 in the 60 cities are -0.11 and 0.24, respectively, which are better than the MFB (-0.25 to -0.16) and MFE (0.26-0.31) of individual simulations. The ensemble annual daily maximum 1 h O3 (O3-1h) concentrations are also improved, with mean normalized bias (MNB) of 0.03 and mean normalized errors (MNE) of 0.14, compared to MNB of 0.06-0.19 and

  7. Adaptive calibration of (u,v)‐wind ensemble forecasts

    DEFF Research Database (Denmark)

    Pinson, Pierre

    2012-01-01

    of sufficient reliability. The original framework introduced here allows for an adaptive bivariate calibration of these ensemble forecasts. The originality of this methodology lies in the fact that calibrated ensembles still consist of a set of (space–time) trajectories, after translation and dilation...... of translation and dilation factors are discussed. Copyright © 2012 Royal Meteorological Society...

  8. Spectral properties of embedded Gaussian unitary ensemble of random matrices with Wigner's SU(4) symmetry

    International Nuclear Information System (INIS)

    Vyas, Manan; Kota, V.K.B.

    2010-01-01

    For m fermions in Ω number of single particle orbitals, each fourfold degenerate, we introduce and analyze in detail embedded Gaussian unitary ensemble of random matrices generated by random two-body interactions that are SU(4) scalar [EGUE(2)-SU(4)]. Here the SU(4) algebra corresponds to the Wigner's supermultiplet SU(4) symmetry in nuclei. Embedding algebra for the EGUE(2)-SU(4) ensemble is U(4Ω) contains U(Ω) x SU(4). Exploiting the Wigner-Racah algebra of the embedding algebra, analytical expression for the ensemble average of the product of any two m particle Hamiltonian matrix elements is derived. Using this, formulas for a special class of U(Ω) irreducible representations (irreps) {4 r , p}, p = 0, 1, 2, 3 are derived for the ensemble averaged spectral variances and also for the covariances in energy centroids and spectral variances. On the other hand, simplifying the tabulations of Hecht for SU(Ω) Racah coefficients, numerical calculations are carried out for general U(Ω) irreps. Spectral variances clearly show, by applying Jacquod and Stone prescription, that the EGUE(2)-SU(4) ensemble generates ground state structure just as the quadratic Casimir invariant (C 2 ) of SU(4). This is further corroborated by the calculation of the expectation values of C 2 [SU(4)] and the four periodicity in the ground state energies. Secondly, it is found that the covariances in energy centroids and spectral variances increase in magnitude considerably as we go from EGUE(2) for spinless fermions to EGUE(2) for fermions with spin to EGUE(2)-SU(4) implying that the differences in ensemble and spectral averages grow with increasing symmetry. Also for EGUE(2)-SU(4) there are, unlike for GUE, non-zero cross-correlations in energy centroids and spectral variances defined over spaces with different particle numbers and/or U(Ω) [equivalently SU(4)] irreps. In the dilute limit defined by Ω → ∞, r >> 1 and r/Ω → 0, for the {4 r , p} irreps, we have derived analytical

  9. Correlating phospholipid fatty acids (PLFA) in a landfill leachate polluted aquifer with biogeochemical factors by multivariate statistical methods

    DEFF Research Database (Denmark)

    Ludvigsen, Liselotte; Albrechtsen, Hans-Jørgen; Rootzén, Helle

    1997-01-01

    Different multivariate statistical analyses were applied to phospholipid fatty acids representing the biomass composition and to different biogeochemical parameters measured in 37 samples from a landfill contaminated aquifer at Grindsted Landfill (Denmark). Principal component analysis...... and correspondence analysis were used to identify groups of samples showing similar patterns with respect to biogeochemical variables and phospholipid fatty acid composition. The principal component analysis revealed that for the biogeochemical parameters the first principal component was linked to the pollution...... was used to allocate samples of phospholipid fatty acids into predefined classes. A large percentages of samples were classified correctly when discriminating samples into groups of dissolved organic carbon and specific conductivity, indicating that the biomass is highly influenced by the pollution...

  10. Modeling polydispersive ensembles of diamond nanoparticles

    International Nuclear Information System (INIS)

    Barnard, Amanda S

    2013-01-01

    While significant progress has been made toward production of monodispersed samples of a variety of nanoparticles, in cases such as diamond nanoparticles (nanodiamonds) a significant degree of polydispersivity persists, so scaling-up of laboratory applications to industrial levels has its challenges. In many cases, however, monodispersivity is not essential for reliable application, provided that the inevitable uncertainties are just as predictable as the functional properties. As computational methods of materials design are becoming more widespread, there is a growing need for robust methods for modeling ensembles of nanoparticles, that capture the structural complexity characteristic of real specimens. In this paper we present a simple statistical approach to modeling of ensembles of nanoparticles, and apply it to nanodiamond, based on sets of individual simulations that have been carefully selected to describe specific structural sources that are responsible for scattering of fundamental properties, and that are typically difficult to eliminate experimentally. For the purposes of demonstration we show how scattering in the Fermi energy and the electronic band gap are related to different structural variations (sources), and how these results can be combined strategically to yield statistically significant predictions of the properties of an entire ensemble of nanodiamonds, rather than merely one individual ‘model’ particle or a non-representative sub-set. (paper)

  11. Multivariate localization methods for ensemble Kalman filtering

    KAUST Repository

    Roh, S.

    2015-12-03

    In ensemble Kalman filtering (EnKF), the small number of ensemble members that is feasible to use in a practical data assimilation application leads to sampling variability of the estimates of the background error covariances. The standard approach to reducing the effects of this sampling variability, which has also been found to be highly efficient in improving the performance of EnKF, is the localization of the estimates of the covariances. One family of localization techniques is based on taking the Schur (element-wise) product of the ensemble-based sample covariance matrix and a correlation matrix whose entries are obtained by the discretization of a distance-dependent correlation function. While the proper definition of the localization function for a single state variable has been extensively investigated, a rigorous definition of the localization function for multiple state variables that exist at the same locations has been seldom considered. This paper introduces two strategies for the construction of localization functions for multiple state variables. The proposed localization functions are tested by assimilating simulated observations experiments into the bivariate Lorenz 95 model with their help.

  12. Multivariate localization methods for ensemble Kalman filtering

    KAUST Repository

    Roh, S.

    2015-05-08

    In ensemble Kalman filtering (EnKF), the small number of ensemble members that is feasible to use in a practical data assimilation application leads to sampling variability of the estimates of the background error covariances. The standard approach to reducing the effects of this sampling variability, which has also been found to be highly efficient in improving the performance of EnKF, is the localization of the estimates of the covariances. One family of localization techniques is based on taking the Schur (entry-wise) product of the ensemble-based sample covariance matrix and a correlation matrix whose entries are obtained by the discretization of a distance-dependent correlation function. While the proper definition of the localization function for a single state variable has been extensively investigated, a rigorous definition of the localization function for multiple state variables has been seldom considered. This paper introduces two strategies for the construction of localization functions for multiple state variables. The proposed localization functions are tested by assimilating simulated observations experiments into the bivariate Lorenz 95 model with their help.

  13. Multivariate localization methods for ensemble Kalman filtering

    KAUST Repository

    Roh, S.; Jun, M.; Szunyogh, I.; Genton, Marc G.

    2015-01-01

    In ensemble Kalman filtering (EnKF), the small number of ensemble members that is feasible to use in a practical data assimilation application leads to sampling variability of the estimates of the background error covariances. The standard approach to reducing the effects of this sampling variability, which has also been found to be highly efficient in improving the performance of EnKF, is the localization of the estimates of the covariances. One family of localization techniques is based on taking the Schur (entry-wise) product of the ensemble-based sample covariance matrix and a correlation matrix whose entries are obtained by the discretization of a distance-dependent correlation function. While the proper definition of the localization function for a single state variable has been extensively investigated, a rigorous definition of the localization function for multiple state variables has been seldom considered. This paper introduces two strategies for the construction of localization functions for multiple state variables. The proposed localization functions are tested by assimilating simulated observations experiments into the bivariate Lorenz 95 model with their help.

  14. Multivariate localization methods for ensemble Kalman filtering

    Science.gov (United States)

    Roh, S.; Jun, M.; Szunyogh, I.; Genton, M. G.

    2015-12-01

    In ensemble Kalman filtering (EnKF), the small number of ensemble members that is feasible to use in a practical data assimilation application leads to sampling variability of the estimates of the background error covariances. The standard approach to reducing the effects of this sampling variability, which has also been found to be highly efficient in improving the performance of EnKF, is the localization of the estimates of the covariances. One family of localization techniques is based on taking the Schur (element-wise) product of the ensemble-based sample covariance matrix and a correlation matrix whose entries are obtained by the discretization of a distance-dependent correlation function. While the proper definition of the localization function for a single state variable has been extensively investigated, a rigorous definition of the localization function for multiple state variables that exist at the same locations has been seldom considered. This paper introduces two strategies for the construction of localization functions for multiple state variables. The proposed localization functions are tested by assimilating simulated observations experiments into the bivariate Lorenz 95 model with their help.

  15. Lessons from Climate Modeling on the Design and Use of Ensembles for Crop Modeling

    Science.gov (United States)

    Wallach, Daniel; Mearns, Linda O.; Ruane, Alexander C.; Roetter, Reimund P.; Asseng, Senthold

    2016-01-01

    Working with ensembles of crop models is a recent but important development in crop modeling which promises to lead to better uncertainty estimates for model projections and predictions, better predictions using the ensemble mean or median, and closer collaboration within the modeling community. There are numerous open questions about the best way to create and analyze such ensembles. Much can be learned from the field of climate modeling, given its much longer experience with ensembles. We draw on that experience to identify questions and make propositions that should help make ensemble modeling with crop models more rigorous and informative. The propositions include defining criteria for acceptance of models in a crop MME, exploring criteria for evaluating the degree of relatedness of models in a MME, studying the effect of number of models in the ensemble, development of a statistical model of model sampling, creation of a repository for MME results, studies of possible differential weighting of models in an ensemble, creation of single model ensembles based on sampling from the uncertainty distribution of parameter values or inputs specifically oriented toward uncertainty estimation, the creation of super ensembles that sample more than one source of uncertainty, the analysis of super ensemble results to obtain information on total uncertainty and the separate contributions of different sources of uncertainty and finally further investigation of the use of the multi-model mean or median as a predictor.

  16. Single-particle model of a strongly driven, dense, nanoscale quantum ensemble

    Science.gov (United States)

    DiLoreto, C. S.; Rangan, C.

    2018-01-01

    We study the effects of interatomic interactions on the quantum dynamics of a dense, nanoscale, atomic ensemble driven by a strong electromagnetic field. We use a self-consistent, mean-field technique based on the pseudospectral time-domain method and a full, three-directional basis to solve the coupled Maxwell-Liouville equations. We find that interatomic interactions generate a decoherence in the state of an ensemble on a much faster time scale than the excited-state lifetime of individual atoms. We present a single-particle model of the driven, dense ensemble by incorporating interactions into a dephasing rate. This single-particle model reproduces the essential physics of the full simulation and is an efficient way of rapidly estimating the collective dynamics of a dense ensemble.

  17. Benefits of an ultra large and multiresolution ensemble for estimating available wind power

    Science.gov (United States)

    Berndt, Jonas; Hoppe, Charlotte; Elbern, Hendrik

    2016-04-01

    In this study we investigate the benefits of an ultra large ensemble with up to 1000 members including multiple nesting with a target horizontal resolution of 1 km. The ensemble shall be used as a basis to detect events of extreme errors in wind power forecasting. Forecast value is the wind vector at wind turbine hub height (~ 100 m) in the short range (1 to 24 hour). Current wind power forecast systems rest already on NWP ensemble models. However, only calibrated ensembles from meteorological institutions serve as input so far, with limited spatial resolution (˜10 - 80 km) and member number (˜ 50). Perturbations related to the specific merits of wind power production are yet missing. Thus, single extreme error events which are not detected by such ensemble power forecasts occur infrequently. The numerical forecast model used in this study is the Weather Research and Forecasting Model (WRF). Model uncertainties are represented by stochastic parametrization of sub-grid processes via stochastically perturbed parametrization tendencies and in conjunction via the complementary stochastic kinetic-energy backscatter scheme already provided by WRF. We perform continuous ensemble updates by comparing each ensemble member with available observations using a sequential importance resampling filter to improve the model accuracy while maintaining ensemble spread. Additionally, we use different ensemble systems from global models (ECMWF and GFS) as input and boundary conditions to capture different synoptic conditions. Critical weather situations which are connected to extreme error events are located and corresponding perturbation techniques are applied. The demanding computational effort is overcome by utilising the supercomputer JUQUEEN at the Forschungszentrum Juelich.

  18. Simulated Tempering Distributed Replica Sampling, Virtual Replica Exchange, and Other Generalized-Ensemble Methods for Conformational Sampling.

    Science.gov (United States)

    Rauscher, Sarah; Neale, Chris; Pomès, Régis

    2009-10-13

    Generalized-ensemble algorithms in temperature space have become popular tools to enhance conformational sampling in biomolecular simulations. A random walk in temperature leads to a corresponding random walk in potential energy, which can be used to cross over energetic barriers and overcome the problem of quasi-nonergodicity. In this paper, we introduce two novel methods: simulated tempering distributed replica sampling (STDR) and virtual replica exchange (VREX). These methods are designed to address the practical issues inherent in the replica exchange (RE), simulated tempering (ST), and serial replica exchange (SREM) algorithms. RE requires a large, dedicated, and homogeneous cluster of CPUs to function efficiently when applied to complex systems. ST and SREM both have the drawback of requiring extensive initial simulations, possibly adaptive, for the calculation of weight factors or potential energy distribution functions. STDR and VREX alleviate the need for lengthy initial simulations, and for synchronization and extensive communication between replicas. Both methods are therefore suitable for distributed or heterogeneous computing platforms. We perform an objective comparison of all five algorithms in terms of both implementation issues and sampling efficiency. We use disordered peptides in explicit water as test systems, for a total simulation time of over 42 μs. Efficiency is defined in terms of both structural convergence and temperature diffusion, and we show that these definitions of efficiency are in fact correlated. Importantly, we find that ST-based methods exhibit faster temperature diffusion and correspondingly faster convergence of structural properties compared to RE-based methods. Within the RE-based methods, VREX is superior to both SREM and RE. On the basis of our observations, we conclude that ST is ideal for simple systems, while STDR is well-suited for complex systems.

  19. Assessing an ensemble Kalman filter inference of Manning’s n coefficient of an idealized tidal inlet against a polynomial chaos-based MCMC

    KAUST Repository

    Siripatana, Adil

    2017-06-08

    Bayesian estimation/inversion is commonly used to quantify and reduce modeling uncertainties in coastal ocean model, especially in the framework of parameter estimation. Based on Bayes rule, the posterior probability distribution function (pdf) of the estimated quantities is obtained conditioned on available data. It can be computed either directly, using a Markov chain Monte Carlo (MCMC) approach, or by sequentially processing the data following a data assimilation approach, which is heavily exploited in large dimensional state estimation problems. The advantage of data assimilation schemes over MCMC-type methods arises from the ability to algorithmically accommodate a large number of uncertain quantities without significant increase in the computational requirements. However, only approximate estimates are generally obtained by this approach due to the restricted Gaussian prior and noise assumptions that are generally imposed in these methods. This contribution aims at evaluating the effectiveness of utilizing an ensemble Kalman-based data assimilation method for parameter estimation of a coastal ocean model against an MCMC polynomial chaos (PC)-based scheme. We focus on quantifying the uncertainties of a coastal ocean ADvanced CIRCulation (ADCIRC) model with respect to the Manning’s n coefficients. Based on a realistic framework of observation system simulation experiments (OSSEs), we apply an ensemble Kalman filter and the MCMC method employing a surrogate of ADCIRC constructed by a non-intrusive PC expansion for evaluating the likelihood, and test both approaches under identical scenarios. We study the sensitivity of the estimated posteriors with respect to the parameters of the inference methods, including ensemble size, inflation factor, and PC order. A full analysis of both methods, in the context of coastal ocean model, suggests that an ensemble Kalman filter with appropriate ensemble size and well-tuned inflation provides reliable mean estimates and

  20. Multiscale Investigation on Biofilm Distribution and Its Impact on Macroscopic Biogeochemical Reaction Rates: BIOFILM DISTRIBUTION AND RATE SCALING

    Energy Technology Data Exchange (ETDEWEB)

    Yan, Zhifeng [Institute of Surface-Earth System Science, Tianjin University, Tianjin China; Pacific Northwest National Laboratory, Richland WA USA; Liu, Chongxuan [Pacific Northwest National Laboratory, Richland WA USA; School of Environmental Science and Engineering, Southern University of Science and Technology, Shenzhen China; Liu, Yuanyuan [Pacific Northwest National Laboratory, Richland WA USA; School of Earth Science and Engineering, Nanjing University, Nanjing China; Bailey, Vanessa L. [Pacific Northwest National Laboratory, Richland WA USA

    2017-11-01

    Biofilms are critical locations for biogeochemical reactions in the subsurface environment. The occurrence and distribution of biofilms at microscale as well as their impacts on macroscopic biogeochemical reaction rates are still poorly understood. This paper investigated the formation and distributions of biofilms in heterogeneous sediments using multiscale models, and evaluated the effects of biofilm heterogeneity on local and macroscopic biogeochemical reaction rates. Sediment pore structures derived from X-ray computed tomography were used to simulate the microscale flow dynamics and biofilm distribution in the sediment column. The response of biofilm formation and distribution to the variations in hydraulic and chemical properties was first examined. One representative biofilm distribution was then utilized to evaluate its effects on macroscopic reaction rates using nitrate reduction as an example. The results revealed that microorganisms primarily grew on the surfaces of grains and aggregates near preferential flow paths where both electron donor and acceptor were readily accessible, leading to the heterogeneous distribution of biofilms in the sediments. The heterogeneous biofilm distribution decreased the macroscopic rate of biogeochemical reactions as compared with those in homogeneous cases. Operationally considering the heterogeneous biofilm distribution in macroscopic reactive transport models such as using dual porosity domain concept can significantly improve the prediction of biogeochemical reaction rates. Overall, this study provided important insights into the biofilm formation and distribution in soils and sediments as well as their impacts on the macroscopic manifestation of reaction rates.

  1. Ensemble models of neutrophil trafficking in severe sepsis.

    Directory of Open Access Journals (Sweden)

    Sang Ok Song

    Full Text Available A hallmark of severe sepsis is systemic inflammation which activates leukocytes and can result in their misdirection. This leads to both impaired migration to the locus of infection and increased infiltration into healthy tissues. In order to better understand the pathophysiologic mechanisms involved, we developed a coarse-grained phenomenological model of the acute inflammatory response in CLP (cecal ligation and puncture-induced sepsis in rats. This model incorporates distinct neutrophil kinetic responses to the inflammatory stimulus and the dynamic interactions between components of a compartmentalized inflammatory response. Ensembles of model parameter sets consistent with experimental observations were statistically generated using a Markov-Chain Monte Carlo sampling. Prediction uncertainty in the model states was quantified over the resulting ensemble parameter sets. Forward simulation of the parameter ensembles successfully captured experimental features and predicted that systemically activated circulating neutrophils display impaired migration to the tissue and neutrophil sequestration in the lung, consequently contributing to tissue damage and mortality. Principal component and multiple regression analyses of the parameter ensembles estimated from survivor and non-survivor cohorts provide insight into pathologic mechanisms dictating outcome in sepsis. Furthermore, the model was extended to incorporate hypothetical mechanisms by which immune modulation using extracorporeal blood purification results in improved outcome in septic rats. Simulations identified a sub-population (about 18% of the treated population that benefited from blood purification. Survivors displayed enhanced neutrophil migration to tissue and reduced sequestration of lung neutrophils, contributing to improved outcome. The model ensemble presented herein provides a platform for generating and testing hypotheses in silico, as well as motivating further experimental

  2. Multi-model ensembles for assessment of flood losses and associated uncertainty

    Science.gov (United States)

    Figueiredo, Rui; Schröter, Kai; Weiss-Motz, Alexander; Martina, Mario L. V.; Kreibich, Heidi

    2018-05-01

    Flood loss modelling is a crucial part of risk assessments. However, it is subject to large uncertainty that is often neglected. Most models available in the literature are deterministic, providing only single point estimates of flood loss, and large disparities tend to exist among them. Adopting any one such model in a risk assessment context is likely to lead to inaccurate loss estimates and sub-optimal decision-making. In this paper, we propose the use of multi-model ensembles to address these issues. This approach, which has been applied successfully in other scientific fields, is based on the combination of different model outputs with the aim of improving the skill and usefulness of predictions. We first propose a model rating framework to support ensemble construction, based on a probability tree of model properties, which establishes relative degrees of belief between candidate models. Using 20 flood loss models in two test cases, we then construct numerous multi-model ensembles, based both on the rating framework and on a stochastic method, differing in terms of participating members, ensemble size and model weights. We evaluate the performance of ensemble means, as well as their probabilistic skill and reliability. Our results demonstrate that well-designed multi-model ensembles represent a pragmatic approach to consistently obtain more accurate flood loss estimates and reliable probability distributions of model uncertainty.

  3. The Ensembl Web site: mechanics of a genome browser.

    Science.gov (United States)

    Stalker, James; Gibbins, Brian; Meidl, Patrick; Smith, James; Spooner, William; Hotz, Hans-Rudolf; Cox, Antony V

    2004-05-01

    The Ensembl Web site (http://www.ensembl.org/) is the principal user interface to the data of the Ensembl project, and currently serves >500,000 pages (approximately 2.5 million hits) per week, providing access to >80 GB (gigabyte) of data to users in more than 80 countries. Built atop an open-source platform comprising Apache/mod_perl and the MySQL relational database management system, it is modular, extensible, and freely available. It is being actively reused and extended in several different projects, and has been downloaded and installed in companies and academic institutions worldwide. Here, we describe some of the technical features of the site, with particular reference to its dynamic configuration that enables it to handle disparate data from multiple species.

  4. Evaluation of LDA Ensembles Classifiers for Brain Computer Interface

    International Nuclear Information System (INIS)

    Arjona, Cristian; Pentácolo, José; Gareis, Iván; Atum, Yanina; Gentiletti, Gerardo; Acevedo, Rubén; Rufiner, Leonardo

    2011-01-01

    The Brain Computer Interface (BCI) translates brain activity into computer commands. To increase the performance of the BCI, to decode the user intentions it is necessary to get better the feature extraction and classification techniques. In this article the performance of a three linear discriminant analysis (LDA) classifiers ensemble is studied. The system based on ensemble can theoretically achieved better classification results than the individual counterpart, regarding individual classifier generation algorithm and the procedures for combine their outputs. Classic algorithms based on ensembles such as bagging and boosting are discussed here. For the application on BCI, it was concluded that the generated results using ER and AUC as performance index do not give enough information to establish which configuration is better.

  5. Verification of Ensemble Forecasts for the New York City Operations Support Tool

    Science.gov (United States)

    Day, G.; Schaake, J. C.; Thiemann, M.; Draijer, S.; Wang, L.

    2012-12-01

    The New York City water supply system operated by the Department of Environmental Protection (DEP) serves nine million people. It covers 2,000 square miles of portions of the Catskill, Delaware, and Croton watersheds, and it includes nineteen reservoirs and three controlled lakes. DEP is developing an Operations Support Tool (OST) to support its water supply operations and planning activities. OST includes historical and real-time data, a model of the water supply system complete with operating rules, and lake water quality models developed to evaluate alternatives for managing turbidity in the New York City Catskill reservoirs. OST will enable DEP to manage turbidity in its unfiltered system while satisfying its primary objective of meeting the City's water supply needs, in addition to considering secondary objectives of maintaining ecological flows, supporting fishery and recreation releases, and mitigating downstream flood peaks. The current version of OST relies on statistical forecasts of flows in the system based on recent observed flows. To improve short-term decision making, plans are being made to transition to National Weather Service (NWS) ensemble forecasts based on hydrologic models that account for short-term weather forecast skill, longer-term climate information, as well as the hydrologic state of the watersheds and recent observed flows. To ensure that the ensemble forecasts are unbiased and that the ensemble spread reflects the actual uncertainty of the forecasts, a statistical model has been developed to post-process the NWS ensemble forecasts to account for hydrologic model error as well as any inherent bias and uncertainty in initial model states, meteorological data and forecasts. The post-processor is designed to produce adjusted ensemble forecasts that are consistent with the DEP historical flow sequences that were used to develop the system operating rules. A set of historical hindcasts that is representative of the real-time ensemble

  6. Extracellular Electron Transport Coupling Biogeochemical Processes Centimeters

    DEFF Research Database (Denmark)

    Risgaard-Petersen, Nils; Fossing, Henrik; Christensen, Peter Bondo

    2010-01-01

    of the oxygen uptake in laboratory incubations of initially homogenized and stabilized sediment. Using microsensors and process rate measurements we further investigated the effect of the electric currents on sediment biogeochemistry. Dissolved sulfide readily donated electrons to the networks and could...... confirmed the depth range of the electric communication and indicated donation of electrons directly from organotrophic bacteria. The separation of oxidation and reduction processes created steep pH gradients eventually causing carbonate precipitation at the surface. The results indicate that electron...... exchanging organisms have major biogeochemical importance as they allow widely separated electron donors and acceptors to react with one another....

  7. Distribution of the Largest Eigenvalues of the Levi-Smirnov Ensemble

    International Nuclear Information System (INIS)

    Wieczorek, W.

    2004-01-01

    We calculate the distribution of the k-th largest eigenvalue in the random matrix Levi - Smirnov Ensemble (LSE), using the spectral dualism between LSE and chiral Gaussian Unitary Ensemble (GUE). Then we reconstruct universal spectral oscillations and we investigate an asymptotic behavior of the spectral distribution. (author)

  8. A Single-column Model Ensemble Approach Applied to the TWP-ICE Experiment

    Science.gov (United States)

    Davies, L.; Jakob, C.; Cheung, K.; DelGenio, A.; Hill, A.; Hume, T.; Keane, R. J.; Komori, T.; Larson, V. E.; Lin, Y.; hide

    2013-01-01

    Single-column models (SCM) are useful test beds for investigating the parameterization schemes of numerical weather prediction and climate models. The usefulness of SCM simulations are limited, however, by the accuracy of the best estimate large-scale observations prescribed. Errors estimating the observations will result in uncertainty in modeled simulations. One method to address the modeled uncertainty is to simulate an ensemble where the ensemble members span observational uncertainty. This study first derives an ensemble of large-scale data for the Tropical Warm Pool International Cloud Experiment (TWP-ICE) based on an estimate of a possible source of error in the best estimate product. These data are then used to carry out simulations with 11 SCM and two cloud-resolving models (CRM). Best estimate simulations are also performed. All models show that moisture-related variables are close to observations and there are limited differences between the best estimate and ensemble mean values. The models, however, show different sensitivities to changes in the forcing particularly when weakly forced. The ensemble simulations highlight important differences in the surface evaporation term of the moisture budget between the SCM and CRM. Differences are also apparent between the models in the ensemble mean vertical structure of cloud variables, while for each model, cloud properties are relatively insensitive to forcing. The ensemble is further used to investigate cloud variables and precipitation and identifies differences between CRM and SCM particularly for relationships involving ice. This study highlights the additional analysis that can be performed using ensemble simulations and hence enables a more complete model investigation compared to using the more traditional single best estimate simulation only.

  9. ENSEMBLE methods to reconcile disparate national long range dispersion forecasting

    Energy Technology Data Exchange (ETDEWEB)

    Mikkelsen, T; Galmarini, S; Bianconi, R; French, S [eds.

    2003-11-01

    ENSEMBLE is a web-based decision support system for real-time exchange and evaluation of national long-range dispersion forecasts of nuclear releases with cross-boundary consequences. The system is developed with the purpose to reconcile among disparate national forecasts for long-range dispersion. ENSEMBLE addresses the problem of achieving a common coherent strategy across European national emergency management when national long-range dispersion forecasts differ from one another during an accidental atmospheric release of radioactive material. A series of new decision-making 'ENSEMBLE' procedures and Web-based software evaluation and exchange tools have been created for real-time reconciliation and harmonisation of real-time dispersion forecasts from meteorological and emergency centres across Europe during an accident. The new ENSEMBLE software tools is available to participating national emergency and meteorological forecasting centres, which may choose to integrate them directly into operational emergency information systems, or possibly use them as a basis for future system development. (au)

  10. ENSEMBLE methods to reconcile disparate national long range dispersion forecasting

    Energy Technology Data Exchange (ETDEWEB)

    Mikkelsen, T.; Galmarini, S.; Bianconi, R.; French, S. (eds.)

    2003-11-01

    ENSEMBLE is a web-based decision support system for real-time exchange and evaluation of national long-range dispersion forecasts of nuclear releases with cross-boundary consequences. The system is developed with the purpose to reconcile among disparate national forecasts for long-range dispersion. ENSEMBLE addresses the problem of achieving a common coherent strategy across European national emergency management when national long-range dispersion forecasts differ from one another during an accidental atmospheric release of radioactive material. A series of new decision-making 'ENSEMBLE' procedures and Web-based software evaluation and exchange tools have been created for real-time reconciliation and harmonisation of real-time dispersion forecasts from meteorological and emergency centres across Europe during an accident. The new ENSEMBLE software tools is available to participating national emergency and meteorological forecasting centres, which may choose to integrate them directly into operational emergency information systems, or possibly use them as a basis for future system development. (au)

  11. A study of fuzzy logic ensemble system performance on face recognition problem

    Science.gov (United States)

    Polyakova, A.; Lipinskiy, L.

    2017-02-01

    Some problems are difficult to solve by using a single intelligent information technology (IIT). The ensemble of the various data mining (DM) techniques is a set of models which are able to solve the problem by itself, but the combination of which allows increasing the efficiency of the system as a whole. Using the IIT ensembles can improve the reliability and efficiency of the final decision, since it emphasizes on the diversity of its components. The new method of the intellectual informational technology ensemble design is considered in this paper. It is based on the fuzzy logic and is designed to solve the classification and regression problems. The ensemble consists of several data mining algorithms: artificial neural network, support vector machine and decision trees. These algorithms and their ensemble have been tested by solving the face recognition problems. Principal components analysis (PCA) is used for feature selection.

  12. Tweet-based Target Market Classification Using Ensemble Method

    Directory of Open Access Journals (Sweden)

    Muhammad Adi Khairul Anshary

    2016-09-01

    Full Text Available Target market classification is aimed at focusing marketing activities on the right targets. Classification of target markets can be done through data mining and by utilizing data from social media, e.g. Twitter. The end result of data mining are learning models that can classify new data. Ensemble methods can improve the accuracy of the models and therefore provide better results. In this study, classification of target markets was conducted on a dataset of 3000 tweets in order to extract features. Classification models were constructed to manipulate the training data using two ensemble methods (bagging and boosting. To investigate the effectiveness of the ensemble methods, this study used the CART (classification and regression tree algorithm for comparison. Three categories of consumer goods (computers, mobile phones and cameras and three categories of sentiments (positive, negative and neutral were classified towards three target-market categories. Machine learning was performed using Weka 3.6.9. The results of the test data showed that the bagging method improved the accuracy of CART with 1.9% (to 85.20%. On the other hand, for sentiment classification, the ensemble methods were not successful in increasing the accuracy of CART. The results of this study may be taken into consideration by companies who approach their customers through social media, especially Twitter.

  13. Notes on Well-Posed, Ensemble Averaged Conservation Equations for Multiphase, Multi-Component, and Multi-Material Flows

    International Nuclear Information System (INIS)

    Ray A. Berry

    2005-01-01

    ensemble averaging can also be used to produce the governing equation systems. In fact volume and time averaging can be viewed as special cases of ensemble averaging. Ensemble averaging is beginning to gain some notice, for example the general-purpose multi-material flow simulation code CFDLib under continuing developed at the Los Alamos National Laboratory [Kashiwa and Rauenzahn 1994] is based on an ensemble averaged formulation. The purpose of this short note is to give an introduction to the ensemble averaging methodology and to show how ensemble averaged balance equations and entropy inequality can be obtained from the microscopic balances. It then details some seven-equation, two-pressure, two-velocity hyperbolic, well-posed models for two-phase flows. Lastly, a simple example is presented of a model in which the flow consists of two barotropic fluids with no phase change in which an equilibrium pressure equation is obtained in the spirit of pressure-based methods of computational fluid dynamics

  14. Biogeochemical cycles of Chernobyl-born radionuclides in the contaminated forest ecosystems: long-term dynamics of the migration processes

    Science.gov (United States)

    Shcheglov, Alexey; Tsvetnova, Ol'ga; Klyashtorin, Alexey

    2013-04-01

    Biogeochemical migration is a dominant factor of the radionuclide transport through the biosphere. In the early XX century, V.I. Vernadskii, a Russian scientist known, noted about a special role living things play in transport and accumulation of natural radionuclide in various environments. The role of biogeochemical processes in migration and redistribution of technogenic radionuclides is not less important. In Russia, V. M. Klechkovskii and N.V. Timofeev-Ressovskii showed some important biogeochemical aspects of radionuclide migration by the example of global fallout and Kyshtym accident. Their followers, R.M. Alexakhin, M.A. Naryshkin, N.V. Kulikov, F.A. Tikhomirov, E.B. Tyuryukanova, and others also contributed a lot to biogeochemistry of radionuclides. In the post-Chernobyl period, this area of knowledge received a lot of data that allowed building the radioactive element balance and flux estimation in various biogeochemical cycles [Shcheglov et al., 1999]. Regrettably, many of recent radioecological studies are only focused on specific radionuclide fluxes or pursue some applied tasks, missing the holistic approach. Most of the studies consider biogeochemical fluxes of radioactive isotopes in terms of either dose estimation or radionuclide migration rates in various food chains. However, to get a comprehensive picture and develop a reliable forecast of environmental, ecological, and social consequences of radioactive pollution in a vast contaminated area, it is necessary to investigate all the radionuclide fluxes associated with the biogeochemical cycles in affected ecosystems. We believe such an integrated approach would be useful to study long-term environmental consequences of the Fukushima accident as well. In our long-term research, we tried to characterize the flux dynamics of the Chernobyl-born radionuclides in the contaminated forest ecosystems and landscapes as a part of the integrated biogeochemical process. Our field studies were started in June of

  15. Reanalysis of biogeochemical properties in the Mediterranean Sea

    Science.gov (United States)

    Cossarini, Gianpiero; Teruzzi, Anna; Salon, Stefano; Solidoro, Cosimo

    2014-05-01

    In the 3D variational (3DVAR) assimilation approach the error covariance matrix can be decomposed in a series of operators. The decomposition makes the 3DVAR particularly suitable for marine biogeochemistry data assimilation, because of the reduced computational costs of the method and its modularity, which allows to define the covariance among the biogeochemical variables in a specific operator. In the present work, the results of 3DVAR assimilation of surface chlorophyll concentration in a multi-annual simulation of the Mediterranean Sea biogeochemistry are presented. The assimilated chlorophyll concentrations are obtained from satellite observations (Volpe et al. 2012). The multi-annual simulation is carried out using the OPATM-BFM model (Lazzari et al. 2012), which describes the low trophic web dynamics and is offline coupled with the MFS physical model (Oddo et al. 2009). In the OPATM-BFM four types of phytoplankton are simulated in terms of their content in carbon, nitrogen, phosphorous, silicon and chlorophyll. In the 3DVAR the error covariance matrix has been decomposed in three different operators, which account for the vertical, the horizontal and the biogeochemical covariance (Teruzzi et al. 2014). The biogeochemical operator propagates the result of the assimilation to the OPATM-BFM variables, providing innovation for the components of the four phytoplankton types. The biogeochemical covariance has been designed supposing that the assimilation preserves the physiological status and the relative abundances of phytoplankton types. Practically, the assimilation preserves the internal quotas of the components for each phytoplankton as long as the optimal growth rate condition are maintained. The quotas preservation is not applied when the phytoplankton is in severe declining growth phase, and the correction provided by the assimilation is set equal to zero. Moreover, the relative abundances among the phytoplankton functional types are preserved. The 3DVAR

  16. Ocean heat content variability and change in an ensemble of ocean reanalyses

    Science.gov (United States)

    Palmer, M. D.; Roberts, C. D.; Balmaseda, M.; Chang, Y.-S.; Chepurin, G.; Ferry, N.; Fujii, Y.; Good, S. A.; Guinehut, S.; Haines, K.; Hernandez, F.; Köhl, A.; Lee, T.; Martin, M. J.; Masina, S.; Masuda, S.; Peterson, K. A.; Storto, A.; Toyoda, T.; Valdivieso, M.; Vernieres, G.; Wang, O.; Xue, Y.

    2017-08-01

    heat uptake of approximately 0.9 ± 0.8 W m-2 (expressed relative to Earth's surface area) between 1995 and 2002, which reduces to about 0.2 ± 0.6 W m-2 between 2004 and 2006, in qualitative agreement with recent analysis of Earth's energy imbalance. There is a marked reduction in the ensemble spread of OHC trends below 300 m as the Argo profiling float observations become available in the early 2000s. In general, we suggest that ORAs should be treated with caution when employed to understand past ocean warming trends—especially when considering the deeper ocean where there is little in the way of observational constraints. The current work emphasizes the need to better observe the deep ocean, both for providing observational constraints for future ocean state estimation efforts and also to develop improved models and data assimilation methods.

  17. Geophysical Monitoring of Hydrological and Biogeochemical Transformations associated with Cr(VI) Bioremediation

    International Nuclear Information System (INIS)

    Hubbard, Susan; Williams, Kenneth H.; Conrad, Mark E.; Faybishenko, Boris; Peterson, John; Chen, Jinsong; Long, Philip E.; Hazen, Terry C.

    2008-01-01

    Understanding how hydrological and biogeochemical properties change over space and time in response to remedial treatments is hindered by our ability to monitor these processes with sufficient resolution and over field relevant scales. Here, we explored the use of geophysical approaches for monitoring the spatiotemporal distribution of hydrological and biogeochemical transformations associated with a Cr(VI)bioremediation experiment performed at Hanford, WA. We first integrated hydrological wellbore and geophysical tomographic datasets to estimate hydrological zonation at the study site. Using results from laboratory biogeophysical experiments and constraints provided by field geochemical datasets, we then interpreted time-lapse seismic and radar tomographic datasets, collected during thirteen acquisition campaigns over a three year experimental period, in terms of hydrological and biogeochemical transformations. The geophysical monitoring datasets were used to infer: the spatial distribution of injected electron donor; the evolution of gas bubbles; variations in total dissolved solids (nitrate and sulfate) as a function of pumping activity; the formation of precipitates and dissolution of calcites; and concomitant changes in porosity. Although qualitative in nature, the integrated interpretation illustrates how geophysical techniques have the potential to provide a wealth of information about coupled hydrobiogeochemical responses to remedial treatments in high spatial resolution and in a minimally invasive manner. Particularly novel aspects of our study include the use of multiple lines of evidence to constrain the interpretation of a long-term, field-scale geophysical monitoring dataset and the interpretation of the transformations as a function of hydrological heterogeneity and pumping activity

  18. Error assessment of biogeochemical models by lower bound methods (NOMMA-1.0

    Directory of Open Access Journals (Sweden)

    V. Sauerland

    2018-03-01

    Full Text Available Biogeochemical models, capturing the major feedbacks of the pelagic ecosystem of the world ocean, are today often embedded into Earth system models which are increasingly used for decision making regarding climate policies. These models contain poorly constrained parameters (e.g., maximum phytoplankton growth rate, which are typically adjusted until the model shows reasonable behavior. Systematic approaches determine these parameters by minimizing the misfit between the model and observational data. In most common model approaches, however, the underlying functions mimicking the biogeochemical processes are nonlinear and non-convex. Thus, systematic optimization algorithms are likely to get trapped in local minima and might lead to non-optimal results. To judge the quality of an obtained parameter estimate, we propose determining a preferably large lower bound for the global optimum that is relatively easy to obtain and that will help to assess the quality of an optimum, generated by an optimization algorithm. Due to the unavoidable noise component in all observations, such a lower bound is typically larger than zero. We suggest deriving such lower bounds based on typical properties of biogeochemical models (e.g., a limited number of extremes and a bounded time derivative. We illustrate the applicability of the method with two real-world examples. The first example uses real-world observations of the Baltic Sea in a box model setup. The second example considers a three-dimensional coupled ocean circulation model in combination with satellite chlorophyll a.

  19. Error assessment of biogeochemical models by lower bound methods (NOMMA-1.0)

    Science.gov (United States)

    Sauerland, Volkmar; Löptien, Ulrike; Leonhard, Claudine; Oschlies, Andreas; Srivastav, Anand

    2018-03-01

    Biogeochemical models, capturing the major feedbacks of the pelagic ecosystem of the world ocean, are today often embedded into Earth system models which are increasingly used for decision making regarding climate policies. These models contain poorly constrained parameters (e.g., maximum phytoplankton growth rate), which are typically adjusted until the model shows reasonable behavior. Systematic approaches determine these parameters by minimizing the misfit between the model and observational data. In most common model approaches, however, the underlying functions mimicking the biogeochemical processes are nonlinear and non-convex. Thus, systematic optimization algorithms are likely to get trapped in local minima and might lead to non-optimal results. To judge the quality of an obtained parameter estimate, we propose determining a preferably large lower bound for the global optimum that is relatively easy to obtain and that will help to assess the quality of an optimum, generated by an optimization algorithm. Due to the unavoidable noise component in all observations, such a lower bound is typically larger than zero. We suggest deriving such lower bounds based on typical properties of biogeochemical models (e.g., a limited number of extremes and a bounded time derivative). We illustrate the applicability of the method with two real-world examples. The first example uses real-world observations of the Baltic Sea in a box model setup. The second example considers a three-dimensional coupled ocean circulation model in combination with satellite chlorophyll a.

  20. Programming in the Zone: Repertoire Selection for the Large Ensemble

    Science.gov (United States)

    Hopkins, Michael

    2013-01-01

    One of the great challenges ensemble directors face is selecting high-quality repertoire that matches the musical and technical levels of their ensembles. Thoughtful repertoire selection can lead to increased student motivation as well as greater enthusiasm for the music program from parents, administrators, teachers, and community members. Common…

  1. A decade of physical and biogeochemical measurements in the Northern Indian Ocean.

    Digital Repository Service at National Institute of Oceanography (India)

    PrasannaKumar, S.; Sardesai, S.; Ramaiah, N.

    at understanding the seasonal variability of physical and biogeochemical parameters. The results showed strongest seasonal cycle in the Arabian Sea with blooms during summer and winter. Upwelling, advection and wind-mixing drive the summer bloom, while the winter...

  2. Improved ensemble-mean forecast skills of ENSO events by a zero-mean stochastic model-error model of an intermediate coupled model

    Science.gov (United States)

    Zheng, F.; Zhu, J.

    2015-12-01

    To perform an ensemble-based ENSO probabilistic forecast, the crucial issue is to design a reliable ensemble prediction strategy that should include the major uncertainties of a forecast system. In this study, we developed a new general ensemble perturbation technique to improve the ensemble-mean predictive skill of forecasting ENSO using an intermediate coupled model (ICM). The model uncertainties are first estimated and analyzed from EnKF analysis results through assimilating observed SST. Then, based on the pre-analyzed properties of the model errors, a zero-mean stochastic model-error model is developed to mainly represent the model uncertainties induced by some important physical processes missed in the coupled model (i.e., stochastic atmospheric forcing/MJO, extra-tropical cooling and warming, Indian Ocean Dipole mode, etc.). Each member of an ensemble forecast is perturbed by the stochastic model-error model at each step during the 12-month forecast process, and the stochastical perturbations are added into the modeled physical fields to mimic the presence of these high-frequency stochastic noises and model biases and their effect on the predictability of the coupled system. The impacts of stochastic model-error perturbations on ENSO deterministic predictions are examined by performing two sets of 21-yr retrospective forecast experiments. The two forecast schemes are differentiated by whether they considered the model stochastic perturbations, with both initialized by the ensemble-mean analysis states from EnKF. The comparison results suggest that the stochastic model-error perturbations have significant and positive impacts on improving the ensemble-mean prediction skills during the entire 12-month forecast process. Because the nonlinear feature of the coupled model can induce the nonlinear growth of the added stochastic model errors with model integration, especially through the nonlinear heating mechanism with the vertical advection term of the model, the

  3. A Brief Tutorial on the Ensemble Kalman Filter

    OpenAIRE

    Mandel, Jan

    2009-01-01

    The ensemble Kalman filter (EnKF) is a recursive filter suitable for problems with a large number of variables, such as discretizations of partial differential equations in geophysical models. The EnKF originated as a version of the Kalman filter for large problems (essentially, the covariance matrix is replaced by the sample covariance), and it is now an important data assimilation component of ensemble forecasting. EnKF is related to the particle filter (in this context, a particle is the s...

  4. A comparison of resampling schemes for estimating model observer performance with small ensembles

    Science.gov (United States)

    Elshahaby, Fatma E. A.; Jha, Abhinav K.; Ghaly, Michael; Frey, Eric C.

    2017-09-01

    In objective assessment of image quality, an ensemble of images is used to compute the 1st and 2nd order statistics of the data. Often, only a finite number of images is available, leading to the issue of statistical variability in numerical observer performance. Resampling-based strategies can help overcome this issue. In this paper, we compared different combinations of resampling schemes (the leave-one-out (LOO) and the half-train/half-test (HT/HT)) and model observers (the conventional channelized Hotelling observer (CHO), channelized linear discriminant (CLD) and channelized quadratic discriminant). Observer performance was quantified by the area under the ROC curve (AUC). For a binary classification task and for each observer, the AUC value for an ensemble size of 2000 samples per class served as a gold standard for that observer. Results indicated that each observer yielded a different performance depending on the ensemble size and the resampling scheme. For a small ensemble size, the combination [CHO, HT/HT] had more accurate rankings than the combination [CHO, LOO]. Using the LOO scheme, the CLD and CHO had similar performance for large ensembles. However, the CLD outperformed the CHO and gave more accurate rankings for smaller ensembles. As the ensemble size decreased, the performance of the [CHO, LOO] combination seriously deteriorated as opposed to the [CLD, LOO] combination. Thus, it might be desirable to use the CLD with the LOO scheme when smaller ensemble size is available.

  5. NATO Advanced Research Workshop on The Biogeochemical Cycling of Sulfur and Nitrogen in the Remote Atmosphere

    CERN Document Server

    Charlson, Robert; Andreae, Meinrat; Rodhe, Henning

    1985-01-01

    Viewed from space, the Earth appears as a globe without a beginning or an end. Encompassing the globe is the atmosphere with its three phases-­ gaseous, liquid, and solid--moving in directions influenced by sunlight, gravity, and rotation. The chemical compositions of these phases are determined by biogeochemical cycles. Over the past hundred years, the processes governing the rates and reactions in the atmospheric biogeochemical cycles have typically been studied in regions where scientists lived. Hence, as time has gone by, the advances in our knowledge of atmospheric chemical cycles in remote areas have lagged substantially behind those for more populated areas. Not only are the data less abundant, they are also scattered. Therefore, we felt a workshop would be an excellent mechanism to assess the state­ of-knowledge of the atmospheric cycles of sulfur and nitrogen in remote areas and to make recommendations for future research. Thus, a NATO Advanced Research Workshop '~he Biogeochemical Cycling of Sulfu...

  6. Weight Distribution for Non-binary Cluster LDPC Code Ensemble

    Science.gov (United States)

    Nozaki, Takayuki; Maehara, Masaki; Kasai, Kenta; Sakaniwa, Kohichi

    In this paper, we derive the average weight distributions for the irregular non-binary cluster low-density parity-check (LDPC) code ensembles. Moreover, we give the exponential growth rate of the average weight distribution in the limit of large code length. We show that there exist $(2,d_c)$-regular non-binary cluster LDPC code ensembles whose normalized typical minimum distances are strictly positive.

  7. Hole digging in ensembles of tunneling molecular magnets

    Science.gov (United States)

    Tupitsyn, I. S.; Stamp, P. C.; Prokof'ev, N. V.

    2004-04-01

    The nuclear spin-mediated quantum relaxation of ensembles of tunneling magnetic molecules causes a “hole” to appear in the distribution of internal fields in the system. The form of this hole and its time evolution, are studied using Monte Carlo simulations. It is shown that the line shape of the tunneling hole in a partially depolarized sample must have a Lorentzian line shape. The short-time half-width ξo in Fe8 crystals should be ˜E0, the half-width of the nuclear spin multiplet, but this result is not generally true. The Lorentzian hole line shape and the short-time √(t) relaxation in weakly polarized samples are both connected to a correlation time τde(ξ) for bias diffusion, whose inverse value also has a Lorentzian dependence on ξ.

  8. Statistical mechanics of Fermi-Pasta-Ulam chains with the canonical ensemble

    Science.gov (United States)

    Demirel, Melik C.; Sayar, Mehmet; Atılgan, Ali R.

    1997-03-01

    Low-energy vibrations of a Fermi-Pasta-Ulam-Β (FPU-Β) chain with 16 repeat units are analyzed with the aid of numerical experiments and the statistical mechanics equations of the canonical ensemble. Constant temperature numerical integrations are performed by employing the cubic coupling scheme of Kusnezov et al. [Ann. Phys. 204, 155 (1990)]. Very good agreement is obtained between numerical results and theoretical predictions for the probability distributions of the generalized coordinates and momenta both of the chain and of the thermal bath. It is also shown that the average energy of the chain scales linearly with the bath temperature.

  9. Biogeochemical control points in a water-limited critical zone

    Science.gov (United States)

    Chorover, J.; Brooks, P. D.; Gallery, R. E.; McIntosh, J. C.; Olshansky, Y.; Rasmussen, C.

    2017-12-01

    The routing of water and carbon through complex terrain is postulated to control structure evolution in the sub-humid critical zone of the southwestern US. By combining measurements of land-atmosphere exchange, ecohydrologic partitioning, and subsurface biogeochemistry, we seek to quantify how a heterogeneous (in time and space) distribution of "reactants" impacts both short-term (sub-)catchment response (e.g., pore and surface water chemical dynamics) and long-term landscape evolution (e.g., soil geochemistry/morphology and regolith weathering depth) in watersheds underlain by rhyolite and schist. Instrumented pedons in convergent, planar, and divergent landscape positions show distinct depth-dependent responses to precipitation events. Wetting front propagation, dissolved carbon flux and associated biogeochemical responses (e.g., pulses of CO2 production, O2 depletion, solute release) vary with topography, revealing the influence of lateral subsidies of water and carbon. The impacts of these episodes on the evolution of porous media heterogeneity is being investigated by statistical analysis of pore water chemistry, chemical/spectroscopic studies of solid phase organo-mineral products, sensor-derived water characteristic curves, and quantification of co-located microbial community activity/composition. Our results highlight the interacting effects of critical zone structure and convergent hydrologic flows in the evolution of biogeochemical control points.

  10. Skill of Global Raw and Postprocessed Ensemble Predictions of Rainfall over Northern Tropical Africa

    Science.gov (United States)

    Vogel, Peter; Knippertz, Peter; Fink, Andreas H.; Schlueter, Andreas; Gneiting, Tilmann

    2018-04-01

    Accumulated precipitation forecasts are of high socioeconomic importance for agriculturally dominated societies in northern tropical Africa. In this study, we analyze the performance of nine operational global ensemble prediction systems (EPSs) relative to climatology-based forecasts for 1 to 5-day accumulated precipitation based on the monsoon seasons 2007-2014 for three regions within northern tropical Africa. To assess the full potential of raw ensemble forecasts across spatial scales, we apply state-of-the-art statistical postprocessing methods in form of Bayesian Model Averaging (BMA) and Ensemble Model Output Statistics (EMOS), and verify against station and spatially aggregated, satellite-based gridded observations. Raw ensemble forecasts are uncalibrated, unreliable, and underperform relative to climatology, independently of region, accumulation time, monsoon season, and ensemble. Differences between raw ensemble and climatological forecasts are large, and partly stem from poor prediction for low precipitation amounts. BMA and EMOS postprocessed forecasts are calibrated, reliable, and strongly improve on the raw ensembles, but - somewhat disappointingly - typically do not outperform climatology. Most EPSs exhibit slight improvements over the period 2007-2014, but overall have little added value compared to climatology. We suspect that the parametrization of convection is a potential cause for the sobering lack of ensemble forecast skill in a region dominated by mesoscale convective systems.

  11. A novel least squares support vector machine ensemble model for NOx emission prediction of a coal-fired boiler

    International Nuclear Information System (INIS)

    Lv, You; Liu, Jizhen; Yang, Tingting; Zeng, Deliang

    2013-01-01

    Real operation data of power plants are inclined to be concentrated in some local areas because of the operators’ habits and control system design. In this paper, a novel least squares support vector machine (LSSVM)-based ensemble learning paradigm is proposed to predict NO x emission of a coal-fired boiler using real operation data. In view of the plant data characteristics, a soft fuzzy c-means cluster algorithm is proposed to decompose the original data and guarantee the diversity of individual learners. Subsequently the base LSSVM is trained in each individual subset to solve the subtask. Finally, partial least squares (PLS) is applied as the combination strategy to eliminate the collinear and redundant information of the base learners. Considering that the fuzzy membership also has an effect on the ensemble output, the membership degree is added as one of the variables of the combiner. The single LSSVM and other ensemble models using different decomposition and combination strategies are also established to make a comparison. The result shows that the new soft FCM-LSSVM-PLS ensemble method can predict NO x emission accurately. Besides, because of the divide and conquer frame, the total time consumed in the searching the parameters and training also decreases evidently. - Highlights: • A novel LSSVM ensemble model to predict NO x emissions is presented. • LSSVM is used as the base learner and PLS is employed as the combiner. • The model is applied to process data from a 660 MW coal-fired boiler. • The generalization ability of the model is enhanced. • The time consuming in training and searching the parameters decreases sharply

  12. Winter flooding in Dutch stream valley floodplains: biogeochemical effects and vegetation consequences

    NARCIS (Netherlands)

    Beumer, V.

    2009-01-01

    Winter flooding in Dutch stream valley floodplains: biogeochemical effects and vegetation consequences Victor Beumer Climatic change has great impacts on stream catchments and their ecology. Expectations are that more extreme climate events will result in undesired flooding in stream catchments. In

  13. Ensemble Clustering using Semidefinite Programming with Applications.

    Science.gov (United States)

    Singh, Vikas; Mukherjee, Lopamudra; Peng, Jiming; Xu, Jinhui

    2010-05-01

    In this paper, we study the ensemble clustering problem, where the input is in the form of multiple clustering solutions. The goal of ensemble clustering algorithms is to aggregate the solutions into one solution that maximizes the agreement in the input ensemble. We obtain several new results for this problem. Specifically, we show that the notion of agreement under such circumstances can be better captured using a 2D string encoding rather than a voting strategy, which is common among existing approaches. Our optimization proceeds by first constructing a non-linear objective function which is then transformed into a 0-1 Semidefinite program (SDP) using novel convexification techniques. This model can be subsequently relaxed to a polynomial time solvable SDP. In addition to the theoretical contributions, our experimental results on standard machine learning and synthetic datasets show that this approach leads to improvements not only in terms of the proposed agreement measure but also the existing agreement measures based on voting strategies. In addition, we identify several new application scenarios for this problem. These include combining multiple image segmentations and generating tissue maps from multiple-channel Diffusion Tensor brain images to identify the underlying structure of the brain.

  14. 2 × 2 random matrix ensembles with reduced symmetry: from Hermitian to PT -symmetric matrices

    International Nuclear Information System (INIS)

    Gong Jiangbin; Wang Qinghai

    2012-01-01

    A possibly fruitful extension of conventional random matrix ensembles is proposed by imposing symmetry constraints on conventional Hermitian matrices or parity–time (PT)-symmetric matrices. To illustrate the main idea, we first study 2 × 2 complex Hermitian matrix ensembles with O(2)-invariant constraints, yielding novel level-spacing statistics such as singular distributions, the half-Gaussian distribution, distributions interpolating between the GOE (Gaussian orthogonal ensemble) distribution and half-Gaussian distributions, as well as the gapped-GOE distribution. Such a symmetry-reduction strategy is then used to explore 2 × 2 PT-symmetric matrix ensembles with real eigenvalues. In particular, PT-symmetric random matrix ensembles with U(2) invariance can be constructed, with the conventional complex Hermitian random matrix ensemble being a special case. In two examples of PT-symmetric random matrix ensembles, the level-spacing distributions are found to be the standard GUE (Gaussian unitary ensemble) statistics or the ‘truncated-GUE’ statistics. This article is part of a special issue of Journal of Physics A: Mathematical and Theoretical devoted to ‘Quantum physics with non-Hermitian operators’. (paper)

  15. Wetland biogeochemical processes and simulation modeling

    Science.gov (United States)

    Bai, Junhong; Huang, Laibin; Gao, Haifeng; Jia, Jia; Wang, Xin

    2018-02-01

    As the important landscape with rich biodiversity and high productivity, wetlands can provide numerous ecological services including playing an important role in regulating global biogeochemical cycles, filteringpollutants from terrestrial runoff and atmospheric deposition, protecting and improving water quality, providing living habitats for plants and animals, controlling floodwaters, and retaining surface water flow during dry periods (Reddy and DeLaune, 2008; Qin and Mitsch, 2009; Zhao et al., 2016). However, more than 50% of the world's wetlands had been altered, degraded or lost through a wide range of human activities in the past 150 years, and only a small percentage of the original wetlands remained around the world after over two centuries of intensive development and urbanization (O'connell, 2003; Zhao et al., 2016).

  16. New technologies for examining neuronal ensembles in drug addiction and fear

    Science.gov (United States)

    Cruz, Fabio C.; Koya, Eisuke; Guez-Barber, Danielle H.; Bossert, Jennifer M.; Lupica, Carl R.; Shaham, Yavin; Hope, Bruce T.

    2015-01-01

    Correlational data suggest that learned associations are encoded within neuronal ensembles. However, it has been difficult to prove that neuronal ensembles mediate learned behaviours because traditional pharmacological and lesion methods, and even newer cell type-specific methods, affect both activated and non-activated neurons. Additionally, previous studies on synaptic and molecular alterations induced by learning did not distinguish between behaviourally activated and non-activated neurons. Here, we describe three new approaches—Daun02 inactivation, FACS sorting of activated neurons and c-fos-GFP transgenic rats — that have been used to selectively target and study activated neuronal ensembles in models of conditioned drug effects and relapse. We also describe two new tools — c-fos-tTA mice and inactivation of CREB-overexpressing neurons — that have been used to study the role of neuronal ensembles in conditioned fear. PMID:24088811

  17. Evaluation of heavy metal pollution in bogs of Tomsk region on change in biogeochemical activity of ericaceous shrubs

    Science.gov (United States)

    Gaskova, L. P.

    2018-01-01

    The article discusses the change in biogeochemical activity of plant species in bogs under the influence of various types of human impact (roads, cities, drainage of mires, fire). It has been established that ericaceous shrubs, depending on the species, react with varying degrees of intensity to anthropogenic influences. The biogeochemical activity of species increased by 2.5 to 4.8 times in polluted sites.

  18. Risk assessments of regional climate change over Europe: generation of probabilistic ensemble and uncertainty assessment for EURO-CODEX

    Science.gov (United States)

    Yuan, J.; Kopp, R. E.

    2017-12-01

    Quantitative risk analysis of regional climate change is crucial for risk management and impact assessment of climate change. Two major challenges to assessing the risks of climate change are: CMIP5 model runs, which drive EURO-CODEX downscaling runs, do not cover the full range of uncertainty of future projections; Climate models may underestimate the probability of tail risks (i.e. extreme events). To overcome the difficulties, this study offers a viable avenue, where a set of probabilistic climate ensemble is generated using the Surrogate/Model Mixed Ensemble (SMME) method. The probabilistic ensembles for temperature and precipitation are used to assess the range of uncertainty covered by five bias-corrected simulations from the high-resolution (0.11º) EURO-CODEX database, which are selected by the PESETA (The Projection of Economic impacts of climate change in Sectors of the European Union based on bottom-up Analysis) III project. Results show that the distribution of SMME ensemble is notably wider than both distribution of raw ensemble of GCMs and the spread of the five EURO-CORDEX in RCP8.5. Tail risks are well presented by the SMME ensemble. Both SMME ensemble and EURO-CORDEX projections are aggregated to administrative level, and are integrated into impact functions of PESETA III to assess climate risks in Europe. To further evaluate the uncertainties introduced by the downscaling process, we compare the 5 runs from EURO-CORDEX with runs from the corresponding GCMs. Time series of regional mean, spatial patterns, and climate indices are examined for the future climate (2080-2099) deviating from the present climate (1981-2010). The downscaling processes do not appear to be trend-preserving, e.g. the increase in regional mean temperature from EURO-CORDEX is slower than that from the corresponding GCM. The spatial pattern comparison reveals that the differences between each pair of GCM and EURO-CORDEX are small in winter. In summer, the temperatures of EURO

  19. Residence-time framework for modeling multicomponent reactive transport in stream hyporheic zones

    Science.gov (United States)

    Painter, S. L.; Coon, E. T.; Brooks, S. C.

    2017-12-01

    Process-based models for transport and transformation of nutrients and contaminants in streams require tractable representations of solute exchange between the stream channel and biogeochemically active hyporheic zones. Residence-time based formulations provide an alternative to detailed three-dimensional simulations and have had good success in representing hyporheic exchange of non-reacting solutes. We extend the residence-time formulation for hyporheic transport to accommodate general multicomponent reactive transport. To that end, the integro-differential form of previous residence time models is replaced by an equivalent formulation based on a one-dimensional advection dispersion equation along the channel coupled at each channel location to a one-dimensional transport model in Lagrangian travel-time form. With the channel discretized for numerical solution, the associated Lagrangian model becomes a subgrid model representing an ensemble of streamlines that are diverted into the hyporheic zone before returning to the channel. In contrast to the previous integro-differential forms of the residence-time based models, the hyporheic flowpaths have semi-explicit spatial representation (parameterized by travel time), thus allowing coupling to general biogeochemical models. The approach has been implemented as a stream-corridor subgrid model in the open-source integrated surface/subsurface modeling software ATS. We use bedform-driven flow coupled to a biogeochemical model with explicit microbial biomass dynamics as an example to show that the subgrid representation is able to represent redox zonation in sediments and resulting effects on metal biogeochemical dynamics in a tractable manner that can be scaled to reach scales.

  20. Optical properties of single semiconductor nanowires and nanowire ensembles. Probing surface physics by photoluminescence spectroscopy

    Energy Technology Data Exchange (ETDEWEB)

    Pfueller, Carsten

    2011-06-27

    This thesis presents a detailed investigation of the optical properties of semiconductor nanowires (NWs) in general and single GaN NWs and GaN NW ensembles in particular by photoluminescence (PL) spectroscopy. NWs are often considered as potential building blocks for future nanometer-scaled devices. This vision is based on several attractive features that are generally ascribed to NWs. For instance, they are expected to grow virtually free of strain and defects even on substrates with a large structural mismatch. In the first part of the thesis, some of these expectations are examined using semiconductor NWs of different materials. On the basis of the temperature-dependent PL of Au- and selfassisted GaAs/(Al,Ga)As core-shell NWs, the influence of foreign catalyst particles on the optical properties of NWs is investigated. For the Au-assisted NWs, we find a thermally activated, nonradiative recombination channel, possibly related to Auatoms incorporated from the catalyst. These results indicate the limited suitability of catalyst-assisted NWs for optoelectronic applications. The effect of the substrate choice is studied by comparing the PL of ZnO NWs grown on Si, Al{sub 2}O{sub 3}, and ZnO substrates. Their virtually identical optical characteristics indicate that the synthesis of NWs may indeed overcome the constraints that limit the heteroepitaxial deposition of thin films. The major part of this thesis discusses the optical properties of GaN NWs grown on Si substrates. The investigation of the PL of single GaN NWs and GaN NW ensembles reveals the significance of their large surface-to-volume ratio. Differences in the recombination behavior of GaNNW ensembles and GaN layers are observed. First, the large surface-to-volume ratio is discussed to be responsible for the different recombination mechanisms apparent in NWs. Second, certain optical features are only found in the PL of GaN NWs, but not in that of GaN layers. An unexpected broadening of the donor