Directory of Open Access Journals (Sweden)
Githure John I
2009-09-01
Full Text Available Abstract Background Autoregressive regression coefficients for Anopheles arabiensis aquatic habitat models are usually assessed using global error techniques and are reported as error covariance matrices. A global statistic, however, will summarize error estimates from multiple habitat locations. This makes it difficult to identify where there are clusters of An. arabiensis aquatic habitats of acceptable prediction. It is therefore useful to conduct some form of spatial error analysis to detect clusters of An. arabiensis aquatic habitats based on uncertainty residuals from individual sampled habitats. In this research, a method of error estimation for spatial simulation models was demonstrated using autocorrelation indices and eigenfunction spatial filters to distinguish among the effects of parameter uncertainty on a stochastic simulation of ecological sampled Anopheles aquatic habitat covariates. A test for diagnostic checking error residuals in an An. arabiensis aquatic habitat model may enable intervention efforts targeting productive habitats clusters, based on larval/pupal productivity, by using the asymptotic distribution of parameter estimates from a residual autocovariance matrix. The models considered in this research extends a normal regression analysis previously considered in the literature. Methods Field and remote-sampled data were collected during July 2006 to December 2007 in Karima rice-village complex in Mwea, Kenya. SAS 9.1.4® was used to explore univariate statistics, correlations, distributions, and to generate global autocorrelation statistics from the ecological sampled datasets. A local autocorrelation index was also generated using spatial covariance parameters (i.e., Moran's Indices in a SAS/GIS® database. The Moran's statistic was decomposed into orthogonal and uncorrelated synthetic map pattern components using a Poisson model with a gamma-distributed mean (i.e. negative binomial regression. The eigenfunction
Asympotic efficiency of signed - rank symmetry tests under skew alternatives.
Alessandra Durio; Yakov Nikitin
2002-01-01
The efficiency of some known tests for symmetry such as the sign test, the Wilcoxon signed-rank test or more general linear signed rank tests was studied mainly under the classical alternatives of location. However it is interesting to compare the efficiencies of these tests under asymmetric alternatives like the so-called skew alternative proposed in Azzalini (1985). We find and compare local Bahadur efficiencies of linear signed-rank statistics for skew alternatives and discuss also the con...
Efficient, Differentially Private Point Estimators
Smith, Adam
2008-01-01
Differential privacy is a recent notion of privacy for statistical databases that provides rigorous, meaningful confidentiality guarantees, even in the presence of an attacker with access to arbitrary side information. We show that for a large class of parametric probability models, one can construct a differentially private estimator whose distribution converges to that of the maximum likelihood estimator. In particular, it is efficient and asymptotically unbiased. This result provides (furt...
Estimating the NIH efficient frontier.
Directory of Open Access Journals (Sweden)
Dimitrios Bisias
Full Text Available BACKGROUND: The National Institutes of Health (NIH is among the world's largest investors in biomedical research, with a mandate to: "…lengthen life, and reduce the burdens of illness and disability." Its funding decisions have been criticized as insufficiently focused on disease burden. We hypothesize that modern portfolio theory can create a closer link between basic research and outcome, and offer insight into basic-science related improvements in public health. We propose portfolio theory as a systematic framework for making biomedical funding allocation decisions-one that is directly tied to the risk/reward trade-off of burden-of-disease outcomes. METHODS AND FINDINGS: Using data from 1965 to 2007, we provide estimates of the NIH "efficient frontier", the set of funding allocations across 7 groups of disease-oriented NIH institutes that yield the greatest expected return on investment for a given level of risk, where return on investment is measured by subsequent impact on U.S. years of life lost (YLL. The results suggest that NIH may be actively managing its research risk, given that the volatility of its current allocation is 17% less than that of an equal-allocation portfolio with similar expected returns. The estimated efficient frontier suggests that further improvements in expected return (89% to 119% vs. current or reduction in risk (22% to 35% vs. current are available holding risk or expected return, respectively, constant, and that 28% to 89% greater decrease in average years-of-life-lost per unit risk may be achievable. However, these results also reflect the imprecision of YLL as a measure of disease burden, the noisy statistical link between basic research and YLL, and other known limitations of portfolio theory itself. CONCLUSIONS: Our analysis is intended to serve as a proof-of-concept and starting point for applying quantitative methods to allocating biomedical research funding that are objective, systematic, transparent
Estimating the NIH efficient frontier.
Bisias, Dimitrios; Lo, Andrew W; Watkins, James F
2012-01-01
The National Institutes of Health (NIH) is among the world's largest investors in biomedical research, with a mandate to: "…lengthen life, and reduce the burdens of illness and disability." Its funding decisions have been criticized as insufficiently focused on disease burden. We hypothesize that modern portfolio theory can create a closer link between basic research and outcome, and offer insight into basic-science related improvements in public health. We propose portfolio theory as a systematic framework for making biomedical funding allocation decisions-one that is directly tied to the risk/reward trade-off of burden-of-disease outcomes. Using data from 1965 to 2007, we provide estimates of the NIH "efficient frontier", the set of funding allocations across 7 groups of disease-oriented NIH institutes that yield the greatest expected return on investment for a given level of risk, where return on investment is measured by subsequent impact on U.S. years of life lost (YLL). The results suggest that NIH may be actively managing its research risk, given that the volatility of its current allocation is 17% less than that of an equal-allocation portfolio with similar expected returns. The estimated efficient frontier suggests that further improvements in expected return (89% to 119% vs. current) or reduction in risk (22% to 35% vs. current) are available holding risk or expected return, respectively, constant, and that 28% to 89% greater decrease in average years-of-life-lost per unit risk may be achievable. However, these results also reflect the imprecision of YLL as a measure of disease burden, the noisy statistical link between basic research and YLL, and other known limitations of portfolio theory itself. Our analysis is intended to serve as a proof-of-concept and starting point for applying quantitative methods to allocating biomedical research funding that are objective, systematic, transparent, repeatable, and expressly designed to reduce the burden of
Estimating the NIH Efficient Frontier
2012-01-01
Background The National Institutes of Health (NIH) is among the world’s largest investors in biomedical research, with a mandate to: “…lengthen life, and reduce the burdens of illness and disability.” Its funding decisions have been criticized as insufficiently focused on disease burden. We hypothesize that modern portfolio theory can create a closer link between basic research and outcome, and offer insight into basic-science related improvements in public health. We propose portfolio theory as a systematic framework for making biomedical funding allocation decisions–one that is directly tied to the risk/reward trade-off of burden-of-disease outcomes. Methods and Findings Using data from 1965 to 2007, we provide estimates of the NIH “efficient frontier”, the set of funding allocations across 7 groups of disease-oriented NIH institutes that yield the greatest expected return on investment for a given level of risk, where return on investment is measured by subsequent impact on U.S. years of life lost (YLL). The results suggest that NIH may be actively managing its research risk, given that the volatility of its current allocation is 17% less than that of an equal-allocation portfolio with similar expected returns. The estimated efficient frontier suggests that further improvements in expected return (89% to 119% vs. current) or reduction in risk (22% to 35% vs. current) are available holding risk or expected return, respectively, constant, and that 28% to 89% greater decrease in average years-of-life-lost per unit risk may be achievable. However, these results also reflect the imprecision of YLL as a measure of disease burden, the noisy statistical link between basic research and YLL, and other known limitations of portfolio theory itself. Conclusions Our analysis is intended to serve as a proof-of-concept and starting point for applying quantitative methods to allocating biomedical research funding that are objective, systematic, transparent
Flexible and efficient estimating equations for variogram estimation
Sun, Ying; Chang, Xiaohui; Guan, Yongtao
2018-01-01
Variogram estimation plays a vastly important role in spatial modeling. Different methods for variogram estimation can be largely classified into least squares methods and likelihood based methods. A general framework to estimate the variogram through a set of estimating equations is proposed. This approach serves as an alternative approach to likelihood based methods and includes commonly used least squares approaches as its special cases. The proposed method is highly efficient as a low dimensional representation of the weight matrix is employed. The statistical efficiency of various estimators is explored and the lag effect is examined. An application to a hydrology dataset is also presented.
Flexible and efficient estimating equations for variogram estimation
Sun, Ying
2018-01-11
Variogram estimation plays a vastly important role in spatial modeling. Different methods for variogram estimation can be largely classified into least squares methods and likelihood based methods. A general framework to estimate the variogram through a set of estimating equations is proposed. This approach serves as an alternative approach to likelihood based methods and includes commonly used least squares approaches as its special cases. The proposed method is highly efficient as a low dimensional representation of the weight matrix is employed. The statistical efficiency of various estimators is explored and the lag effect is examined. An application to a hydrology dataset is also presented.
Efficiently adapting graphical models for selectivity estimation
DEFF Research Database (Denmark)
Tzoumas, Kostas; Deshpande, Amol; Jensen, Christian S.
2013-01-01
cardinality estimation without making the independence assumption. By carefully using concepts from the field of graphical models, we are able to factor the joint probability distribution over all the attributes in the database into small, usually two-dimensional distributions, without a significant loss...... in estimation accuracy. We show how to efficiently construct such a graphical model from the database using only two-way join queries, and we show how to perform selectivity estimation in a highly efficient manner. We integrate our algorithms into the PostgreSQL DBMS. Experimental results indicate...
An efficient quantum algorithm for spectral estimation
Steffens, Adrian; Rebentrost, Patrick; Marvian, Iman; Eisert, Jens; Lloyd, Seth
2017-03-01
We develop an efficient quantum implementation of an important signal processing algorithm for line spectral estimation: the matrix pencil method, which determines the frequencies and damping factors of signals consisting of finite sums of exponentially damped sinusoids. Our algorithm provides a quantum speedup in a natural regime where the sampling rate is much higher than the number of sinusoid components. Along the way, we develop techniques that are expected to be useful for other quantum algorithms as well—consecutive phase estimations to efficiently make products of asymmetric low rank matrices classically accessible and an alternative method to efficiently exponentiate non-Hermitian matrices. Our algorithm features an efficient quantum-classical division of labor: the time-critical steps are implemented in quantum superposition, while an interjacent step, requiring much fewer parameters, can operate classically. We show that frequencies and damping factors can be obtained in time logarithmic in the number of sampling points, exponentially faster than known classical algorithms.
Efficient Frontier - Comparing Different Volatility Estimators
Tea Poklepović; Zdravka Aljinović; Mario Matković
2015-01-01
Modern Portfolio Theory (MPT) according to Markowitz states that investors form mean-variance efficient portfolios which maximizes their utility. Markowitz proposed the standard deviation as a simple measure for portfolio risk and the lower semi-variance as the only risk measure of interest to rational investors. This paper uses a third volatility estimator based on intraday data and compares three efficient frontiers on the Croatian Stock Market. The results show that ra...
An efficient estimator for Gibbs random fields
Czech Academy of Sciences Publication Activity Database
Janžura, Martin
2014-01-01
Roč. 50, č. 6 (2014), s. 883-895 ISSN 0023-5954 R&D Projects: GA ČR(CZ) GBP402/12/G097 Institutional support: RVO:67985556 Keywords : Gibbs random field * efficient estimator * empirical estimator Subject RIV: BA - General Mathematics Impact factor: 0.541, year: 2014 http://library.utia.cas.cz/separaty/2015/SI/janzura-0441325.pdf
Piecewise Loglinear Estimation of Efficient Production Surfaces
Rajiv D. Banker; Ajay Maindiratta
1986-01-01
Linear programming formulations for piecewise loglinear estimation of efficient production surfaces are derived from a set of basic properties postulated for the underlying production possibility sets. Unlike the piecewise linear model of Banker, Charnes, and Cooper (Banker R. D., A. Charnes, W. W. Cooper. 1984. Models for the estimation of technical and scale inefficiencies in data envelopment analysis. Management Sci. 30 (September) 1078--1092.), this approach permits the identification of ...
Efficient Methods of Estimating Switchgrass Biomass Supplies
Switchgrass (Panicum virgatum L.) is being developed as a biofuel feedstock for the United States. Efficient and accurate methods to estimate switchgrass biomass feedstock supply within a production area will be required by biorefineries. Our main objective was to determine the effectiveness of in...
MILITARY MISSION COMBAT EFFICIENCY ESTIMATION SYSTEM
Directory of Open Access Journals (Sweden)
Ighoyota B. AJENAGHUGHRURE
2017-04-01
Full Text Available Military infantry recruits, although trained, lacks experience in real-time combat operations, despite the combat simulations training. Therefore, the choice of including them in military operations is a thorough and careful process. This has left top military commanders with the tough task of deciding, the best blend of inexperienced and experienced infantry soldiers, for any military operation, based on available information on enemy strength and capability. This research project delves into the design of a mission combat efficiency estimator (MCEE. It is a decision support system that aids top military commanders in estimating the best combination of soldiers suitable for different military operations, based on available information on enemy’s combat experience. Hence, its advantages consist of reducing casualties and other risks that compromises the entire operation overall success, and also boosting the morals of soldiers in an operation, with such information as an estimation of combat efficiency of their enemies. The system was developed using Microsoft Asp.Net and Sql server backend. A case study test conducted with the MECEE system, reveals clearly that the MECEE system is an efficient tool for military mission planning in terms of team selection. Hence, when the MECEE system is fully deployed it will aid military commanders in the task of decision making on team members’ combination for any given operation based on enemy personnel information that is well known beforehand. Further work on the MECEE will be undertaken to explore fire power types and impact in mission combat efficiency estimation.
Efficient multidimensional regularization for Volterra series estimation
Birpoutsoukis, Georgios; Csurcsia, Péter Zoltán; Schoukens, Johan
2018-05-01
This paper presents an efficient nonparametric time domain nonlinear system identification method. It is shown how truncated Volterra series models can be efficiently estimated without the need of long, transient-free measurements. The method is a novel extension of the regularization methods that have been developed for impulse response estimates of linear time invariant systems. To avoid the excessive memory needs in case of long measurements or large number of estimated parameters, a practical gradient-based estimation method is also provided, leading to the same numerical results as the proposed Volterra estimation method. Moreover, the transient effects in the simulated output are removed by a special regularization method based on the novel ideas of transient removal for Linear Time-Varying (LTV) systems. Combining the proposed methodologies, the nonparametric Volterra models of the cascaded water tanks benchmark are presented in this paper. The results for different scenarios varying from a simple Finite Impulse Response (FIR) model to a 3rd degree Volterra series with and without transient removal are compared and studied. It is clear that the obtained models capture the system dynamics when tested on a validation dataset, and their performance is comparable with the white-box (physical) models.
Efficient estimation of diffusion during dendritic solidification
Yeum, K. S.; Poirier, D. R.; Laxmanan, V.
1989-01-01
A very efficient finite difference method has been developed to estimate the solute redistribution during solidification with diffusion in the solid. This method is validated by comparing the computed results with the results of an analytical solution derived by Kobayashi (1988) for the assumptions of a constant diffusion coefficient, a constant equilibrium partition ratio, and a parabolic rate of the advancement of the solid/liquid interface. The flexibility of the method is demonstrated by applying it to the dendritic solidification of a Pb-15 wt pct Sn alloy, for which the equilibrium partition ratio and diffusion coefficient vary substantially during solidification. The fraction eutectic at the end of solidification is also obtained by estimating the fraction solid, in greater resolution, where the concentration of solute in the interdendritic liquid reaches the eutectic composition of the alloy.
Estimation of energy efficiency of residential buildings
Directory of Open Access Journals (Sweden)
Glushkov Sergey
2017-01-01
Full Text Available Increasing energy performance of the residential buildings by means of reducing heat consumption on the heating and ventilation is the last segment in the system of energy resources saving. The first segments in the energy saving process are heat producing and transportation over the main lines and outside distribution networks. In the period from 2006 to 2013. by means of the heat-supply schemes optimization and modernization of the heating systems. using expensive (200–300 $US per 1 m though hugely effective preliminary coated pipes. the economy reached 2.7 mln tons of fuel equivalent. Considering the multi-stage and multifactorial nature (electricity. heat and water supply of the residential sector energy saving. the reasonable estimate of the efficiency of the saving of residential buildings energy should be performed in tons of fuel equivalent per unit of time.
Fast and Statistically Efficient Fundamental Frequency Estimation
DEFF Research Database (Denmark)
Nielsen, Jesper Kjær; Jensen, Tobias Lindstrøm; Jensen, Jesper Rindom
2016-01-01
Fundamental frequency estimation is a very important task in many applications involving periodic signals. For computational reasons, fast autocorrelation-based estimation methods are often used despite parametric estimation methods having superior estimation accuracy. However, these parametric...... a recursive solver. Via benchmarks, we demonstrate that the computation time is reduced by approximately two orders of magnitude. The proposed fast algorithm is available for download online....
Efficient Estimating Functions for Stochastic Differential Equations
DEFF Research Database (Denmark)
Jakobsen, Nina Munkholt
The overall topic of this thesis is approximate martingale estimating function-based estimationfor solutions of stochastic differential equations, sampled at high frequency. Focuslies on the asymptotic properties of the estimators. The first part of the thesis deals with diffusions observed over...
Efficient estimation of semiparametric copula models for bivariate survival data
Cheng, Guang
2014-01-01
A semiparametric copula model for bivariate survival data is characterized by a parametric copula model of dependence and nonparametric models of two marginal survival functions. Efficient estimation for the semiparametric copula model has been recently studied for the complete data case. When the survival data are censored, semiparametric efficient estimation has only been considered for some specific copula models such as the Gaussian copulas. In this paper, we obtain the semiparametric efficiency bound and efficient estimation for general semiparametric copula models for possibly censored data. We construct an approximate maximum likelihood estimator by approximating the log baseline hazard functions with spline functions. We show that our estimates of the copula dependence parameter and the survival functions are asymptotically normal and efficient. Simple consistent covariance estimators are also provided. Numerical results are used to illustrate the finite sample performance of the proposed estimators. © 2013 Elsevier Inc.
Efficient Estimating Functions for Stochastic Differential Equations
DEFF Research Database (Denmark)
Jakobsen, Nina Munkholt
The overall topic of this thesis is approximate martingale estimating function-based estimationfor solutions of stochastic differential equations, sampled at high frequency. Focuslies on the asymptotic properties of the estimators. The first part of the thesis deals with diffusions observed over...... a fixed time interval. Rate optimal and effcient estimators areobtained for a one-dimensional diffusion parameter. Stable convergence in distribution isused to achieve a practically applicable Gaussian limit distribution for suitably normalisedestimators. In a simulation example, the limit distributions...... multidimensional parameter. Conditions for rate optimality and effciency of estimatorsof drift-jump and diffusion parameters are given in some special cases. Theseconditions are found to extend the pre-existing conditions applicable to continuous diffusions,and impose much stronger requirements on the estimating...
Efficient channel estimation in massive MIMO systems - a distributed approach
Al-Naffouri, Tareq Y.
2016-01-01
We present two efficient algorithms for distributed estimation of channels in massive MIMO systems. The two cases of 1) generic, and 2) sparse channels is considered. The algorithms estimate the impulse response for each channel observed
Efficient bootstrap estimates for tail statistics
Breivik, Øyvind; Aarnes, Ole Johan
2017-03-01
Bootstrap resamples can be used to investigate the tail of empirical distributions as well as return value estimates from the extremal behaviour of the sample. Specifically, the confidence intervals on return value estimates or bounds on in-sample tail statistics can be obtained using bootstrap techniques. However, non-parametric bootstrapping from the entire sample is expensive. It is shown here that it suffices to bootstrap from a small subset consisting of the highest entries in the sequence to make estimates that are essentially identical to bootstraps from the entire sample. Similarly, bootstrap estimates of confidence intervals of threshold return estimates are found to be well approximated by using a subset consisting of the highest entries. This has practical consequences in fields such as meteorology, oceanography and hydrology where return values are calculated from very large gridded model integrations spanning decades at high temporal resolution or from large ensembles of independent and identically distributed model fields. In such cases the computational savings are substantial.
Estimation of gas turbine blades cooling efficiency
Moskalenko, A.B.; Kozhevnikov, A.
2016-01-01
This paper outlines the results of the evaluation of the most thermally stressed gas turbine elements, first stage power turbine blades, cooling efficiency. The calculations were implemented using a numerical simulation based on the Finite Element Method. The volume average temperature of the blade
PFP total operating efficiency calculation and basis of estimate
International Nuclear Information System (INIS)
SINCLAIR, J.C.
1999-01-01
The purpose of the Plutonium Finishing Plant (PFP) Total Operating Efficiency Calculation and Basis of Estimate document is to provide the calculated value and basis of estimate for the Total Operating Efficiency (TOE) for the material stabilization operations to be conducted in 234-52 Building. This information will be used to support both the planning and execution of the Plutonium Finishing Plant (PFP) Stabilization and Deactivation Project's (hereafter called the Project) resource-loaded, integrated schedule
Estimation of farm level technical efficiency and its determinants ...
African Journals Online (AJOL)
With the difficulties encountered by the farmers in adopting improved technologies, increasing resource use efficiency has become a very significant factor in increasing productivity. Therefore, this study was designed to estimate the farm level technical efficiency and its determinants among male and female sweet potato ...
Estimating Production Technical Efficiency of Irvingia Seed (Ogbono ...
African Journals Online (AJOL)
This study estimated the production technical efficiency of irvingia seed (Ogbono) farmers in Nsukka agricultural zone in Enugu State, Nigeria. This is against the backdrop of the importance of efficiency as a factor of productivity in a growing economy like Nigeria where resources are scarce and opportunities for new ...
Efficient estimation for ergodic diffusions sampled at high frequency
DEFF Research Database (Denmark)
Sørensen, Michael
A general theory of efficient estimation for ergodic diffusions sampled at high fre- quency is presented. High frequency sampling is now possible in many applications, in particular in finance. The theory is formulated in term of approximate martingale estimating functions and covers a large class...
Highly Efficient Estimators of Multivariate Location with High Breakdown Point
Lopuhaa, H.P.
1991-01-01
We propose an affine equivariant estimator of multivariate location that combines a high breakdown point and a bounded influence function with high asymptotic efficiency. This proposal is basically a location $M$-estimator based on the observations obtained after scaling with an affine equivariant
Extrapolated HPGe efficiency estimates based on a single calibration measurement
International Nuclear Information System (INIS)
Winn, W.G.
1994-01-01
Gamma spectroscopists often must analyze samples with geometries for which their detectors are not calibrated. The effort to experimentally recalibrate a detector for a new geometry can be quite time consuming, causing delay in reporting useful results. Such concerns have motivated development of a method for extrapolating HPGe efficiency estimates from an existing single measured efficiency. Overall, the method provides useful preliminary results for analyses that do not require exceptional accuracy, while reliably bracketing the credible range. The estimated efficiency element-of for a uniform sample in a geometry with volume V is extrapolated from the measured element-of 0 of the base sample of volume V 0 . Assuming all samples are centered atop the detector for maximum efficiency, element-of decreases monotonically as V increases about V 0 , and vice versa. Extrapolation of high and low efficiency estimates element-of h and element-of L provides an average estimate of element-of = 1/2 [element-of h + element-of L ] ± 1/2 [element-of h - element-of L ] (general) where an uncertainty D element-of = 1/2 (element-of h - element-of L ] brackets limits for a maximum possible error. The element-of h and element-of L both diverge from element-of 0 as V deviates from V 0 , causing D element-of to increase accordingly. The above concepts guided development of both conservative and refined estimates for element-of
A Computationally Efficient Method for Polyphonic Pitch Estimation
Directory of Open Access Journals (Sweden)
Ruohua Zhou
2009-01-01
Full Text Available This paper presents a computationally efficient method for polyphonic pitch estimation. The method employs the Fast Resonator Time-Frequency Image (RTFI as the basic time-frequency analysis tool. The approach is composed of two main stages. First, a preliminary pitch estimation is obtained by means of a simple peak-picking procedure in the pitch energy spectrum. Such spectrum is calculated from the original RTFI energy spectrum according to harmonic grouping principles. Then the incorrect estimations are removed according to spectral irregularity and knowledge of the harmonic structures of the music notes played on commonly used music instruments. The new approach is compared with a variety of other frame-based polyphonic pitch estimation methods, and results demonstrate the high performance and computational efficiency of the approach.
Phytoremediation: realistic estimation of modern efficiency and future possibility
International Nuclear Information System (INIS)
Kravets, A.; Pavlenko, Y.; Kusmenko, L.; Ermak, M.
1996-01-01
Kinetic peculiarities of the radionuclides migration in the system 'soil-plant' of the Chernobyl region have been investigated by means of numerical modelling. Quantitative estimation of half-time of natural cleaning of soil has been realised. Potential possibility and efficiency of the modem phytoremediation technology has been estimated. Outlines of the general demands and future possibility of biotechnology of the phytoremediation creation have been formulated. (author)
Phytoremediation: realistic estimation of modern efficiency and future possibility
Energy Technology Data Exchange (ETDEWEB)
Kravets, A; Pavlenko, Y [Institute of Cell Biology and Genetic Engineering NAS, Kiev (Ukraine); Kusmenko, L; Ermak, M [Institute of Plant Physiology and Genetic NAS, Vasilkovsky, Kiev (Ukraine)
1996-11-01
Kinetic peculiarities of the radionuclides migration in the system 'soil-plant' of the Chernobyl region have been investigated by means of numerical modelling. Quantitative estimation of half-time of natural cleaning of soil has been realised. Potential possibility and efficiency of the modem phytoremediation technologyhas been estimated. Outlines of the general demands and future possibility of biotechnology of the phytoremediation creation have been formulated. (author)
Efficient channel estimation in massive MIMO systems - a distributed approach
Al-Naffouri, Tareq Y.
2016-01-21
We present two efficient algorithms for distributed estimation of channels in massive MIMO systems. The two cases of 1) generic, and 2) sparse channels is considered. The algorithms estimate the impulse response for each channel observed by the antennas at the receiver (base station) in a coordinated manner by sharing minimal information among neighboring antennas. Simulations demonstrate the superior performance of the proposed methods as compared to other methods.
The relative efficiency of three methods of estimating herbage mass ...
African Journals Online (AJOL)
The methods involved were randomly placed circular quadrats; randomly placed narrow strips; and disc meter sampling. Disc meter and quadrat sampling appear to be more efficient than strip sampling. In a subsequent small plot grazing trial the estimates of herbage mass, using the disc meter, had a consistent precision ...
Efficient estimation for high similarities using odd sketches
DEFF Research Database (Denmark)
Mitzenmacher, Michael; Pagh, Rasmus; Pham, Ninh Dang
2014-01-01
. This means that Odd Sketches provide a highly space-efficient estimator for sets of high similarity, which is relevant in applications such as web duplicate detection, collaborative filtering, and association rule learning. The method extends to weighted Jaccard similarity, relevant e.g. for TF-IDF vector...... and web duplicate detection tasks....
Stoichiometric estimates of the biochemical conversion efficiencies in tsetse metabolism
Directory of Open Access Journals (Sweden)
Custer Adrian V
2005-08-01
Full Text Available Abstract Background The time varying flows of biomass and energy in tsetse (Glossina can be examined through the construction of a dynamic mass-energy budget specific to these flies but such a budget depends on efficiencies of metabolic conversion which are unknown. These efficiencies of conversion determine the overall yields when food or storage tissue is converted into body tissue or into metabolic energy. A biochemical approach to the estimation of these efficiencies uses stoichiometry and a simplified description of tsetse metabolism to derive estimates of the yields, for a given amount of each substrate, of conversion product, by-products, and exchanged gases. This biochemical approach improves on estimates obtained through calorimetry because the stoichiometric calculations explicitly include the inefficiencies and costs of the reactions of conversion. However, the biochemical approach still overestimates the actual conversion efficiency because the approach ignores all the biological inefficiencies and costs such as the inefficiencies of leaky membranes and the costs of molecular transport, enzyme production, and cell growth. Results This paper presents estimates of the net amounts of ATP, fat, or protein obtained by tsetse from a starting milligram of blood, and provides estimates of the net amounts of ATP formed from the catabolism of a milligram of fat along two separate pathways, one used for resting metabolism and one for flight. These estimates are derived from stoichiometric calculations constructed based on a detailed quantification of the composition of food and body tissue and on a description of the major metabolic pathways in tsetse simplified to single reaction sequences between substrates and products. The estimates include the expected amounts of uric acid formed, oxygen required, and carbon dioxide released during each conversion. The calculated estimates of uric acid egestion and of oxygen use compare favorably to
Computationally Efficient and Noise Robust DOA and Pitch Estimation
DEFF Research Database (Denmark)
Karimian-Azari, Sam; Jensen, Jesper Rindom; Christensen, Mads Græsbøll
2016-01-01
Many natural signals, such as voiced speech and some musical instruments, are approximately periodic over short intervals. These signals are often described in mathematics by the sum of sinusoids (harmonics) with frequencies that are proportional to the fundamental frequency, or pitch. In sensor...... a joint DOA and pitch estimator. In white Gaussian noise, we derive even more computationally efficient solutions which are designed using the narrowband power spectrum of the harmonics. Numerical results reveal the performance of the estimators in colored noise compared with the Cram\\'{e}r-Rao lower...
de-Graft Acquah, Henry
2014-01-01
This paper highlights the sensitivity of technical efficiency estimates to estimation approaches using empirical data. Firm specific technical efficiency and mean technical efficiency are estimated using the non parametric Data Envelope Analysis (DEA) and the parametric Corrected Ordinary Least Squares (COLS) and Stochastic Frontier Analysis (SFA) approaches. Mean technical efficiency is found to be sensitive to the choice of estimation technique. Analysis of variance and Tukeyâ€™s test sugge...
Statistically and Computationally Efficient Estimating Equations for Large Spatial Datasets
Sun, Ying
2014-11-07
For Gaussian process models, likelihood based methods are often difficult to use with large irregularly spaced spatial datasets, because exact calculations of the likelihood for n observations require O(n3) operations and O(n2) memory. Various approximation methods have been developed to address the computational difficulties. In this paper, we propose new unbiased estimating equations based on score equation approximations that are both computationally and statistically efficient. We replace the inverse covariance matrix that appears in the score equations by a sparse matrix to approximate the quadratic forms, then set the resulting quadratic forms equal to their expected values to obtain unbiased estimating equations. The sparse matrix is constructed by a sparse inverse Cholesky approach to approximate the inverse covariance matrix. The statistical efficiency of the resulting unbiased estimating equations are evaluated both in theory and by numerical studies. Our methods are applied to nearly 90,000 satellite-based measurements of water vapor levels over a region in the Southeast Pacific Ocean.
Statistically and Computationally Efficient Estimating Equations for Large Spatial Datasets
Sun, Ying; Stein, Michael L.
2014-01-01
For Gaussian process models, likelihood based methods are often difficult to use with large irregularly spaced spatial datasets, because exact calculations of the likelihood for n observations require O(n3) operations and O(n2) memory. Various approximation methods have been developed to address the computational difficulties. In this paper, we propose new unbiased estimating equations based on score equation approximations that are both computationally and statistically efficient. We replace the inverse covariance matrix that appears in the score equations by a sparse matrix to approximate the quadratic forms, then set the resulting quadratic forms equal to their expected values to obtain unbiased estimating equations. The sparse matrix is constructed by a sparse inverse Cholesky approach to approximate the inverse covariance matrix. The statistical efficiency of the resulting unbiased estimating equations are evaluated both in theory and by numerical studies. Our methods are applied to nearly 90,000 satellite-based measurements of water vapor levels over a region in the Southeast Pacific Ocean.
FASTSim: A Model to Estimate Vehicle Efficiency, Cost and Performance
Energy Technology Data Exchange (ETDEWEB)
Brooker, A.; Gonder, J.; Wang, L.; Wood, E.; Lopp, S.; Ramroth, L.
2015-05-04
The Future Automotive Systems Technology Simulator (FASTSim) is a high-level advanced vehicle powertrain systems analysis tool supported by the U.S. Department of Energy’s Vehicle Technologies Office. FASTSim provides a quick and simple approach to compare powertrains and estimate the impact of technology improvements on light- and heavy-duty vehicle efficiency, performance, cost, and battery batches of real-world drive cycles. FASTSim’s calculation framework and balance among detail, accuracy, and speed enable it to simulate thousands of driven miles in minutes. The key components and vehicle outputs have been validated by comparing the model outputs to test data for many different vehicles to provide confidence in the results. A graphical user interface makes FASTSim easy and efficient to use. FASTSim is freely available for download from the National Renewable Energy Laboratory’s website (see www.nrel.gov/fastsim).
Efficient Smoothed Concomitant Lasso Estimation for High Dimensional Regression
Ndiaye, Eugene; Fercoq, Olivier; Gramfort, Alexandre; Leclère, Vincent; Salmon, Joseph
2017-10-01
In high dimensional settings, sparse structures are crucial for efficiency, both in term of memory, computation and performance. It is customary to consider ℓ 1 penalty to enforce sparsity in such scenarios. Sparsity enforcing methods, the Lasso being a canonical example, are popular candidates to address high dimension. For efficiency, they rely on tuning a parameter trading data fitting versus sparsity. For the Lasso theory to hold this tuning parameter should be proportional to the noise level, yet the latter is often unknown in practice. A possible remedy is to jointly optimize over the regression parameter as well as over the noise level. This has been considered under several names in the literature: Scaled-Lasso, Square-root Lasso, Concomitant Lasso estimation for instance, and could be of interest for uncertainty quantification. In this work, after illustrating numerical difficulties for the Concomitant Lasso formulation, we propose a modification we coined Smoothed Concomitant Lasso, aimed at increasing numerical stability. We propose an efficient and accurate solver leading to a computational cost no more expensive than the one for the Lasso. We leverage on standard ingredients behind the success of fast Lasso solvers: a coordinate descent algorithm, combined with safe screening rules to achieve speed efficiency, by eliminating early irrelevant features.
Efficient estimation of feedback effects with application to climate models
International Nuclear Information System (INIS)
Cacugi, D.G.; Hall, M.C.G.
1984-01-01
This work presents an efficient method for calculating the sensitivity of a mathematical model's result to feedback. Feedback is defined in terms of an operator acting on the model's dependent variables. The sensitivity to feedback is defined as a functional derivative, and a method is presented to evaluate this derivative using adjoint functions. Typically, this method allows the individual effect of many different feedbacks to be estimated with a total additional computing time comparable to only one recalculation. The effects on a CO 2 -doubling experiment of actually incorporating surface albedo and water vapor feedbacks in radiative-convective model are compared with sensivities calculated using adjoint functions. These sensitivities predict the actual effects of feedback with at least the correct sign and order of magnitude. It is anticipated that this method of estimation the effect of feedback will be useful for more complex models where extensive recalculations for each of a variety of different feedbacks is impractical
Efficient AM Algorithms for Stochastic ML Estimation of DOA
Directory of Open Access Journals (Sweden)
Haihua Chen
2016-01-01
Full Text Available The estimation of direction-of-arrival (DOA of signals is a basic and important problem in sensor array signal processing. To solve this problem, many algorithms have been proposed, among which the Stochastic Maximum Likelihood (SML is one of the most concerned algorithms because of its high accuracy of DOA. However, the estimation of SML generally involves the multidimensional nonlinear optimization problem. As a result, its computational complexity is rather high. This paper addresses the issue of reducing computational complexity of SML estimation of DOA based on the Alternating Minimization (AM algorithm. We have the following two contributions. First using transformation of matrix and properties of spatial projection, we propose an efficient AM (EAM algorithm by dividing the SML criterion into two components. One depends on a single variable parameter while the other does not. Second when the array is a uniform linear array, we get the irreducible form of the EAM criterion (IAM using polynomial forms. Simulation results show that both EAM and IAM can reduce the computational complexity of SML estimation greatly, while IAM is the best. Another advantage of IAM is that this algorithm can avoid the numerical instability problem which may happen in AM and EAM algorithms when more than one parameter converges to an identical value.
Experimental study on source efficiencies for estimating surface contamination level
International Nuclear Information System (INIS)
Ichiji, Takeshi; Ogino, Haruyuki
2008-01-01
Source efficiency was measured experimentally for various materials, such as metals, nonmetals, flooring materials, sheet materials and other materials, contaminated by alpha and beta emitter radioactive nuclides. Five nuclides, 147 Pm, 60 Co, 137 Cs, 204 Tl and 90 Sr- 90 Y, were used as the beta emitters, and one nuclide 241 Am was used as the alpha emitter. The test samples were prepared by placing drops of the radioactive standardized solutions uniformly on the various materials using an automatic quantitative dispenser system from Musashi Engineering, Inc. After placing drops of the radioactive standardized solutions, the test materials were allowed to dry for more than 12 hours in a draft chamber with a hood. The radioactivity of each test material was about 30 Bq. Beta rays or alpha rays from the test materials were measured with a 2-pi gas flow proportional counter from Aloka Co., Ltd. The source efficiencies of the metals, nonmetals and sheet materials were higher than 0.5 in the case of contamination by the 137 Cs, 204 Tl and 90 Sr- 90 Y radioactive standardized solutions, higher than 0.4 in the case of contamination by the 60 Co radioactive standardized solution, and higher than 0.25 in the case of contamination by the alpha emitter the 241 Am radioactive standardized solution. These values were higher than those given in Japanese Industrial Standards (JIS) documents. In contrast, the source efficiencies of some permeable materials were lower than those given in JIS documents, because source efficiency varies depending on whether the materials or radioactive sources are wet or dry. This study provides basic data on source efficiency, which is useful for estimating the surface contamination level of materials. (author)
Commercial Discount Rate Estimation for Efficiency Standards Analysis
Energy Technology Data Exchange (ETDEWEB)
Fujita, K. Sydny [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)
2016-04-13
Underlying each of the Department of Energy's (DOE's) federal appliance and equipment standards are a set of complex analyses of the projected costs and benefits of regulation. Any new or amended standard must be designed to achieve significant additional energy conservation, provided that it is technologically feasible and economically justified (42 U.S.C. 6295(o)(2)(A)). A proposed standard is considered economically justified when its benefits exceed its burdens, as represented by the projected net present value of costs and benefits. DOE performs multiple analyses to evaluate the balance of costs and benefits of commercial appliance and equipment e efficiency standards, at the national and individual building or business level, each framed to capture different nuances of the complex impact of standards on the commercial end user population. The Life-Cycle Cost (LCC) analysis models the combined impact of appliance first cost and operating cost changes on a representative commercial building sample in order to identify the fraction of customers achieving LCC savings or incurring net cost at the considered efficiency levels.1 Thus, the choice of commercial discount rate value(s) used to calculate the present value of energy cost savings within the Life-Cycle Cost model implicitly plays a key role in estimating the economic impact of potential standard levels.2 This report is intended to provide a more in-depth discussion of the commercial discount rate estimation process than can be readily included in standard rulemaking Technical Support Documents (TSDs).
Estimation of Maize photosynthesis Efficiency Under Deficit Irrigation and Mulch
International Nuclear Information System (INIS)
Al-Hadithi, S.
2004-01-01
This research aims at estimating maize photosynthesis efficiency under deficit irrigation and soil mulching. A split-split plot design experiment was conducted with three replicates during the fall season 2000 and spring season 2001 at the experimental Station of Soil Dept./ Iraq Atomic Energy Commission. The main plots were assigned to full and deficit irrigation treatments: (C) control. The deficit irrigation treatment included the omission of one irrigation at establishment (S1, 15 days), vegetation (S2, 35 days), flowering (S3, 40 days) and yield formation (S4, 30 days) stages. The sub-plots were allocated for the two varieties, Synthetic 5012 (V1) and Haybrid 2052 (V2). The sub-sub-plots were assigned to mulch (M1) with wheat straw and no mulch (M0). Results showed that the deficit irrigation did not affect photosynthesis efficiency in both seasons, which ranged between 1.90 to 2.15% in fall season and between 1.18 and 1.45% in spring season. The hybrid variety was superior 9.39 and 9.15% over synthetic variety in fall and spring seasons, respectively. Deficit irrigation, varieties and mulch had no significant effects on harvest index in both seasons. This indicates that the two varieties were stable in their partitioning efficiency of nutrient matter between plant organ and grains under the condition of this experiment. (Author) 21 refs., 3 figs., 6 tabs
Efficient Implementation of a Symbol Timing Estimator for Broadband PLC.
Nombela, Francisco; García, Enrique; Mateos, Raúl; Hernández, Álvaro
2015-08-21
Broadband Power Line Communications (PLC) have taken advantage of the research advances in multi-carrier modulations to mitigate frequency selective fading, and their adoption opens up a myriad of applications in the field of sensory and automation systems, multimedia connectivity or smart spaces. Nonetheless, the use of these multi-carrier modulations, such as Wavelet-OFDM, requires a highly accurate symbol timing estimation for reliably recovering of transmitted data. Furthermore, the PLC channel presents some particularities that prevent the direct use of previous synchronization algorithms proposed in wireless communication systems. Therefore more research effort should be involved in the design and implementation of novel and robust synchronization algorithms for PLC, thus enabling real-time synchronization. This paper proposes a symbol timing estimator for broadband PLC based on cross-correlation with multilevel complementary sequences or Zadoff-Chu sequences and its efficient implementation in a FPGA; the obtained results show a 90% of success rate in symbol timing estimation for a certain PLC channel model and a reduced resource consumption for its implementation in a Xilinx Kyntex FPGA.
Efficient Implementation of a Symbol Timing Estimator for Broadband PLC
Directory of Open Access Journals (Sweden)
Francisco Nombela
2015-08-01
Full Text Available Broadband Power Line Communications (PLC have taken advantage of the research advances in multi-carrier modulations to mitigate frequency selective fading, and their adoption opens up a myriad of applications in the field of sensory and automation systems, multimedia connectivity or smart spaces. Nonetheless, the use of these multi-carrier modulations, such as Wavelet-OFDM, requires a highly accurate symbol timing estimation for reliably recovering of transmitted data. Furthermore, the PLC channel presents some particularities that prevent the direct use of previous synchronization algorithms proposed in wireless communication systems. Therefore more research effort should be involved in the design and implementation of novel and robust synchronization algorithms for PLC, thus enabling real-time synchronization. This paper proposes a symbol timing estimator for broadband PLC based on cross-correlation with multilevel complementary sequences or Zadoff-Chu sequences and its efficient implementation in a FPGA; the obtained results show a 90% of success rate in symbol timing estimation for a certain PLC channel model and a reduced resource consumption for its implementation in a Xilinx Kyntex FPGA.
An efficient algebraic approach to observability analysis in state estimation
Energy Technology Data Exchange (ETDEWEB)
Pruneda, R.E.; Solares, C.; Conejo, A.J. [University of Castilla-La Mancha, 13071 Ciudad Real (Spain); Castillo, E. [University of Cantabria, 39005 Santander (Spain)
2010-03-15
An efficient and compact algebraic approach to state estimation observability is proposed. It is based on transferring rows to columns and vice versa in the Jacobian measurement matrix. The proposed methodology provides a unified approach to observability checking, critical measurement identification, determination of observable islands, and selection of pseudo-measurements to restore observability. Additionally, the observability information obtained from a given set of measurements can provide directly the observability obtained from any subset of measurements of the given set. Several examples are used to illustrate the capabilities of the proposed methodology, and results from a large case study are presented to demonstrate the appropriate computational behavior of the proposed algorithms. Finally, some conclusions are drawn. (author)
Efficient Spectral Power Estimation on an Arbitrary Frequency Scale
Directory of Open Access Journals (Sweden)
F. Zaplata
2015-04-01
Full Text Available The Fast Fourier Transform is a very efficient algorithm for the Fourier spectrum estimation, but has the limitation of a linear frequency scale spectrum, which may not be suitable for every system. For example, audio and speech analysis needs a logarithmic frequency scale due to the characteristic of a human’s ear. The Fast Fourier Transform algorithms are not able to efficiently give the desired results and modified techniques have to be used in this case. In the following text a simple technique using the Goertzel algorithm allowing the evaluation of the power spectra on an arbitrary frequency scale will be introduced. Due to its simplicity the algorithm suffers from imperfections which will be discussed and partially solved in this paper. The implementation into real systems and the impact of quantization errors appeared to be critical and have to be dealt with in special cases. The simple method dealing with the quantization error will also be introduced. Finally, the proposed method will be compared to other methods based on its computational demands and its potential speed.
The estimation of energy efficiency for hybrid refrigeration system
International Nuclear Information System (INIS)
Gazda, Wiesław; Kozioł, Joachim
2013-01-01
Highlights: ► We present the experimental setup and the model of the hybrid cooling system. ► We examine impact of the operating parameters of the hybrid cooling system on the energy efficiency indicators. ► A comparison of the final and the primary energy use for a combination of the cooling systems is carried out. ► We explain the relationship between the COP and PER values for the analysed cooling systems. -- Abstract: The concept of the air blast-cryogenic freezing method (ABCF) is based on an innovative hybrid refrigeration system with one common cooling space. The hybrid cooling system consists of a vapor compression refrigeration system and a cryogenic refrigeration system. The prototype experimental setup for this method on the laboratory scale is discussed. The application of the results of experimental investigations and the theoretical–empirical model makes it possible to calculate the cooling capacity as well as the final and primary energy use in the hybrid system. The energetic analysis has been carried out for the operating modes of the refrigerating systems for the required temperatures inside the cooling chamber of −5 °C, −10 °C and −15 °C. For the estimation of the energy efficiency the coefficient of performance COP and the primary energy ratio PER for the hybrid refrigeration system are proposed. A comparison of these coefficients for the vapor compression refrigeration and the cryogenic refrigeration system has also been presented.
ESTIMATION OF EFFICIENCY PARTNERSHIP LARGE AND SMALL BUSINESS
Directory of Open Access Journals (Sweden)
Олег Васильевич Чабанюк
2014-05-01
Full Text Available In this article, based on the definition of key factors and its components, developed an algorithm consistent, logically connected stages of the transition from the traditional enterprise to enterprise innovation typebased becoming intrapreneurship. Аnalysis of economic efficiency of innovative business idea is: based on the determination of experts the importance of the model parameters ensure the effectiveness of intrapreneurship by using methods of kvalimetricheskogo modeling expert estimates score calculated "efficiency intrapreneurship". On the author's projected optimum level indicator should exceed 0.5, but it should be noted that the achievement of this level is possible with the 2 - 3rd year of existence intraprenerskoy structure. The proposed method was tested in practice and can be used for the formation of intrapreneurship in large and medium-sized enterprises as one of the methods of implementation of the innovation activities of small businesses.DOI: http://dx.doi.org/10.12731/2218-7405-2013-10-50
Asympotics with positive cosmological constant
Bonga, Beatrice; Ashtekar, Abhay; Kesavan, Aruna
2014-03-01
Since observations to date imply that our universe has a positive cosmological constant, one needs an extension of the theory of isolated systems and gravitational radiation in full general relativity from the asymptotically flat to asymptotically de Sitter space-times. In current definitions, one mimics the boundary conditions used in asymptotically AdS context to conclude that the asymptotic symmetry group is the de Sitter group. However, these conditions severely restricts radiation and in fact rules out non-zero flux of energy, momentum and angular momentum carried by gravitational waves. Therefore, these formulations of asymptotically de Sitter space-times are uninteresting beyond non-radiative spacetimes. The situation is compared and contrasted with conserved charges and fluxes at null infinity in asymptotically flat space-times.
Public-Private Investment Partnerships: Efficiency Estimation Methods
Directory of Open Access Journals (Sweden)
Aleksandr Valeryevich Trynov
2016-06-01
Full Text Available The article focuses on assessing the effectiveness of investment projects implemented on the principles of public-private partnership (PPP. This article puts forward the hypothesis that the inclusion of multiplicative economic effects will increase the attractiveness of public-private partnership projects, which in turn will contribute to the more efficient use of budgetary resources. The author proposed a methodological approach and methods of evaluating the economic efficiency of PPP projects. The author’s technique is based upon the synthesis of approaches to evaluation of the project implemented in the private and public sector and in contrast to the existing methods allows taking into account the indirect (multiplicative effect arising during the implementation of project. In the article, to estimate the multiplier effect, the model of regional economy — social accounting matrix (SAM was developed. The matrix is based on the data of the Sverdlovsk region for 2013. In the article, the genesis of the balance models of economic systems is presented. The evolution of balance models in the Russian (Soviet and foreign sources from their emergence up to now are observed. It is shown that SAM is widely used in the world for a wide range of applications, primarily to assess the impact on the regional economy of various exogenous factors. In order to clarify the estimates of multiplicative effects, the disaggregation of the account of the “industry” of the matrix of social accounts was carried out in accordance with the All-Russian Classifier of Types of Economic Activities (OKVED. This step allows to consider the particular characteristics of the industry of the estimated investment project. The method was tested on the example of evaluating the effectiveness of the construction of a toll road in the Sverdlovsk region. It is proved that due to the multiplier effect, the more capital-intensive version of the project may be more beneficial in
Efficient estimation of an additive quantile regression model
Cheng, Y.; de Gooijer, J.G.; Zerom, D.
2011-01-01
In this paper, two non-parametric estimators are proposed for estimating the components of an additive quantile regression model. The first estimator is a computationally convenient approach which can be viewed as a more viable alternative to existing kernel-based approaches. The second estimator
Statistically Efficient Methods for Pitch and DOA Estimation
DEFF Research Database (Denmark)
Jensen, Jesper Rindom; Christensen, Mads Græsbøll; Jensen, Søren Holdt
2013-01-01
, it was recently considered to estimate the DOA and pitch jointly. In this paper, we propose two novel methods for DOA and pitch estimation. They both yield maximum-likelihood estimates in white Gaussian noise scenar- ios, where the SNR may be different across channels, as opposed to state-of-the-art methods......Traditionally, direction-of-arrival (DOA) and pitch estimation of multichannel, periodic sources have been considered as two separate problems. Separate estimation may render the task of resolving sources with similar DOA or pitch impossible, and it may decrease the estimation accuracy. Therefore...
International Nuclear Information System (INIS)
Lang, Corey; Siler, Matthew
2013-01-01
Energy efficiency upgrades have been gaining widespread attention across global channels as a cost-effective approach to addressing energy challenges. The cost-effectiveness of these projects is generally predicted using engineering estimates pre-implementation, often with little ex post analysis of project success. In this paper, for a suite of energy efficiency projects, we directly compare ex ante engineering estimates of energy savings to ex post econometric estimates that use 15-min interval, building-level energy consumption data. In contrast to most prior literature, our econometric results confirm the engineering estimates, even suggesting the engineering estimates were too modest. Further, we find heterogeneous efficiency impacts by time of day, suggesting select efficiency projects can be useful in reducing peak load. - Highlights: • Regression discontinuity used to estimate energy savings from efficiency projects. • Ex post econometric estimates validate ex ante engineering estimates of energy savings. • Select efficiency projects shown to reduce peak load
Efficient estimation of an additive quantile regression model
Cheng, Y.; de Gooijer, J.G.; Zerom, D.
2009-01-01
In this paper two kernel-based nonparametric estimators are proposed for estimating the components of an additive quantile regression model. The first estimator is a computationally convenient approach which can be viewed as a viable alternative to the method of De Gooijer and Zerom (2003). By
Efficient estimation of an additive quantile regression model
Cheng, Y.; de Gooijer, J.G.; Zerom, D.
2010-01-01
In this paper two kernel-based nonparametric estimators are proposed for estimating the components of an additive quantile regression model. The first estimator is a computationally convenient approach which can be viewed as a viable alternative to the method of De Gooijer and Zerom (2003). By
On efficiency of some ratio estimators in double sampling design ...
African Journals Online (AJOL)
In this paper, three sampling ratio estimators in double sampling design were proposed with the intention of finding an alternative double sampling design estimator to the conventional ratio estimator in double sampling design discussed by Cochran (1997), Okafor (2002) , Raj (1972) and Raj and Chandhok (1999).
Global CO2 efficiency: Country-wise estimates using a stochastic cost frontier
International Nuclear Information System (INIS)
Herrala, Risto; Goel, Rajeev K.
2012-01-01
This paper examines global carbon dioxide (CO 2 ) efficiency by employing a stochastic cost frontier analysis of about 170 countries in 1997 and 2007. The main contribution lies in providing a new approach to environmental efficiency estimation, in which the efficiency estimates quantify the distance from the policy objective of minimum emissions. We are able to examine a very large pool of nations and provide country-wise efficiency estimates. We estimate three econometric models, corresponding with alternative interpretations of the Cancun vision (Conference of the Parties 2011). The models reveal progress in global environmental efficiency during a preceding decade. The estimates indicate vast differences in efficiency levels, and efficiency changes across countries. The highest efficiency levels are observed in Africa and Europe, while the lowest are clustered around China. The largest efficiency gains were observed in central and eastern Europe. CO 2 efficiency also improved in the US and China, the two largest emitters, but their ranking in terms of CO 2 efficiency deteriorated. Policy implications are discussed. - Highlights: ► We estimate global environmental efficiency in line with the Cancun vision, using a stochastic cost frontier. ► The study covers 170 countries during a 10 year period, ending in 2007. ► The biggest improvements occurred in Europe, and efficiency falls in South America. ► The efficiency ranking of US and China, the largest emitters, deteriorated. ► In 2007, highest efficiency was observed in Africa and Europe, and the lowest around China.
Efficient collaborative sparse channel estimation in massive MIMO
Masood, Mudassir; Afify, Laila H.; Al-Naffouri, Tareq Y.
2015-01-01
We propose a method for estimation of sparse frequency selective channels within MIMO-OFDM systems. These channels are independently sparse and share a common support. The method estimates the impulse response for each channel observed by the antennas at the receiver. Estimation is performed in a coordinated manner by sharing minimal information among neighboring antennas to achieve results better than many contemporary methods. Simulations demonstrate the superior performance of the proposed method.
Efficient collaborative sparse channel estimation in massive MIMO
Masood, Mudassir
2015-08-12
We propose a method for estimation of sparse frequency selective channels within MIMO-OFDM systems. These channels are independently sparse and share a common support. The method estimates the impulse response for each channel observed by the antennas at the receiver. Estimation is performed in a coordinated manner by sharing minimal information among neighboring antennas to achieve results better than many contemporary methods. Simulations demonstrate the superior performance of the proposed method.
Cheng, Guang; Zhou, Lan; Huang, Jianhua Z.
2014-01-01
We consider efficient estimation of the Euclidean parameters in a generalized partially linear additive models for longitudinal/clustered data when multiple covariates need to be modeled nonparametrically, and propose an estimation procedure based
Energy-Efficient Channel Estimation in MIMO Systems
Directory of Open Access Journals (Sweden)
2006-01-01
Full Text Available The emergence of MIMO communications systems as practical high-data-rate wireless communications systems has created several technical challenges to be met. On the one hand, there is potential for enhancing system performance in terms of capacity and diversity. On the other hand, the presence of multiple transceivers at both ends has created additional cost in terms of hardware and energy consumption. For coherent detection as well as to do optimization such as water filling and beamforming, it is essential that the MIMO channel is known. However, due to the presence of multiple transceivers at both the transmitter and receiver, the channel estimation problem is more complicated and costly compared to a SISO system. Several solutions have been proposed to minimize the computational cost, and hence the energy spent in channel estimation of MIMO systems. We present a novel method of minimizing the overall energy consumption. Unlike existing methods, we consider the energy spent during the channel estimation phase which includes transmission of training symbols, storage of those symbols at the receiver, and also channel estimation at the receiver. We develop a model that is independent of the hardware or software used for channel estimation, and use a divide-and-conquer strategy to minimize the overall energy consumption.
Efficient estimates of cochlear hearing loss parameters in individual listeners
DEFF Research Database (Denmark)
Fereczkowski, Michal; Jepsen, Morten Løve; Dau, Torsten
2013-01-01
It has been suggested that the level corresponding to the knee-point of the basilar membrane (BM) input/output (I/O) function can be used to estimate the amount of inner- and outer hair-cell loss (IHL, OHL) in listeners with a moderate cochlear hearing impairment Plack et al. (2004). According...... to Jepsen and Dau (2011) IHL + OHL = HLT [dB], where HLT stands for total hearing loss. Hence having estimates of the total hearing loss and OHC loss, one can estimate the IHL. In the present study, results from forward masking experiments based on temporal masking curves (TMC; Nelson et al., 2001...... estimates of the knee-point level. Further, it is explored whether it is possible to estimate the compression ratio using only on-frequency TMCs. 10 normal-hearing and 10 hearing-impaired listeners (with mild-to-moderate sensorineural hearing loss) were tested at 1, 2 and 4 kHz. The results showed...
Agent-based Security and Efficiency Estimation in Airport Terminals
Janssen, S.A.M.
We investigate the use of an Agent-based framework to identify and quantify the relationship between security and efficiency within airport terminals. In this framework, we define a novel Security Risk Assessment methodology that explicitly models attacker and defender behavior in a security
Thermodynamic framework for estimating the efficiencies of alkaline batteries
Energy Technology Data Exchange (ETDEWEB)
Pound, B G; Singh, R P; MacDonald, D D
1986-06-01
A thermodynamic framework has been developed to evaluate the efficiencies of alkaline battery systems for electrolyte (MOH) concentrations from 1 to 8 mol kg/sup -1/ and over the temperature range -10 to 120/sup 0/C. An analysis of the thermodynamic properties of concentrated LiOH, NaOH, and KOH solutions was carried out to provide data for the activity of water, the activity coefficient of the hydroxide ion, and the pH of the electrolyte. Potential-pH relations were then derived for various equilibrium phenomena for the metals Li, Al, Fe, Ni, and Zn in aqueous solutions and, using the data for the alkali metal hydroxides, equilibrium potentials were computed as a function of composition and temperature. These data were then used to calculate reversible cell voltages for a number of battery systems, assuming a knowledge of the cell reactions. Finally, some of the calculated cell voltages were compared with observed cell voltages to compute voltage efficiencies for various alkaline batteries. The voltage efficiencies of H/sub 2//Ni, Fe/Ni, and Zn/Ni test cells were found to be between 90 and 100%, implying that, at least at open circuit, there is little, if any, contribution from parasitic redox couples to the cell potentials for these systems. The efficiency of an Fe/air test cell was relatively low (72%). This is probably due to the less-than-theoretical voltage of the air electrode.
Sampling strategies for efficient estimation of tree foliage biomass
Hailemariam Temesgen; Vicente Monleon; Aaron Weiskittel; Duncan Wilson
2011-01-01
Conifer crowns can be highly variable both within and between trees, particularly with respect to foliage biomass and leaf area. A variety of sampling schemes have been used to estimate biomass and leaf area at the individual tree and stand scales. Rarely has the effectiveness of these sampling schemes been compared across stands or even across species. In addition,...
Cross sectional efficient estimation of stochastic volatility short rate models
Danilov, Dmitri; Mandal, Pranab K.
2001-01-01
We consider the problem of estimation of term structure of interest rates. Filtering theory approach is very natural here with the underlying setup being non-linear and non-Gaussian. Earlier works make use of Extended Kalman Filter (EKF). However, as indicated by de Jong (2000), the EKF in this
Cross sectional efficient estimation of stochastic volatility short rate models
Danilov, Dmitri; Mandal, Pranab K.
2002-01-01
We consider the problem of estimation of term structure of interest rates. Filtering theory approach is very natural here with the underlying setup being non-linear and non-Gaussian. Earlier works make use of Extended Kalman Filter (EKF). However, the EKF in this situation leads to inconsistent
SCoPE: an efficient method of Cosmological Parameter Estimation
International Nuclear Information System (INIS)
Das, Santanu; Souradeep, Tarun
2014-01-01
Markov Chain Monte Carlo (MCMC) sampler is widely used for cosmological parameter estimation from CMB and other data. However, due to the intrinsic serial nature of the MCMC sampler, convergence is often very slow. Here we present a fast and independently written Monte Carlo method for cosmological parameter estimation named as Slick Cosmological Parameter Estimator (SCoPE), that employs delayed rejection to increase the acceptance rate of a chain, and pre-fetching that helps an individual chain to run on parallel CPUs. An inter-chain covariance update is also incorporated to prevent clustering of the chains allowing faster and better mixing of the chains. We use an adaptive method for covariance calculation to calculate and update the covariance automatically as the chains progress. Our analysis shows that the acceptance probability of each step in SCoPE is more than 95% and the convergence of the chains are faster. Using SCoPE, we carry out some cosmological parameter estimations with different cosmological models using WMAP-9 and Planck results. One of the current research interests in cosmology is quantifying the nature of dark energy. We analyze the cosmological parameters from two illustrative commonly used parameterisations of dark energy models. We also asses primordial helium fraction in the universe can be constrained by the present CMB data from WMAP-9 and Planck. The results from our MCMC analysis on the one hand helps us to understand the workability of the SCoPE better, on the other hand it provides a completely independent estimation of cosmological parameters from WMAP-9 and Planck data
Optimizing Sampling Efficiency for Biomass Estimation Across NEON Domains
Abercrombie, H. H.; Meier, C. L.; Spencer, J. J.
2013-12-01
Over the course of 30 years, the National Ecological Observatory Network (NEON) will measure plant biomass and productivity across the U.S. to enable an understanding of terrestrial carbon cycle responses to ecosystem change drivers. Over the next several years, prior to operational sampling at a site, NEON will complete construction and characterization phases during which a limited amount of sampling will be done at each site to inform sampling designs, and guide standardization of data collection across all sites. Sampling biomass in 60+ sites distributed among 20 different eco-climatic domains poses major logistical and budgetary challenges. Traditional biomass sampling methods such as clip harvesting and direct measurements of Leaf Area Index (LAI) involve collecting and processing plant samples, and are time and labor intensive. Possible alternatives include using indirect sampling methods for estimating LAI such as digital hemispherical photography (DHP) or using a LI-COR 2200 Plant Canopy Analyzer. These LAI estimations can then be used as a proxy for biomass. The biomass estimates calculated can then inform the clip harvest sampling design during NEON operations, optimizing both sample size and number so that standardized uncertainty limits can be achieved with a minimum amount of sampling effort. In 2011, LAI and clip harvest data were collected from co-located sampling points at the Central Plains Experimental Range located in northern Colorado, a short grass steppe ecosystem that is the NEON Domain 10 core site. LAI was measured with a LI-COR 2200 Plant Canopy Analyzer. The layout of the sampling design included four, 300 meter transects, with clip harvests plots spaced every 50m, and LAI sub-transects spaced every 10m. LAI was measured at four points along 6m sub-transects running perpendicular to the 300m transect. Clip harvest plots were co-located 4m from corresponding LAI transects, and had dimensions of 0.1m by 2m. We conducted regression analyses
Robust efficient estimation of heart rate pulse from video
Xu, Shuchang; Sun, Lingyun; Rohde, Gustavo Kunde
2014-01-01
We describe a simple but robust algorithm for estimating the heart rate pulse from video sequences containing human skin in real time. Based on a model of light interaction with human skin, we define the change of blood concentration due to arterial pulsation as a pixel quotient in log space, and successfully use the derived signal for computing the pulse heart rate. Various experiments with different cameras, different illumination condition, and different skin locations were conducted to demonstrate the effectiveness and robustness of the proposed algorithm. Examples computed with normal illumination show the algorithm is comparable with pulse oximeter devices both in accuracy and sensitivity. PMID:24761294
Efficient Topology Estimation for Large Scale Optical Mapping
Elibol, Armagan; Garcia, Rafael
2013-01-01
Large scale optical mapping methods are in great demand among scientists who study different aspects of the seabed, and have been fostered by impressive advances in the capabilities of underwater robots in gathering optical data from the seafloor. Cost and weight constraints mean that low-cost ROVs usually have a very limited number of sensors. When a low-cost robot carries out a seafloor survey using a down-looking camera, it usually follows a predefined trajectory that provides several non time-consecutive overlapping image pairs. Finding these pairs (a process known as topology estimation) is indispensable to obtaining globally consistent mosaics and accurate trajectory estimates, which are necessary for a global view of the surveyed area, especially when optical sensors are the only data source. This book contributes to the state-of-art in large area image mosaicing methods for underwater surveys using low-cost vehicles equipped with a very limited sensor suite. The main focus has been on global alignment...
Estimation of Economic Efficiency of Regional Touristic Complex
Directory of Open Access Journals (Sweden)
Kurchenkov Vladimir Viktorovich
2015-09-01
Full Text Available The article describes the features of the development of the regional touristic complex in modern conditions and determines the direction of realizing the potential of the regional market of tourist services. The authors reveal the multiplicative interrelation for analyzing the interaction of the primary and secondary sectors of the regional market of tourist services. The key indicators of efficiency are outlined, the extent of their relevance for assessing the potential of international tourism in the region is revealed. The authors calculate the relative indicators reflecting the dynamics of incomes from inbound, outbound and domestic tourism in relation to the total income from tourism activities in the region during the reporting period, usually for one calendar year. On the basis of these parameters, the classification of the regions of the Southern Federal District in terms of tourist attraction is carried out. The authors determine the reasons of the low tourist attractiveness of the Volgograd region in comparison with other regions of the Southern Federal District. It is substantiated that the potential of expanding tourism activity is not fully realized today in the Volgograd region. The technique of analysis and evaluation of the effectiveness of regional touristic complex on the basis of cluster approach is suggested. For analyzing the effectiveness of regional tourism cluster the authors propose to use indicators that reflect the overall performance of the regional tourism cluster, characterizing the impact of cluster development of the area, or the regional market, as well as evaluating the performance of each of the companies cooperating in the framework of the cluster. The article contains recommendations to the regional authorities on improving the efficiency of regional touristic complex in the short- and long-term prospects.
Motion estimation for video coding efficient algorithms and architectures
Chakrabarti, Indrajit; Chatterjee, Sumit Kumar
2015-01-01
The need of video compression in the modern age of visual communication cannot be over-emphasized. This monograph will provide useful information to the postgraduate students and researchers who wish to work in the domain of VLSI design for video processing applications. In this book, one can find an in-depth discussion of several motion estimation algorithms and their VLSI implementation as conceived and developed by the authors. It records an account of research done involving fast three step search, successive elimination, one-bit transformation and its effective combination with diamond search and dynamic pixel truncation techniques. Two appendices provide a number of instances of proof of concept through Matlab and Verilog program segments. In this aspect, the book can be considered as first of its kind. The architectures have been developed with an eye to their applicability in everyday low-power handheld appliances including video camcorders and smartphones.
Semiparametric Gaussian copula models : Geometry and efficient rank-based estimation
Segers, J.; van den Akker, R.; Werker, B.J.M.
2014-01-01
We propose, for multivariate Gaussian copula models with unknown margins and structured correlation matrices, a rank-based, semiparametrically efficient estimator for the Euclidean copula parameter. This estimator is defined as a one-step update of a rank-based pilot estimator in the direction of
Alternative Approaches to Technical Efficiency Estimation in the Stochastic Frontier Model
Acquah, H. de-Graft; Onumah, E. E.
2014-01-01
Estimating the stochastic frontier model and calculating technical efficiency of decision making units are of great importance in applied production economic works. This paper estimates technical efficiency from the stochastic frontier model using Jondrow, and Battese and Coelli approaches. In order to compare alternative methods, simulated data with sample sizes of 60 and 200 are generated from stochastic frontier model commonly applied to agricultural firms. Simulated data is employed to co...
Operator Bias in the Estimation of Arc Efficiency in Gas Tungsten Arc Welding
Directory of Open Access Journals (Sweden)
Fredrik Sikström
2015-03-01
Full Text Available In this paper the operator bias in the measurement process of arc efficiency in stationary direct current electrode negative gas tungsten arc welding is discussed. An experimental study involving 15 operators (enough to reach statistical significance has been carried out with the purpose to estimate the arc efficiency from a specific procedure for calorimetric experiments. The measurement procedure consists of three manual operations which introduces operator bias in the measurement process. An additional relevant experiment highlights the consequences of estimating the arc voltage by measuring the potential between the terminals of the welding power source instead of measuring the potential between the electrode contact tube and the workpiece. The result of the study is a statistical evaluation of the operator bias influence on the estimate, showing that operator bias is negligible in the estimate considered here. On the contrary the consequences of neglecting welding leads voltage drop results in a significant under estimation of the arc efficiency.
Singbo, Alphonse G.; Lansink, Alfons Oude; Emvalomatis, Grigorios
2015-01-01
This paper analyzes technical efficiency and the value of the marginal product of productive inputs vis-a-vis pesticide use to measure allocative efficiency of pesticide use along productive inputs. We employ the data envelopment analysis framework and marginal cost techniques to estimate
Reinhard, S.; Lovell, C.A.K.; Thijssen, G.J.
2000-01-01
The objective of this paper is to estimate comprehensive environmental efficiency measures for Dutch dairy farms. The environmental efficiency scores are based on the nitrogen surplus, phosphate surplus and the total (direct and indirect) energy use of an unbalanced panel of dairy farms. We define
DEFF Research Database (Denmark)
Gørgens, Tue; Skeels, Christopher L.; Wurtz, Allan
This paper explores estimation of a class of non-linear dynamic panel data models with additive unobserved individual-specific effects. The models are specified by moment restrictions. The class includes the panel data AR(p) model and panel smooth transition models. We derive an efficient set...... of moment restrictions for estimation and apply the results to estimation of panel smooth transition models with fixed effects, where the transition may be determined endogenously. The performance of the GMM estimator, both in terms of estimation precision and forecasting performance, is examined in a Monte...
International Nuclear Information System (INIS)
Pollitt, Michael
2005-01-01
Electricity regulators around the world make use of efficiency analysis (or benchmarking) to produce estimates of the likely amount of cost reduction which regulated electric utilities can achieve. This short paper examines the use of such efficiency estimates by the UK electricity regulator (Ofgem) within electricity distribution and transmission price reviews. It highlights the place of efficiency analysis within the calculation of X factors. We suggest a number of problems with the current approach and make suggestions for the future development of X factor setting. (author)
An organic group contribution approach to radiative efficiency estimation of organic working fluid
International Nuclear Information System (INIS)
Zhang, Xinxin; Kobayashi, Noriyuki; He, Maogang; Wang, Jingfu
2016-01-01
Highlights: • We use group contribution method to estimate radiative efficiency. • CFC, HCFC, HFC, HFE, and PFC were estimated using this method. • In most cases, the estimation value has a good precision. • The method is reliable for the estimation of molecule with a symmetric structure. • This estimation method can offer good reference for working fluid development. - Abstract: The ratification of the Montreal Protocol in 1987 and the Kyoto Protocol in 1997 mark an environment protection era of the development of organic working fluid. Ozone depletion potential (ODP) and global warming potential (GWP) are two most important indices for the quantitative comparison of organic working fluid. Nowadays, more and more attention has been paid to GWP. The calculation of GWP is an extremely complicated process which involves interactions between surface and atmosphere such as atmospheric radiative transfer and atmospheric chemical reactions. GWP of a substance is related to its atmospheric abundance and is a variable in itself. However, radiative efficiency is an intermediate parameter for GWP calculation and it is a constant value used to describe inherent property of a substance. In this paper, the group contribution method was adopted to estimate the radiative efficiency of the organic substance which contains more than one carbon atom. In most cases, the estimation value and the standard value are in a good agreement. The biggest estimation error occurs in the estimation of the radiative efficiency of fluorinated ethers due to its plenty of structure groups and its complicated structure compared with hydrocarbon. This estimation method can be used to predict the radiative efficiency of newly developed organic working fluids.
Energy-efficient power allocation of two-hop cooperative systems with imperfect channel estimation
Amin, Osama
2015-06-08
Recently, much attention has been paid to the green design of wireless communication systems using energy efficiency (EE) metrics that should capture all energy consumption sources to deliver the required data. In this paper, we formulate an accurate EE metric for cooperative two-hop systems that use the amplify-and-forward relaying scheme. Different from the existing research that assumes the availability of perfect channel state information (CSI) at the communication cooperative nodes, we assume a practical scenario, where training pilots are used to estimate the channels. The estimated CSI can be used to adapt the available resources of the proposed system in order to maximize the EE. Two estimation strategies are assumed namely disintegrated channel estimation, which assumes the availability of channel estimator at the relay, and cascaded channel estimation, where the relay is not equipped with channel estimator and only forwards the received pilot(s) in order to let the destination estimate the cooperative link. The channel estimation cost is reflected on the EE metric by including the estimation error in the signal-to-noise term and considering the energy consumption during the estimation phase. Based on the formulated EE metric, we propose an energy-aware power allocation algorithm to maximize the EE of the cooperative system with channel estimation. Furthermore, we study the impact of the estimation parameters on the optimized EE performance via simulation examples.
RATIO ESTIMATORS FOR THE CO-EFFICIENT OF VARIATION IN A FINITE POPULATION
Directory of Open Access Journals (Sweden)
Archana V
2011-04-01
Full Text Available The Co-efficient of variation (C.V is a relative measure of dispersion and is free from unit of measurement. Hence it is widely used by the scientists in the disciplines of agriculture, biology, economics and environmental science. Although a lot of work has been reported in the past for the estimation of population C.V in infinite population models, they are not directly applicable for the finite populations. In this paper we have proposed six new estimators of the population C.V in finite population using ratio and product type estimators. The bias and mean square error of these estimators are derived for the simple random sampling design. The performance of the estimators is compared using a real life dataset. The ratio estimator using the information on the population C.V of the auxiliary variable emerges as the best estimator
Cheng, Guang
2014-02-01
We consider efficient estimation of the Euclidean parameters in a generalized partially linear additive models for longitudinal/clustered data when multiple covariates need to be modeled nonparametrically, and propose an estimation procedure based on a spline approximation of the nonparametric part of the model and the generalized estimating equations (GEE). Although the model in consideration is natural and useful in many practical applications, the literature on this model is very limited because of challenges in dealing with dependent data for nonparametric additive models. We show that the proposed estimators are consistent and asymptotically normal even if the covariance structure is misspecified. An explicit consistent estimate of the asymptotic variance is also provided. Moreover, we derive the semiparametric efficiency score and information bound under general moment conditions. By showing that our estimators achieve the semiparametric information bound, we effectively establish their efficiency in a stronger sense than what is typically considered for GEE. The derivation of our asymptotic results relies heavily on the empirical processes tools that we develop for the longitudinal/clustered data. Numerical results are used to illustrate the finite sample performance of the proposed estimators. © 2014 ISI/BS.
Takahashi, Fumitake; Kida, Akiko; Shimaoka, Takayuki
2010-10-15
Although representative removal efficiencies of gaseous mercury for air pollution control devices (APCDs) are important to prepare more reliable atmospheric emission inventories of mercury, they have been still uncertain because they depend sensitively on many factors like the type of APCDs, gas temperature, and mercury speciation. In this study, representative removal efficiencies of gaseous mercury for several types of APCDs of municipal solid waste incineration (MSWI) were offered using a statistical method. 534 data of mercury removal efficiencies for APCDs used in MSWI were collected. APCDs were categorized as fixed-bed absorber (FA), wet scrubber (WS), electrostatic precipitator (ESP), and fabric filter (FF), and their hybrid systems. Data series of all APCD types had Gaussian log-normality. The average removal efficiency with a 95% confidence interval for each APCD was estimated. The FA, WS, and FF with carbon and/or dry sorbent injection systems had 75% to 82% average removal efficiencies. On the other hand, the ESP with/without dry sorbent injection had lower removal efficiencies of up to 22%. The type of dry sorbent injection in the FF system, dry or semi-dry, did not make more than 1% difference to the removal efficiency. The injection of activated carbon and carbon-containing fly ash in the FF system made less than 3% difference. Estimation errors of removal efficiency were especially high for the ESP. The national average of removal efficiency of APCDs in Japanese MSWI plants was estimated on the basis of incineration capacity. Owing to the replacement of old APCDs for dioxin control, the national average removal efficiency increased from 34.5% in 1991 to 92.5% in 2003. This resulted in an additional reduction of about 0.86Mg emission in 2003. Further study using the methodology in this study to other important emission sources like coal-fired power plants will contribute to better emission inventories. Copyright © 2010 Elsevier B.V. All rights
2015-01-01
The recent availability of high frequency data has permitted more efficient ways of computing volatility. However, estimation of volatility from asset price observations is challenging because observed high frequency data are generally affected by noise-microstructure effects. We address this issue by using the Fourier estimator of instantaneous volatility introduced in Malliavin and Mancino 2002. We prove a central limit theorem for this estimator with optimal rate and asymptotic variance. An extensive simulation study shows the accuracy of the spot volatility estimates obtained using the Fourier estimator and its robustness even in the presence of different microstructure noise specifications. An empirical analysis on high frequency data (U.S. S&P500 and FIB 30 indices) illustrates how the Fourier spot volatility estimates can be successfully used to study intraday variations of volatility and to predict intraday Value at Risk. PMID:26421617
Energy Technology Data Exchange (ETDEWEB)
Letschert, Virginie [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Desroches, Louis-Benoit [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Ke, Jing [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); McNeil, Michael [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)
2012-07-01
As part of the ongoing effort to estimate the foreseeable impacts of aggressive minimum efficiency performance standards (MEPS) programs in the world’s major economies, Lawrence Berkeley National Laboratory (LBNL) has developed a scenario to analyze the technical potential of MEPS in 13 major economies around the world1 . The “best available technology” (BAT) scenario seeks to determine the maximum potential savings that would result from diffusion of the most efficient available technologies in these major economies.
Robust and efficient parameter estimation in dynamic models of biological systems.
Gábor, Attila; Banga, Julio R
2015-10-29
Dynamic modelling provides a systematic framework to understand function in biological systems. Parameter estimation in nonlinear dynamic models remains a very challenging inverse problem due to its nonconvexity and ill-conditioning. Associated issues like overfitting and local solutions are usually not properly addressed in the systems biology literature despite their importance. Here we present a method for robust and efficient parameter estimation which uses two main strategies to surmount the aforementioned difficulties: (i) efficient global optimization to deal with nonconvexity, and (ii) proper regularization methods to handle ill-conditioning. In the case of regularization, we present a detailed critical comparison of methods and guidelines for properly tuning them. Further, we show how regularized estimations ensure the best trade-offs between bias and variance, reducing overfitting, and allowing the incorporation of prior knowledge in a systematic way. We illustrate the performance of the presented method with seven case studies of different nature and increasing complexity, considering several scenarios of data availability, measurement noise and prior knowledge. We show how our method ensures improved estimations with faster and more stable convergence. We also show how the calibrated models are more generalizable. Finally, we give a set of simple guidelines to apply this strategy to a wide variety of calibration problems. Here we provide a parameter estimation strategy which combines efficient global optimization with a regularization scheme. This method is able to calibrate dynamic models in an efficient and robust way, effectively fighting overfitting and allowing the incorporation of prior information.
AN ESTIMATION OF TECHNICAL EFFICIENCY OF GARLIC PRODUCTION IN KHYBER PAKHTUNKHWA PAKISTAN
Directory of Open Access Journals (Sweden)
Nabeel Hussain
2014-04-01
Full Text Available This study was conducted to estimate the technical efficiency of farmers in garlic production in Khyber Pakhtunkhwa province, Pakistan. Data was randomly collected from 110 farmers using multistage sampling technique. Maximum likelihood estimation technique was used to estimate Cob-Douglas frontier production function. The analysis revealed that the estimated mean technical efficiency was 77 percent indicating that total output can be further increased with efficient use of resources and technology. The estimated gamma value was found to be 0.93 which shows 93% variation in garlic output due to inefficiency factors. The analysis further revealed that seed rate, tractor hours, fertilizer, FYM and weedicides were positive and statistically significant production factors. The results also show that age and education were statistically significant inefficiency factors, age having positive and education having negative relationship with the output of garlic. This study suggests that in order to increase the production of garlic by taking advantage of their high efficiency level, the government should invest in the research and development aspects for introducing good quality seeds to increase garlic productivity and should organize training programs to educate farmers about garlic production.
Efficient and robust estimation for longitudinal mixed models for binary data
DEFF Research Database (Denmark)
Holst, René
2009-01-01
This paper proposes a longitudinal mixed model for binary data. The model extends the classical Poisson trick, in which a binomial regression is fitted by switching to a Poisson framework. A recent estimating equations method for generalized linear longitudinal mixed models, called GEEP, is used...... as a vehicle for fitting the conditional Poisson regressions, given a latent process of serial correlated Tweedie variables. The regression parameters are estimated using a quasi-score method, whereas the dispersion and correlation parameters are estimated by use of bias-corrected Pearson-type estimating...... equations, using second moments only. Random effects are predicted by BLUPs. The method provides a computationally efficient and robust approach to the estimation of longitudinal clustered binary data and accommodates linear and non-linear models. A simulation study is used for validation and finally...
O. Severyn; O. Shulika
2017-01-01
The results of optimization of gravimetric coefficients for indexes included in the integral criterion of estimation of the efficiency of transport-technological charts of cargo delivery are resulted. The values of gravimetric coefficients are determined on the basis of two methods of experimental researches: questioning of respondents among the specialists of motor transport production and imitation design.
Shrinkage Estimators for Robust and Efficient Inference in Haplotype-Based Case-Control Studies
Chen, Yi-Hau
2009-03-01
Case-control association studies often aim to investigate the role of genes and gene-environment interactions in terms of the underlying haplotypes (i.e., the combinations of alleles at multiple genetic loci along chromosomal regions). The goal of this article is to develop robust but efficient approaches to the estimation of disease odds-ratio parameters associated with haplotypes and haplotype-environment interactions. We consider "shrinkage" estimation techniques that can adaptively relax the model assumptions of Hardy-Weinberg-Equilibrium and gene-environment independence required by recently proposed efficient "retrospective" methods. Our proposal involves first development of a novel retrospective approach to the analysis of case-control data, one that is robust to the nature of the gene-environment distribution in the underlying population. Next, it involves shrinkage of the robust retrospective estimator toward a more precise, but model-dependent, retrospective estimator using novel empirical Bayes and penalized regression techniques. Methods for variance estimation are proposed based on asymptotic theories. Simulations and two data examples illustrate both the robustness and efficiency of the proposed methods.
Shrinkage Estimators for Robust and Efficient Inference in Haplotype-Based Case-Control Studies
Chen, Yi-Hau; Chatterjee, Nilanjan; Carroll, Raymond J.
2009-01-01
Case-control association studies often aim to investigate the role of genes and gene-environment interactions in terms of the underlying haplotypes (i.e., the combinations of alleles at multiple genetic loci along chromosomal regions). The goal of this article is to develop robust but efficient approaches to the estimation of disease odds-ratio parameters associated with haplotypes and haplotype-environment interactions. We consider "shrinkage" estimation techniques that can adaptively relax the model assumptions of Hardy-Weinberg-Equilibrium and gene-environment independence required by recently proposed efficient "retrospective" methods. Our proposal involves first development of a novel retrospective approach to the analysis of case-control data, one that is robust to the nature of the gene-environment distribution in the underlying population. Next, it involves shrinkage of the robust retrospective estimator toward a more precise, but model-dependent, retrospective estimator using novel empirical Bayes and penalized regression techniques. Methods for variance estimation are proposed based on asymptotic theories. Simulations and two data examples illustrate both the robustness and efficiency of the proposed methods.
An Efficient Acoustic Density Estimation Method with Human Detectors Applied to Gibbons in Cambodia.
Directory of Open Access Journals (Sweden)
Darren Kidney
Full Text Available Some animal species are hard to see but easy to hear. Standard visual methods for estimating population density for such species are often ineffective or inefficient, but methods based on passive acoustics show more promise. We develop spatially explicit capture-recapture (SECR methods for territorial vocalising species, in which humans act as an acoustic detector array. We use SECR and estimated bearing data from a single-occasion acoustic survey of a gibbon population in northeastern Cambodia to estimate the density of calling groups. The properties of the estimator are assessed using a simulation study, in which a variety of survey designs are also investigated. We then present a new form of the SECR likelihood for multi-occasion data which accounts for the stochastic availability of animals. In the context of gibbon surveys this allows model-based estimation of the proportion of groups that produce territorial vocalisations on a given day, thereby enabling the density of groups, instead of the density of calling groups, to be estimated. We illustrate the performance of this new estimator by simulation. We show that it is possible to estimate density reliably from human acoustic detections of visually cryptic species using SECR methods. For gibbon surveys we also show that incorporating observers' estimates of bearings to detected groups substantially improves estimator performance. Using the new form of the SECR likelihood we demonstrate that estimates of availability, in addition to population density and detection function parameters, can be obtained from multi-occasion data, and that the detection function parameters are not confounded with the availability parameter. This acoustic SECR method provides a means of obtaining reliable density estimates for territorial vocalising species. It is also efficient in terms of data requirements since since it only requires routine survey data. We anticipate that the low-tech field requirements will
KDE-Track: An Efficient Dynamic Density Estimator for Data Streams
Qahtan, Abdulhakim Ali Ali; Wang, Suojin; Zhang, Xiangliang
2016-01-01
Recent developments in sensors, global positioning system devices and smart phones have increased the availability of spatiotemporal data streams. Developing models for mining such streams is challenged by the huge amount of data that cannot be stored in the memory, the high arrival speed and the dynamic changes in the data distribution. Density estimation is an important technique in stream mining for a wide variety of applications. The construction of kernel density estimators is well studied and documented. However, existing techniques are either expensive or inaccurate and unable to capture the changes in the data distribution. In this paper, we present a method called KDE-Track to estimate the density of spatiotemporal data streams. KDE-Track can efficiently estimate the density function with linear time complexity using interpolation on a kernel model, which is incrementally updated upon the arrival of new samples from the stream. We also propose an accurate and efficient method for selecting the bandwidth value for the kernel density estimator, which increases its accuracy significantly. Both theoretical analysis and experimental validation show that KDE-Track outperforms a set of baseline methods on the estimation accuracy and computing time of complex density structures in data streams.
KDE-Track: An Efficient Dynamic Density Estimator for Data Streams
Qahtan, Abdulhakim Ali Ali
2016-11-08
Recent developments in sensors, global positioning system devices and smart phones have increased the availability of spatiotemporal data streams. Developing models for mining such streams is challenged by the huge amount of data that cannot be stored in the memory, the high arrival speed and the dynamic changes in the data distribution. Density estimation is an important technique in stream mining for a wide variety of applications. The construction of kernel density estimators is well studied and documented. However, existing techniques are either expensive or inaccurate and unable to capture the changes in the data distribution. In this paper, we present a method called KDE-Track to estimate the density of spatiotemporal data streams. KDE-Track can efficiently estimate the density function with linear time complexity using interpolation on a kernel model, which is incrementally updated upon the arrival of new samples from the stream. We also propose an accurate and efficient method for selecting the bandwidth value for the kernel density estimator, which increases its accuracy significantly. Both theoretical analysis and experimental validation show that KDE-Track outperforms a set of baseline methods on the estimation accuracy and computing time of complex density structures in data streams.
A novel method for coil efficiency estimation: Validation with a 13C birdcage
DEFF Research Database (Denmark)
Giovannetti, Giulio; Frijia, Francesca; Hartwig, Valentina
2012-01-01
Coil efficiency, defined as the B1 magnetic field induced at a given point on the square root of supplied power P, is an important parameter that characterizes both the transmit and receive performance of the radiofrequency (RF) coil. Maximizing coil efficiency will maximize also the signal......-to-noise ratio. In this work, we propose a novel method for RF coil efficiency estimation based on the use of a perturbing loop. The proposed method consists of loading the coil with a known resistor by inductive coupling and measuring the quality factor with and without the load. We tested the method...... by measuring the efficiency of a 13C birdcage coil tuned at 32.13 MHz and verified its accuracy by comparing the results with the nuclear magnetic resonance nutation experiment. The method allows coil performance characterization in a short time and with great accuracy, and it can be used both on the bench...
LocExpress: a web server for efficiently estimating expression of novel transcripts.
Hou, Mei; Tian, Feng; Jiang, Shuai; Kong, Lei; Yang, Dechang; Gao, Ge
2016-12-22
The temporal and spatial-specific expression pattern of a transcript in multiple tissues and cell types can indicate key clues about its function. While several gene atlas available online as pre-computed databases for known gene models, it's still challenging to get expression profile for previously uncharacterized (i.e. novel) transcripts efficiently. Here we developed LocExpress, a web server for efficiently estimating expression of novel transcripts across multiple tissues and cell types in human (20 normal tissues/cells types and 14 cell lines) as well as in mouse (24 normal tissues/cell types and nine cell lines). As a wrapper to RNA-Seq quantification algorithm, LocExpress efficiently reduces the time cost by making abundance estimation calls increasingly within the minimum spanning bundle region of input transcripts. For a given novel gene model, such local context-oriented strategy allows LocExpress to estimate its FPKMs in hundreds of samples within minutes on a standard Linux box, making an online web server possible. To the best of our knowledge, LocExpress is the only web server to provide nearly real-time expression estimation for novel transcripts in common tissues and cell types. The server is publicly available at http://loc-express.cbi.pku.edu.cn .
Efficient estimation of dynamic density functions with an application to outlier detection
Qahtan, Abdulhakim Ali Ali; Zhang, Xiangliang; Wang, Suojin
2012-01-01
In this paper, we propose a new method to estimate the dynamic density over data streams, named KDE-Track as it is based on a conventional and widely used Kernel Density Estimation (KDE) method. KDE-Track can efficiently estimate the density with linear complexity by using interpolation on a kernel model, which is incrementally updated upon the arrival of streaming data. Both theoretical analysis and experimental validation show that KDE-Track outperforms traditional KDE and a baseline method Cluster-Kernels on estimation accuracy of the complex density structures in data streams, computing time and memory usage. KDE-Track is also demonstrated on timely catching the dynamic density of synthetic and real-world data. In addition, KDE-Track is used to accurately detect outliers in sensor data and compared with two existing methods developed for detecting outliers and cleaning sensor data. © 2012 ACM.
Ma, Yanyuan
2013-09-01
We propose semiparametric methods to estimate the center and shape of a symmetric population when a representative sample of the population is unavailable due to selection bias. We allow an arbitrary sample selection mechanism determined by the data collection procedure, and we do not impose any parametric form on the population distribution. Under this general framework, we construct a family of consistent estimators of the center that is robust to population model misspecification, and we identify the efficient member that reaches the minimum possible estimation variance. The asymptotic properties and finite sample performance of the estimation and inference procedures are illustrated through theoretical analysis and simulations. A data example is also provided to illustrate the usefulness of the methods in practice. © 2013 American Statistical Association.
International Nuclear Information System (INIS)
Chernov, N.I.; Kurbatov, V.S.; Ososkov, G.A.
1988-01-01
Parameter estimation for multivariate probability distributions is studied in experiments where data are presented as one-dimensional hystograms. For this model a statistics defined as a quadratic form of the observed frequencies which has a limitig x 2 -distribution is proposed. The efficiency of the estimator minimizing the value of that statistics is proved whithin the class of all unibased estimates obtained via minimization of quadratic forms of observed frequencies. The elaborated method was applied to the physical problem of analysis of the secondary pion energy distribution in the isobar model of pion-nucleon interactions with the production of an additional pion. The numerical experiments showed that the accuracy of estimation is twice as much if comparing the conventional methods
Madani, N.; Kimball, J. S.; Running, S. W.
2014-12-01
Remote sensing based light use efficiency (LUE) models, including the MODIS (MODerate resolution Imaging Spectroradiometer) MOD17 algorithm are commonly used for regional estimation and monitoring of vegetation gross primary production (GPP) and photosynthetic carbon (CO2) uptake. A common model assumption is that plants in a biome matrix operate at their photosynthetic capacity under optimal climatic conditions. A prescribed biome maximum light use efficiency parameter defines the maximum photosynthetic carbon conversion rate under prevailing climate conditions and is a large source of model uncertainty. Here, we used tower (FLUXNET) eddy covariance measurement based carbon flux data for estimating optimal LUE (LUEopt) over a North American domain. LUEopt was first estimated using tower observed daily carbon fluxes, meteorology and satellite (MODIS) observed fraction of photosynthetically active radiation (FPAR). LUEopt was then spatially interpolated over the domain using empirical models derived from independent geospatial data including global plant traits, surface soil moisture, terrain aspect, land cover type and percent tree cover. The derived LUEopt maps were then used as primary inputs to the MOD17 LUE algorithm for regional GPP estimation; these results were evaluated against tower observations and alternate MOD17 GPP estimates determined using Biome-specific LUEopt constants. Estimated LUEopt shows large spatial variability within and among different land cover classes indicated from a sparse North American tower network. Leaf nitrogen content and soil moisture are two important factors explaining LUEopt spatial variability. GPP estimated from spatially explicit LUEopt inputs shows significantly improved model accuracy against independent tower observations (R2 = 0.76; Mean RMSE plant trait information can explain spatial heterogeneity in LUEopt, leading to improved GPP estimates from satellite based LUE models.
Sharma, Diksha; Sempau, Josep; Badano, Aldo
2018-02-01
Monte Carlo simulations require large number of histories to obtain reliable estimates of the quantity of interest and its associated statistical uncertainty. Numerous variance reduction techniques (VRTs) have been employed to increase computational efficiency by reducing the statistical uncertainty. We investigate the effect of two VRTs for optical transport methods on accuracy and computing time for the estimation of variance (noise) in x-ray imaging detectors. We describe two VRTs. In the first, we preferentially alter the direction of the optical photons to increase detection probability. In the second, we follow only a fraction of the total optical photons generated. In both techniques, the statistical weight of photons is altered to maintain the signal mean. We use fastdetect2, an open-source, freely available optical transport routine from the hybridmantis package. We simulate VRTs for a variety of detector models and energy sources. The imaging data from the VRT simulations are then compared to the analog case (no VRT) using pulse height spectra, Swank factor, and the variance of the Swank estimate. We analyze the effect of VRTs on the statistical uncertainty associated with Swank factors. VRTs increased the relative efficiency by as much as a factor of 9. We demonstrate that we can achieve the same variance of the Swank factor with less computing time. With this approach, the simulations can be stopped when the variance of the variance estimates reaches the desired level of uncertainty. We implemented analytic estimates of the variance of Swank factor and demonstrated the effect of VRTs on image quality calculations. Our findings indicate that the Swank factor is dominated by the x-ray interaction profile as compared to the additional uncertainty introduced in the optical transport by the use of VRTs. For simulation experiments that aim at reducing the uncertainty in the Swank factor estimate, any of the proposed VRT can be used for increasing the relative
Efficient optimal joint channel estimation and data detection for massive MIMO systems
Alshamary, Haider Ali Jasim
2016-08-15
In this paper, we propose an efficient optimal joint channel estimation and data detection algorithm for massive MIMO wireless systems. Our algorithm is optimal in terms of the generalized likelihood ratio test (GLRT). For massive MIMO systems, we show that the expected complexity of our algorithm grows polynomially in the channel coherence time. Simulation results demonstrate significant performance gains of our algorithm compared with suboptimal non-coherent detection algorithms. To the best of our knowledge, this is the first algorithm which efficiently achieves GLRT-optimal non-coherent detections for massive MIMO systems with general constellations.
Program Potential: Estimates of Federal Energy Cost Savings from Energy Efficient Procurement
Energy Technology Data Exchange (ETDEWEB)
Taylor, Margaret [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Fujita, K. Sydny [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)
2012-09-17
In 2011, energy used by federal buildings cost approximately $7 billion. Reducing federal energy use could help address several important national policy goals, including: (1) increased energy security; (2) lowered emissions of greenhouse gases and other air pollutants; (3) increased return on taxpayer dollars; and (4) increased private sector innovation in energy efficient technologies. This report estimates the impact of efficient product procurement on reducing the amount of wasted energy (and, therefore, wasted money) associated with federal buildings, as well as on reducing the needless greenhouse gas emissions associated with these buildings.
Energy Technology Data Exchange (ETDEWEB)
Lee, Sung Tae [Sungkyunkwan University, Seoul (Korea); Lee, Myunghun [Keimyung University, Taegu (Korea)
2001-03-01
This paper estimates the gasoline price elasticities of demand for automobile fuel efficiency in Korea to examine indirectly whether the government policy of raising fuel prices is effective in inducing less consumption of fuel, relying on a hedonic technique developed by Atkinson and Halvorsen (1984). One of the advantages of this technique is that the data for a single year, without involving variation in the price of gasoline, is sufficient in implementing this study. Moreover, this technique enables us to circumvent the multicollinearity problem, which had reduced reliability of the results in previous hedonic studies. The estimated elasticities of demand for fuel efficiency with respect to the price of gasoline, on average, is 0.42. (author). 30 refs., 3 tabs.
Labibian, Amir; Bahrami, Amir Hossein; Haghshenas, Javad
2017-09-01
This paper presents a computationally efficient algorithm for attitude estimation of remote a sensing satellite. In this study, gyro, magnetometer, sun sensor and star tracker are used in Extended Kalman Filter (EKF) structure for the purpose of Attitude Determination (AD). However, utilizing all of the measurement data simultaneously in EKF structure increases computational burden. Specifically, assuming n observation vectors, an inverse of a 3n×3n matrix is required for gain calculation. In order to solve this problem, an efficient version of EKF, namely Murrell's version, is employed. This method utilizes measurements separately at each sampling time for gain computation. Therefore, an inverse of a 3n×3n matrix is replaced by an inverse of a 3×3 matrix for each measurement vector. Moreover, gyro drifts during the time can reduce the pointing accuracy. Therefore, a calibration algorithm is utilized for estimation of the main gyro parameters.
Al-Salihi, Hayder Qahtan Kshash; Nakhai, Mohammad Reza
2017-01-01
Efficient and highly accurate channel state information (CSI) at the base station (BS) is essential to achieve the potential benefits of massive multiple input multiple output (MIMO) systems. However, the achievable accuracy that is attainable is limited in practice due to the problem of pilot contamination. It has recently been shown that compressed sensing (CS) techniques can address the pilot contamination problem. However, CS-based channel estimation requires prior knowledge of channel sp...
International Nuclear Information System (INIS)
Hanan, N.P.; Prince, S.D.; Begue, A.
1995-01-01
The amount of photosynthetically active radiation (PAR) absorbed by green vegetation is an important determinant of photosynthesis and growth. Methods for the estimation of fractional absorption of PAR (iff PAR ) for areas greater than 1 km 2 using satellite data are discussed, and are applied to sites in the Sahel that have a sparse herb layer and tree cover of less than 5%. Using harvest measurements of seasonal net production, net production efficiencies are calculated. Variation in estimates of seasonal PAR absorption (APAR) caused by the atmospheric correction method and relationship between surface reflectances and iff PAR is considered. The use of maximum value composites of satellite NDVI to reduce the effect of the atmosphere is shown to produce inaccurate APAR estimates. In this data set, however, atmospheric correction using average optical depths was found to give good approximations of the fully corrected data. A simulation of canopy radiative transfer using the SAIL model was used to derive a relationship between canopy NDVI and iff PAR . Seasonal APAR estimates assuming a 1:1 relationship between iff PAR and NDVI overestimated the SAIL modeled results by up to 260%. The use of a modified 1:1 relationship, where iff PAR was assumed to be linearly related to NDVI scaled between minimum (soil) and maximum (infinite canopy) values, underestimated the SAIL modeled results by up to 35%. Estimated net production efficiencies (ϵ n , dry matter per unit APAR) fell in the range 0.12–1.61 g MJ −1 for above ground production, and in the range 0.16–1.88 g MJ −1 for total production. Sites with lower rainfall had reduced efficiencies, probably caused by physiological constraints on photosynthesis during dry conditions. (author)
A possible approach to estimating the operational efficiency of multiprocessor systems
International Nuclear Information System (INIS)
Kuznetsov, N.Y.; Gorlach, S.P.; Sumskaya, A.A.
1984-01-01
This article presents a mathematical model that constructs the upper and lower estimates evaluating the efficiency of solution of a large class of problems using a multiprocessor system with a specific architecture. Efficiency depends on a system's architecture (e.g., the number of processors, memory volume, the number of communication links, commutation speed) and the types of problems it is intended to solve. The behavior of the model is considered in a stationary mode. The model is used to evaluate the efficiency of a particular algorithm implemented in a multiprocessor system. It is concluded that the model is flexible and enables the investigation of a broad class of problems in computational mathematics, including linear algebra and boundary-value problems of mathematical physics
A note on the estimation of the Pareto efficient set for multiobjective matrix permutation problems.
Brusco, Michael J; Steinley, Douglas
2012-02-01
There are a number of important problems in quantitative psychology that require the identification of a permutation of the n rows and columns of an n × n proximity matrix. These problems encompass applications such as unidimensional scaling, paired-comparison ranking, and anti-Robinson forms. The importance of simultaneously incorporating multiple objective criteria in matrix permutation applications is well recognized in the literature; however, to date, there has been a reliance on weighted-sum approaches that transform the multiobjective problem into a single-objective optimization problem. Although exact solutions to these single-objective problems produce supported Pareto efficient solutions to the multiobjective problem, many interesting unsupported Pareto efficient solutions may be missed. We illustrate the limitation of the weighted-sum approach with an example from the psychological literature and devise an effective heuristic algorithm for estimating both the supported and unsupported solutions of the Pareto efficient set. © 2011 The British Psychological Society.
A virtually blind spectrum efficient channel estimation technique for mimo-ofdm system
International Nuclear Information System (INIS)
Ullah, M.O.
2015-01-01
Multiple-Input Multiple-Output antennas in conjunction with Orthogonal Frequency-Division Multiplexing is a dominant air interface for 4G and 5G cellular communication systems. Additionally, MIMO- OFDM based air interface is the foundation for latest wireless Local Area Networks, wireless Personal Area Networks, and digital multimedia broadcasting. Whether it is a single antenna or a multi-antenna OFDM system, accurate channel estimation is required for coherent reception. Training-based channel estimation methods require multiple pilot symbols and therefore waste a significant portion of channel bandwidth. This paper describes a virtually blind spectrum efficient channel estimation scheme for MIMO-OFDM systems which operates well below the Nyquist criterion. (author)
Computationally Efficient 2D DOA Estimation with Uniform Rectangular Array in Low-Grazing Angle
Directory of Open Access Journals (Sweden)
Junpeng Shi
2017-02-01
Full Text Available In this paper, we propose a computationally efficient spatial differencing matrix set (SDMS method for two-dimensional direction of arrival (2D DOA estimation with uniform rectangular arrays (URAs in a low-grazing angle (LGA condition. By rearranging the auto-correlation and cross-correlation matrices in turn among different subarrays, the SDMS method can estimate the two parameters independently with one-dimensional (1D subspace-based estimation techniques, where we only perform difference for auto-correlation matrices and the cross-correlation matrices are kept completely. Then, the pair-matching of two parameters is achieved by extracting the diagonal elements of URA. Thus, the proposed method can decrease the computational complexity, suppress the effect of additive noise and also have little information loss. Simulation results show that, in LGA, compared to other methods, the proposed methods can achieve performance improvement in the white or colored noise conditions.
Efficiency Optimization Control of IPM Synchronous Motor Drives with Online Parameter Estimation
Directory of Open Access Journals (Sweden)
Sadegh Vaez-Zadeh
2011-04-01
Full Text Available This paper describes an efficiency optimization control method for high performance interior permanent magnet synchronous motor drives with online estimation of motor parameters. The control system is based on an input-output feedback linearization method which provides high performance control and simultaneously ensures the minimization of the motor losses. The controllable electrical loss can be minimized by the optimal control of the armature current vector. It is shown that parameter variations except at near the nominal conditions have undesirable effect on the controller performance. Therefore, a parameter estimation method based on the second method of Lyapunov is presented which guarantees the stability and convergence of the estimation. The extensive simulation results show the feasibility of the proposed controller and observer and their desirable performances.
The efficiency of different estimation methods of hydro-physical limits
Directory of Open Access Journals (Sweden)
Emma María Martínez
2012-12-01
Full Text Available The soil water available to crops is defined by specific values of water potential limits. Underlying the estimation of hydro-physical limits, identified as permanent wilting point (PWP and field capacity (FC, is the selection of a suitable method based on a multi-criteria analysis that is not always clear and defined. In this kind of analysis, the time required for measurements must be taken into consideration as well as other external measurement factors, e.g., the reliability and suitability of the study area, measurement uncertainty, cost, effort and labour invested. In this paper, the efficiency of different methods for determining hydro-physical limits is evaluated by using indices that allow for the calculation of efficiency in terms of effort and cost. The analysis evaluates both direct determination methods (pressure plate - PP and water activity meter - WAM and indirect estimation methods (pedotransfer functions - PTFs. The PTFs must be validated for the area of interest before use, but the time and cost associated with this validation are not included in the cost of analysis. Compared to the other methods, the combined use of PP and WAM to determine hydro-physical limits differs significantly in time and cost required and quality of information. For direct methods, increasing sample size significantly reduces cost and time. This paper assesses the effectiveness of combining a general analysis based on efficiency indices and more specific analyses based on the different influencing factors, which were considered separately so as not to mask potential benefits or drawbacks that are not evidenced in efficiency estimation.
Estimating returns to scale and scale efficiency for energy consuming appliances
Energy Technology Data Exchange (ETDEWEB)
Blum, Helcio [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States). Energy Efficiency Standards Group; Okwelum, Edson O. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States). Energy Efficiency Standards Group
2018-01-18
Energy consuming appliances accounted for over 40% of the energy use and $17 billion in sales in the U.S. in 2014. Whether such amounts of money and energy were optimally combined to produce household energy services is not straightforwardly determined. The efficient allocation of capital and energy to provide an energy service has been previously approached, and solved with Data Envelopment Analysis (DEA) under constant returns to scale. That approach, however, lacks the scale dimension of the problem and may restrict the economic efficient models of an appliance available in the market when constant returns to scale does not hold. We expand on that approach to estimate returns to scale for energy using appliances. We further calculate DEA scale efficiency scores for the technically efficient models that comprise the economic efficient frontier of the energy service delivered, under different assumptions of returns to scale. We then apply this approach to evaluate dishwashers available in the market in the U.S. Our results show that (a) for the case of dishwashers scale matters, and (b) the dishwashing energy service is delivered under non-decreasing returns to scale. The results further demonstrate that this method contributes to increase consumers’ choice of appliances.
Kolotii, Andrii; Kussul, Nataliia; Skakun, Sergii; Shelestov, Andrii; Ostapenko, Vadim; Oliinyk, Tamara
2015-04-01
Efficient and timely crop monitoring and yield forecasting are important tasks for ensuring of stability and sustainable economic development [1]. As winter crops pay prominent role in agriculture of Ukraine - the main focus of this study is concentrated on winter wheat. In our previous research [2, 3] it was shown that usage of biophysical parameters of crops such as FAPAR (derived from Geoland-2 portal as for SPOT Vegetation data) is far more efficient for crop yield forecasting to NDVI derived from MODIS data - for available data. In our current work efficiency of usage such biophysical parameters as LAI, FAPAR, FCOVER (derived from SPOT Vegetation and PROBA-V data at resolution of 1 km and simulated within WOFOST model) and NDVI product (derived from MODIS) for winter wheat monitoring and yield forecasting is estimated. As the part of crop monitoring workflow (vegetation anomaly detection, vegetation indexes and products analysis) and yield forecasting SPIRITS tool developed by JRC is used. Statistics extraction is done for landcover maps created in SRI within FP-7 SIGMA project. Efficiency of usage satellite based and modelled with WOFOST model biophysical products is estimated. [1] N. Kussul, S. Skakun, A. Shelestov, O. Kussul, "Sensor Web approach to Flood Monitoring and Risk Assessment", in: IGARSS 2013, 21-26 July 2013, Melbourne, Australia, pp. 815-818. [2] F. Kogan, N. Kussul, T. Adamenko, S. Skakun, O. Kravchenko, O. Kryvobok, A. Shelestov, A. Kolotii, O. Kussul, and A. Lavrenyuk, "Winter wheat yield forecasting in Ukraine based on Earth observation, meteorological data and biophysical models," International Journal of Applied Earth Observation and Geoinformation, vol. 23, pp. 192-203, 2013. [3] Kussul O., Kussul N., Skakun S., Kravchenko O., Shelestov A., Kolotii A, "Assessment of relative efficiency of using MODIS data to winter wheat yield forecasting in Ukraine", in: IGARSS 2013, 21-26 July 2013, Melbourne, Australia, pp. 3235 - 3238.
Madani, Nima; Kimball, John S.; Running, Steven W.
2017-11-01
In the light use efficiency (LUE) approach of estimating the gross primary productivity (GPP), plant productivity is linearly related to absorbed photosynthetically active radiation assuming that plants absorb and convert solar energy into biomass within a maximum LUE (LUEmax) rate, which is assumed to vary conservatively within a given biome type. However, it has been shown that photosynthetic efficiency can vary within biomes. In this study, we used 149 global CO2 flux towers to derive the optimum LUE (LUEopt) under prevailing climate conditions for each tower location, stratified according to model training and test sites. Unlike LUEmax, LUEopt varies according to heterogeneous landscape characteristics and species traits. The LUEopt data showed large spatial variability within and between biome types, so that a simple biome classification explained only 29% of LUEopt variability over 95 global tower training sites. The use of explanatory variables in a mixed effect regression model explained 62.2% of the spatial variability in tower LUEopt data. The resulting regression model was used for global extrapolation of the LUEopt data and GPP estimation. The GPP estimated using the new LUEopt map showed significant improvement relative to global tower data, including a 15% R2 increase and 34% root-mean-square error reduction relative to baseline GPP calculations derived from biome-specific LUEmax constants. The new global LUEopt map is expected to improve the performance of LUE-based GPP algorithms for better assessment and monitoring of global terrestrial productivity and carbon dynamics.
Estimation of combustion flue gas acid dew point during heat recovery and efficiency gain
Energy Technology Data Exchange (ETDEWEB)
Bahadori, A. [Curtin University of Technology, Perth, WA (Australia)
2011-06-15
When cooling combustion flue gas for heat recovery and efficiency gain, the temperature must not be allowed to drop below the sulfur trioxide dew point. Below the SO{sub 3} dew point, very corrosive sulfuric acid forms and leads to operational hazards on metal surfaces. In the present work, simple-to-use predictive tool, which is easier than existing approaches, less complicated with fewer computations is formulated to arrive at an appropriate estimation of acid dew point during combustion flue gas cooling which depends on fuel type, sulfur content in fuel, and excess air levels. The resulting information can then be applied to estimate the acid dew point, for sulfur in various fuels up to 0.10 volume fraction in gas (0.10 mass fraction in liquid), excess air fractions up to 0.25, and elemental concentrations of carbon up to 3. The proposed predictive tool shows a very good agreement with the reported data wherein the average absolute deviation percent was found to be around 3.18%. This approach can be of immense practical value for engineers and scientists for a quick estimation of acid dew point during combustion flue gas cooling for heat recovery and efficiency gain for wide range of operating conditions without the necessity of any pilot plant setup and tedious experimental trials. In particular, process and combustion engineers would find the tool to be user friendly involving transparent calculations with no complex expressions for their applications.
Energy Technology Data Exchange (ETDEWEB)
Hernandez-Bermejo, B. [Departamento de Fisica, Universidad Rey Juan Carlos, Escuela Superior de Ciencias Experimentales y Tecnologia, Edificio Departamental II, Calle Tulipan S/N, 28933-Mostoles-Madrid (Spain)], E-mail: benito.hernandez@urjc.es; Marco-Blanco, J. [Departamento de Fisica, Universidad Rey Juan Carlos, Escuela Superior de Ciencias Experimentales y Tecnologia, Edificio Departamental II, Calle Tulipan S/N, 28933-Mostoles-Madrid (Spain); Romance, M. [Departamento de Matematica Aplicada, Universidad Rey Juan Carlos, Escuela Superior de Ciencias Experimentales y Tecnologia, Edificio Departamental II, Calle Tulipan S/N, 28933-Mostoles-Madrid (Spain)
2009-02-23
Estimates for the efficiency of a tree are derived, leading to new analytical expressions for Barabasi-Albert trees efficiency. These expressions are used to investigate the dynamic behaviour of such networks. It is proved that the preferential attachment leads to an asymptotic conservation of efficiency as the Barabasi-Albert trees grow.
International Nuclear Information System (INIS)
Hernandez-Bermejo, B.; Marco-Blanco, J.; Romance, M.
2009-01-01
Estimates for the efficiency of a tree are derived, leading to new analytical expressions for Barabasi-Albert trees efficiency. These expressions are used to investigate the dynamic behaviour of such networks. It is proved that the preferential attachment leads to an asymptotic conservation of efficiency as the Barabasi-Albert trees grow
International Nuclear Information System (INIS)
Utlu, Zafer; Hepbasli, Arif
2006-01-01
The main objectives in carrying out the present study are twofold, namely to estimate the energy and exergy utilization efficiencies for the residential-commercial sector and to compare those of various countries with each other. In this regard, Turkey is given as an illustrative example with its latest figures in 2002 since the data related to the following years are still being processed. Total energy and exergy inputs in this year are calculated to be 3257.20 and 3212.42 PJ, respectively. Annual fuel consumptions in space heating, water heating and cooking activities as well as electrical energy uses by appliances are also determined. The energy and exergy utilization efficiency values for the Turkish residential-commercial sector are obtained to be 55.58% and 9.33%, respectively. Besides this, Turkey's overall energy and exergy utilization efficiencies are found to be 46.02% and 24.99%, respectively. The present study clearly indicates the necessity of the planned studies toward increasing exergy utilization efficiencies in the sector studied
Efficient estimators for likelihood ratio sensitivity indices of complex stochastic dynamics
Energy Technology Data Exchange (ETDEWEB)
Arampatzis, Georgios; Katsoulakis, Markos A.; Rey-Bellet, Luc [Department of Mathematics and Statistics, University of Massachusetts, Amherst, Massachusetts 01003 (United States)
2016-03-14
We demonstrate that centered likelihood ratio estimators for the sensitivity indices of complex stochastic dynamics are highly efficient with low, constant in time variance and consequently they are suitable for sensitivity analysis in long-time and steady-state regimes. These estimators rely on a new covariance formulation of the likelihood ratio that includes as a submatrix a Fisher information matrix for stochastic dynamics and can also be used for fast screening of insensitive parameters and parameter combinations. The proposed methods are applicable to broad classes of stochastic dynamics such as chemical reaction networks, Langevin-type equations and stochastic models in finance, including systems with a high dimensional parameter space and/or disparate decorrelation times between different observables. Furthermore, they are simple to implement as a standard observable in any existing simulation algorithm without additional modifications.
ESTIMATION OF EFFICIENCY OF OPERATING SYSTEM OF TAX PLANNING IN THE COMMERCIAL ORGANIZATIONS
Directory of Open Access Journals (Sweden)
Evgeniy A. Samsonov
2014-01-01
Full Text Available Present clause is devoted to scientific judgement and estimations of efficiency of stimulating mechanisms (tools of application of system of tax planning in the commercial organizations which allow to estimate разнонаправленное influence of taxes on final financial result of the organization, and also to predict change of business activity of the organization depending on tax loading. The big attention is given to complicated questions of features of management by the taxation and the order of reflection in the tax account of the facts of the economic activities arising between the state, on the one hand, and managing subjects - the commercial organizations - with another.
DEFF Research Database (Denmark)
Gardi, Jonathan Eyal; Nyengaard, Jens Randel; Gundersen, Hans Jørgen Gottlieb
2008-01-01
Quantification of tissue properties is improved using the general proportionator sampling and estimation procedure: automatic image analysis and non-uniform sampling with probability proportional to size (PPS). The complete region of interest is partitioned into fields of view, and every field...... of view is given a weight (the size) proportional to the total amount of requested image analysis features in it. The fields of view sampled with known probabilities proportional to individual weight are the only ones seen by the observer who provides the correct count. Even though the image analysis...... cerebellum, total number of orexin positive neurons in transgenic mice brain, and estimating the absolute area and the areal fraction of β islet cells in dog pancreas. The proportionator was at least eight times more efficient (precision and time combined) than traditional computer controlled sampling....
DEFF Research Database (Denmark)
Fereczkowski, Michal; Jepsen, Morten Løve; Dau, Torsten
2017-01-01
-output (I/O) function have been proposed. However, such measures are very time consuming. The present study investigated possible modifications of the temporal masking curve (TMC) paradigm to improve time and measurement efficiency. In experiment 1, estimates of knee point (KP) and compression ratio (CR......”, was tested. In contrast to the standard TMC paradigm, the maker level was kept fixed and the “gap threshold” was obtained, such that the masker just masks a low-level (12 dB sensation level) signal. It is argued that this modification allows for better control of the tested stimulus level range, which...
SEBAL Model Using to Estimate Irrigation Water Efficiency & Water Requirement of Alfalfa Crop
Zeyliger, Anatoly; Ermolaeva, Olga
2013-04-01
The sustainability of irrigation is a complex and comprehensive undertaking, requiring an attention to much more than hydraulics, chemistry, and agronomy. A special combination of human, environmental, and economic factors exists in each irrigated region and must be recognized and evaluated. A way to evaluate the efficiency of irrigation water use for crop production is to consider the so-called crop-water production functions, which express the relation between the yield of a crop and the quantity of water applied to it or consumed by it. The term has been used in a somewhat ambiguous way. Some authors have defined the Crop-Water Production Functions between yield and the total amount of water applied, whereas others have defined it as a relation between yield and seasonal evapotranspiration (ET). In case of high efficiency of irrigation water use the volume of water applied is less than the potential evapotranspiration (PET), then - assuming no significant change of soil moisture storage from beginning of the growing season to its end-the volume of water may be roughly equal to ET. In other case of low efficiency of irrigation water use the volume of water applied exceeds PET, then the excess of volume of water applied over PET must go to either augmenting soil moisture storage (end-of-season moisture being greater than start-of-season soil moisture) or to runoff or/and deep percolation beyond the root zone. In presented contribution some results of a case study of estimation of biomass and leaf area index (LAI) for irrigated alfalfa by SEBAL algorithm will be discussed. The field study was conducted with aim to compare ground biomass of alfalfa at some irrigated fields (provided by agricultural farm) at Saratov and Volgograd Regions of Russia. The study was conducted during vegetation period of 2012 from April till September. All the operations from importing the data to calculation of the output data were carried by eLEAF company and uploaded in Fieldlook web
Energy efficiency estimation of a steam powered LNG tanker using normal operating data
Directory of Open Access Journals (Sweden)
Sinha Rajendra Prasad
2016-01-01
Full Text Available A ship’s energy efficiency performance is generally estimated by conducting special sea trials of few hours under very controlled environmental conditions of calm sea, standard draft and optimum trim. This indicator is then used as the benchmark for future reference of the ship’s Energy Efficiency Performance (EEP. In practice, however, for greater part of operating life the ship operates in conditions which are far removed from original sea trial conditions and therefore comparing energy performance with benchmark performance indicator is not truly valid. In such situations a higher fuel consumption reading from the ship fuel meter may not be a true indicator of poor machinery performance or dirty underwater hull. Most likely, the reasons for higher fuel consumption may lie in factors other than the condition of hull and machinery, such as head wind, current, low load operations or incorrect trim [1]. Thus a better and more accurate approach to determine energy efficiency of the ship attributable only to main machinery and underwater hull condition will be to filter out the influence of all spurious and non-standard operating conditions from the ship’s fuel consumption [2]. The author in this paper identifies parameters of a suitable filter to be used on the daily report data of a typical LNG tanker of 33000 kW shaft power to remove effects of spurious and non-standard ship operations on its fuel consumption. The filtered daily report data has been then used to estimate actual fuel efficiency of the ship and compared with the sea trials benchmark performance. Results obtained using data filter show closer agreement with the benchmark EEP than obtained from the monthly mini trials . The data filtering method proposed in this paper has the advantage of using the actual operational data of the ship and thus saving cost of conducting special sea trials to estimate ship EEP. The agreement between estimated results and special sea trials EEP is
Efficient 3D movement-based kernel density estimator and application to wildlife ecology
Tracey-PR, Jeff; Sheppard, James K.; Lockwood, Glenn K.; Chourasia, Amit; Tatineni, Mahidhar; Fisher, Robert N.; Sinkovits, Robert S.
2014-01-01
We describe an efficient implementation of a 3D movement-based kernel density estimator for determining animal space use from discrete GPS measurements. This new method provides more accurate results, particularly for species that make large excursions in the vertical dimension. The downside of this approach is that it is much more computationally expensive than simpler, lower-dimensional models. Through a combination of code restructuring, parallelization and performance optimization, we were able to reduce the time to solution by up to a factor of 1000x, thereby greatly improving the applicability of the method.
International Nuclear Information System (INIS)
Baskette, C.; Horii, B.; Price, S.; Kollman, E.
2006-01-01
This paper summarizes the first comprehensive estimation of California's electricity avoided costs since the state reformed its electricity market. It describes avoided cost estimates that vary by time and location, thus facilitating targeted design, funding, and marketing of demand-side management (DSM) and energy efficiency (EE) programs that could not have occurred under the previous methodology of system average cost estimation. The approach, data, and results reflect two important market structure changes: (a) wholesale spot and forward markets now supply electricity commodities to load serving entities; and (b) the evolution of an emissions market that internalizes and prices some of the externalities of electricity generation. The paper also introduces the multiplier effect of a price reduction due to DSM/EE implementation on electricity bills of all consumers. It affirms that area- and time-specific avoided cost estimates can improve the allocation of the state's public funding for DSM/EE programs, a finding that could benefit other parts of North America (e.g. Ontario and New York), which have undergone electricity deregulation. (author)
Computationally Efficient 2D DOA Estimation for L-Shaped Array with Unknown Mutual Coupling
Directory of Open Access Journals (Sweden)
Yang-Yang Dong
2018-01-01
Full Text Available Although L-shaped array can provide good angle estimation performance and is easy to implement, its two-dimensional (2D direction-of-arrival (DOA performance degrades greatly in the presence of mutual coupling. To deal with the mutual coupling effect, a novel 2D DOA estimation method for L-shaped array with low computational complexity is developed in this paper. First, we generalize the conventional mutual coupling model for L-shaped array and compensate the mutual coupling blindly via sacrificing a few sensors as auxiliary elements. Then we apply the propagator method twice to mitigate the effect of strong source signal correlation effect. Finally, the estimations of azimuth and elevation angles are achieved simultaneously without pair matching via the complex eigenvalue technique. Compared with the existing methods, the proposed method is computationally efficient without spectrum search or polynomial rooting and also has fine angle estimation performance for highly correlated source signals. Theoretical analysis and simulation results have demonstrated the effectiveness of the proposed method.
Efficient Monte Carlo Estimation of the Expected Value of Sample Information Using Moment Matching.
Heath, Anna; Manolopoulou, Ioanna; Baio, Gianluca
2018-02-01
The Expected Value of Sample Information (EVSI) is used to calculate the economic value of a new research strategy. Although this value would be important to both researchers and funders, there are very few practical applications of the EVSI. This is due to computational difficulties associated with calculating the EVSI in practical health economic models using nested simulations. We present an approximation method for the EVSI that is framed in a Bayesian setting and is based on estimating the distribution of the posterior mean of the incremental net benefit across all possible future samples, known as the distribution of the preposterior mean. Specifically, this distribution is estimated using moment matching coupled with simulations that are available for probabilistic sensitivity analysis, which is typically mandatory in health economic evaluations. This novel approximation method is applied to a health economic model that has previously been used to assess the performance of other EVSI estimators and accurately estimates the EVSI. The computational time for this method is competitive with other methods. We have developed a new calculation method for the EVSI which is computationally efficient and accurate. This novel method relies on some additional simulation so can be expensive in models with a large computational cost.
The Use of 32P and 15N to Estimate Fertilizer Efficiency in Oil Palm
International Nuclear Information System (INIS)
Sisworo, Elsje L; Sisworo, Widjang H; Havid-Rasjid; Haryanto; Syamsul-Rizal; Poeloengan, Z; Kusnu-Martoyo
2004-01-01
Oil palm has become an important commodity for Indonesia reaching an area of 2.6 million ha at the end of 1998. It is mostly cultivated in highly weathered acid soil usually Ultisols and Oxisols which are known for their low fertility, concerning the major nutrients like N and P. This study most conducted to search for the most active root-zone of oil palm and applied urea fertilizer at such soils to obtain high N-efficiency. Carrier free KH 2 32 PO 4 solution was used to determine the active root-zone of oil palm by applying 32 P around the plant in twenty holes. After the most active root-zone have been determined, urea in one, two and three splits were respectively applied at this zone. To estimate N-fertilizer efficiency of urea labelled 15 N Ammonium Sulphate was used by adding them at the same amount of 16 g 15 N plan -1 . This study showed that the most active root-zone was found at a 1.5 m distance from the plant-stem and at 5 cm soil depth. For urea the highest N-efficiency was obtained from applying it at two splits. The use of 32 P was able to distinguish several root zones: 1.5 m - 2.5 m from the plant-stem at a 5 cm and 15 cm soil depth. Urea placed at the most active root-zone, which was at a 1.5 m distance from the plant-stem and at a 5 cm depth in one, two, and three splits respectively showed difference N-efficiency. The highest N-efficiency of urea was obtained when applying it in two splits at the most active root-zone. (author)
International Nuclear Information System (INIS)
Lu, Ning; Qin, Jun; Yang, Kun; Sun, Jiulin
2011-01-01
Surface global solar radiation (GSR) is the primary renewable energy in nature. Geostationary satellite data are used to map GSR in many inversion algorithms in which ground GSR measurements merely serve to validate the satellite retrievals. In this study, a simple algorithm with artificial neural network (ANN) modeling is proposed to explore the non-linear physical relationship between ground daily GSR measurements and Multi-functional Transport Satellite (MTSAT) all-channel observations in an effort to fully exploit information contained in both data sets. Singular value decomposition is implemented to extract the principal signals from satellite data and a novel method is applied to enhance ANN performance at high altitude. A three-layer feed-forward ANN model is trained with one year of daily GSR measurements at ten ground sites. This trained ANN is then used to map continuous daily GSR for two years, and its performance is validated at all 83 ground sites in China. The evaluation result demonstrates that this algorithm can quickly and efficiently build the ANN model that estimates daily GSR from geostationary satellite data with good accuracy in both space and time. -- Highlights: → A simple and efficient algorithm to estimate GSR from geostationary satellite data. → ANN model fully exploits both the information from satellite and ground measurements. → Good performance of the ANN model is comparable to that of the classical models. → Surface elevation and infrared information enhance GSR inversion.
Estimating the Efficiency and Impacts of Petroleum Product Pricing Reforms in China
Directory of Open Access Journals (Sweden)
Chuxiong Deng
2018-04-01
Full Text Available The efficiency and effects analysis of a new pricing mechanism would have significant policy implications for the further design of a pricing mechanism in an emerging market. Unlike most of the existing literature, which focuses on the impacts to the macro-economy, this paper firstly uses an econometrics model to discuss the efficiency of the new pricing mechanism, and then establishes an augmented Phillips curve to estimate the impact of pricing reform on inflation in China. The results show that: (1 the new pricing mechanism would strengthen the linkage between Chinese oil prices and international oil prices; (2 oil price adjustments are still inadequate in China. (3 The lag in inflation is the most important factor that affects inflation, while the impact of the Chinese government’s price adjustments on inflation is limited and insignificant. In order to improve the efficiency of the petroleum products pricing mechanism and shorten lags, government should shorten the adjustment period and diminish the fluctuation threshold.
An Efficient Estimation Method for Reducing the Axial Intensity Drop in Circular Cone-Beam CT
Directory of Open Access Journals (Sweden)
Lei Zhu
2008-01-01
Full Text Available Reconstruction algorithms for circular cone-beam (CB scans have been extensively studied in the literature. Since insufficient data are measured, an exact reconstruction is impossible for such a geometry. If the reconstruction algorithm assumes zeros for the missing data, such as the standard FDK algorithm, a major type of resulting CB artifacts is the intensity drop along the axial direction. Many algorithms have been proposed to improve image quality when faced with this problem of data missing; however, development of an effective and computationally efficient algorithm remains a major challenge. In this work, we propose a novel method for estimating the unmeasured data and reducing the intensity drop artifacts. Each CB projection is analyzed in the Radon space via Grangeat's first derivative. Assuming the CB projection is taken from a parallel beam geometry, we extract those data that reside in the unmeasured region of the Radon space. These data are then used as in a parallel beam geometry to calculate a correction term, which is added together with Hu’s correction term to the FDK result to form a final reconstruction. More approximations are then made on the calculation of the additional term, and the final formula is implemented very efficiently. The algorithm performance is evaluated using computer simulations on analytical phantoms. The reconstruction comparison with results using other existing algorithms shows that the proposed algorithm achieves a superior performance on the reduction of axial intensity drop artifacts with a high computation efficiency.
Hashemi, Hoda Sadat; Boily, Mathieu; Martineau, Paul A.; Rivaz, Hassan
2017-03-01
Ultrasound elastography entails imaging mechanical properties of tissue and is therefore of significant clinical importance. In elastography, two frames of radio-frequency (RF) ultrasound data that are obtained while the tissue is undergoing deformation, and the time-delay estimate (TDE) between the two frames is used to infer mechanical properties of tissue. TDE is a critical step in elastography, and is challenging due to noise and signal decorrelation. This paper presents a novel and robust technique TDE using all samples of RF data simultaneously. We assume tissue deformation can be approximated by an affine transformation, and hence call our method ATME (Affine Transformation Model Elastography). The affine transformation model is utilized to obtain initial estimates of axial and lateral displacement fields. The affine transformation only has six degrees of freedom (DOF), and as such, can be efficiently estimated. A nonlinear cost function that incorporates similarity of RF data intensity and prior information of displacement continuity is formulated to fine-tune the initial affine deformation field. Optimization of this function involves searching for TDE of all samples of the RF data. The optimization problem is converted to a sparse linear system of equations, which can be solved in real-time. Results on simulation are presented for validation. We further collect RF data from in-vivo patellar tendon and medial collateral ligament (MCL), and show that ATME can be used to accurately track tissue displacement.
Efficient dense blur map estimation for automatic 2D-to-3D conversion
Vosters, L. P. J.; de Haan, G.
2012-03-01
Focus is an important depth cue for 2D-to-3D conversion of low depth-of-field images and video. However, focus can be only reliably estimated on edges. Therefore, Bea et al. [1] first proposed an optimization based approach to propagate focus to non-edge image portions, for single image focus editing. While their approach produces accurate dense blur maps, the computational complexity and memory requirements for solving the resulting sparse linear system with standard multigrid or (multilevel) preconditioning techniques, are infeasible within the stringent requirements of the consumer electronics and broadcast industry. In this paper we propose fast, efficient, low latency, line scanning based focus propagation, which mitigates the need for complex multigrid or (multilevel) preconditioning techniques. In addition we propose facial blur compensation to compensate for false shading edges that cause incorrect blur estimates in people's faces. In general shading leads to incorrect focus estimates, which may lead to unnatural 3D and visual discomfort. Since visual attention mostly tends to faces, our solution solves the most distracting errors. A subjective assessment by paired comparison on a set of challenging low-depth-of-field images shows that the proposed approach achieves equal 3D image quality as optimization based approaches, and that facial blur compensation results in a significant improvement.
Teramoto, Reiji; Saito, Chiaki; Funahashi, Shin-ichi
2014-06-30
Knockdown or overexpression of genes is widely used to identify genes that play important roles in many aspects of cellular functions and phenotypes. Because next-generation sequencing generates high-throughput data that allow us to detect genes, it is important to identify genes that drive functional and phenotypic changes of cells. However, conventional methods rely heavily on the assumption of normality and they often give incorrect results when the assumption is not true. To relax the Gaussian assumption in causal inference, we introduce the non-paranormal method to test conditional independence in the PC-algorithm. Then, we present the non-paranormal intervention-calculus when the directed acyclic graph (DAG) is absent (NPN-IDA), which incorporates the cumulative nature of effects through a cascaded pathway via causal inference for ranking causal genes against a phenotype with the non-paranormal method for estimating DAGs. We demonstrate that causal inference with the non-paranormal method significantly improves the performance in estimating DAGs on synthetic data in comparison with the original PC-algorithm. Moreover, we show that NPN-IDA outperforms the conventional methods in exploring regulators of the flowering time in Arabidopsis thaliana and regulators that control the browning of white adipocytes in mice. Our results show that performance improvement in estimating DAGs contributes to an accurate estimation of causal effects. Although the simplest alternative procedure was used, our proposed method enables us to design efficient intervention experiments and can be applied to a wide range of research purposes, including drug discovery, because of its generality.
An efficient modularized sample-based method to estimate the first-order Sobol' index
International Nuclear Information System (INIS)
Li, Chenzhao; Mahadevan, Sankaran
2016-01-01
Sobol' index is a prominent methodology in global sensitivity analysis. This paper aims to directly estimate the Sobol' index based only on available input–output samples, even if the underlying model is unavailable. For this purpose, a new method to calculate the first-order Sobol' index is proposed. The innovation is that the conditional variance and mean in the formula of the first-order index are calculated at an unknown but existing location of model inputs, instead of an explicit user-defined location. The proposed method is modularized in two aspects: 1) index calculations for different model inputs are separate and use the same set of samples; and 2) model input sampling, model evaluation, and index calculation are separate. Due to this modularization, the proposed method is capable to compute the first-order index if only input–output samples are available but the underlying model is unavailable, and its computational cost is not proportional to the dimension of the model inputs. In addition, the proposed method can also estimate the first-order index with correlated model inputs. Considering that the first-order index is a desired metric to rank model inputs but current methods can only handle independent model inputs, the proposed method contributes to fill this gap. - Highlights: • An efficient method to estimate the first-order Sobol' index. • Estimate the index from input–output samples directly. • Computational cost is not proportional to the number of model inputs. • Handle both uncorrelated and correlated model inputs.
Efficient Estimation of Dynamic Density Functions with Applications in Streaming Data
Qahtan, Abdulhakim
2016-05-11
Recent advances in computing technology allow for collecting vast amount of data that arrive continuously in the form of streams. Mining data streams is challenged by the speed and volume of the arriving data. Furthermore, the underlying distribution of the data changes over the time in unpredicted scenarios. To reduce the computational cost, data streams are often studied in forms of condensed representation, e.g., Probability Density Function (PDF). This thesis aims at developing an online density estimator that builds a model called KDE-Track for characterizing the dynamic density of the data streams. KDE-Track estimates the PDF of the stream at a set of resampling points and uses interpolation to estimate the density at any given point. To reduce the interpolation error and computational complexity, we introduce adaptive resampling where more/less resampling points are used in high/low curved regions of the PDF. The PDF values at the resampling points are updated online to provide up-to-date model of the data stream. Comparing with other existing online density estimators, KDE-Track is often more accurate (as reflected by smaller error values) and more computationally efficient (as reflected by shorter running time). The anytime available PDF estimated by KDE-Track can be applied for visualizing the dynamic density of data streams, outlier detection and change detection in data streams. In this thesis work, the first application is to visualize the taxi traffic volume in New York city. Utilizing KDE-Track allows for visualizing and monitoring the traffic flow on real time without extra overhead and provides insight analysis of the pick up demand that can be utilized by service providers to improve service availability. The second application is to detect outliers in data streams from sensor networks based on the estimated PDF. The method detects outliers accurately and outperforms baseline methods designed for detecting and cleaning outliers in sensor data. The
Towards physiologically meaningful water-use efficiency estimates from eddy covariance data.
Knauer, Jürgen; Zaehle, Sönke; Medlyn, Belinda E; Reichstein, Markus; Williams, Christopher A; Migliavacca, Mirco; De Kauwe, Martin G; Werner, Christiane; Keitel, Claudia; Kolari, Pasi; Limousin, Jean-Marc; Linderson, Maj-Lena
2018-02-01
Intrinsic water-use efficiency (iWUE) characterizes the physiological control on the simultaneous exchange of water and carbon dioxide in terrestrial ecosystems. Knowledge of iWUE is commonly gained from leaf-level gas exchange measurements, which are inevitably restricted in their spatial and temporal coverage. Flux measurements based on the eddy covariance (EC) technique can overcome these limitations, as they provide continuous and long-term records of carbon and water fluxes at the ecosystem scale. However, vegetation gas exchange parameters derived from EC data are subject to scale-dependent and method-specific uncertainties that compromise their ecophysiological interpretation as well as their comparability among ecosystems and across spatial scales. Here, we use estimates of canopy conductance and gross primary productivity (GPP) derived from EC data to calculate a measure of iWUE (G 1 , "stomatal slope") at the ecosystem level at six sites comprising tropical, Mediterranean, temperate, and boreal forests. We assess the following six mechanisms potentially causing discrepancies between leaf and ecosystem-level estimates of G 1 : (i) non-transpirational water fluxes; (ii) aerodynamic conductance; (iii) meteorological deviations between measurement height and canopy surface; (iv) energy balance non-closure; (v) uncertainties in net ecosystem exchange partitioning; and (vi) physiological within-canopy gradients. Our results demonstrate that an unclosed energy balance caused the largest uncertainties, in particular if it was associated with erroneous latent heat flux estimates. The effect of aerodynamic conductance on G 1 was sufficiently captured with a simple representation. G 1 was found to be less sensitive to meteorological deviations between canopy surface and measurement height and, given that data are appropriately filtered, to non-transpirational water fluxes. Uncertainties in the derived GPP and physiological within-canopy gradients and their
International Nuclear Information System (INIS)
Filippini, Massimo; Hunt, Lester C.; Zorić, Jelena
2014-01-01
The promotion of energy efficiency is seen as one of the top priorities of EU energy policy (EC, 2010). In order to design and implement effective energy policy instruments, it is necessary to have information on energy demand price and income elasticities in addition to sound indicators of energy efficiency. This research combines the approaches taken in energy demand modelling and frontier analysis in order to econometrically estimate the level of energy efficiency for the residential sector in the EU-27 member states for the period 1996 to 2009. The estimates for the energy efficiency confirm that the EU residential sector indeed holds a relatively high potential for energy savings from reduced inefficiency. Therefore, despite the common objective to decrease ‘wasteful’ energy consumption, considerable variation in energy efficiency between the EU member states is established. Furthermore, an attempt is made to evaluate the impact of energy-efficiency measures undertaken in the EU residential sector by introducing an additional set of variables into the model and the results suggest that financial incentives and energy performance standards play an important role in promoting energy efficiency improvements, whereas informative measures do not have a significant impact. - Highlights: • The level of energy efficiency of the EU residential sector is estimated. • Considerable potential for energy savings from reduced inefficiency is established. • The impact of introduced energy-efficiency policy measures is also evaluated. • Financial incentives are found to promote energy efficiency improvements. • Energy performance standards also play an important role
Hagen, David R; Tidor, Bruce
2015-02-01
A major effort in systems biology is the development of mathematical models that describe complex biological systems at multiple scales and levels of abstraction. Determining the topology-the set of interactions-of a biological system from observations of the system's behavior is an important and difficult problem. Here we present and demonstrate new methodology for efficiently computing the probability distribution over a set of topologies based on consistency with existing measurements. Key features of the new approach include derivation in a Bayesian framework, incorporation of prior probability distributions of topologies and parameters, and use of an analytically integrable linearization based on the Fisher information matrix that is responsible for large gains in efficiency. The new method was demonstrated on a collection of four biological topologies representing a kinase and phosphatase that operate in opposition to each other with either processive or distributive kinetics, giving 8-12 parameters for each topology. The linearization produced an approximate result very rapidly (CPU minutes) that was highly accurate on its own, as compared to a Monte Carlo method guaranteed to converge to the correct answer but at greater cost (CPU weeks). The Monte Carlo method developed and applied here used the linearization method as a starting point and importance sampling to approach the Bayesian answer in acceptable time. Other inexpensive methods to estimate probabilities produced poor approximations for this system, with likelihood estimation showing its well-known bias toward topologies with more parameters and the Akaike and Schwarz Information Criteria showing a strong bias toward topologies with fewer parameters. These results suggest that this linear approximation may be an effective compromise, providing an answer whose accuracy is near the true Bayesian answer, but at a cost near the common heuristics.
Heidari, A. A.; Moayedi, A.; Abbaspour, R. Ali
2017-09-01
Automated fare collection (AFC) systems are regarded as valuable resources for public transport planners. In this paper, the AFC data are utilized to analysis and extract mobility patterns in a public transportation system. For this purpose, the smart card data are inserted into a proposed metaheuristic-based aggregation model and then converted to O-D matrix between stops, since the size of O-D matrices makes it difficult to reproduce the measured passenger flows precisely. The proposed strategy is applied to a case study from Haaglanden, Netherlands. In this research, moth-flame optimizer (MFO) is utilized and evaluated for the first time as a new metaheuristic algorithm (MA) in estimating transit origin-destination matrices. The MFO is a novel, efficient swarm-based MA inspired from the celestial navigation of moth insects in nature. To investigate the capabilities of the proposed MFO-based approach, it is compared to methods that utilize the K-means algorithm, gray wolf optimization algorithm (GWO) and genetic algorithm (GA). The sum of the intra-cluster distances and computational time of operations are considered as the evaluation criteria to assess the efficacy of the optimizers. The optimality of solutions of different algorithms is measured in detail. The traveler's behavior is analyzed to achieve to a smooth and optimized transport system. The results reveal that the proposed MFO-based aggregation strategy can outperform other evaluated approaches in terms of convergence tendency and optimality of the results. The results show that it can be utilized as an efficient approach to estimating the transit O-D matrices.
Directory of Open Access Journals (Sweden)
A. A. Heidari
2017-09-01
Full Text Available Automated fare collection (AFC systems are regarded as valuable resources for public transport planners. In this paper, the AFC data are utilized to analysis and extract mobility patterns in a public transportation system. For this purpose, the smart card data are inserted into a proposed metaheuristic-based aggregation model and then converted to O-D matrix between stops, since the size of O-D matrices makes it difficult to reproduce the measured passenger flows precisely. The proposed strategy is applied to a case study from Haaglanden, Netherlands. In this research, moth-flame optimizer (MFO is utilized and evaluated for the first time as a new metaheuristic algorithm (MA in estimating transit origin-destination matrices. The MFO is a novel, efficient swarm-based MA inspired from the celestial navigation of moth insects in nature. To investigate the capabilities of the proposed MFO-based approach, it is compared to methods that utilize the K-means algorithm, gray wolf optimization algorithm (GWO and genetic algorithm (GA. The sum of the intra-cluster distances and computational time of operations are considered as the evaluation criteria to assess the efficacy of the optimizers. The optimality of solutions of different algorithms is measured in detail. The traveler's behavior is analyzed to achieve to a smooth and optimized transport system. The results reveal that the proposed MFO-based aggregation strategy can outperform other evaluated approaches in terms of convergence tendency and optimality of the results. The results show that it can be utilized as an efficient approach to estimating the transit O-D matrices.
International Nuclear Information System (INIS)
Horowitz, Marvin J.; Bertoldi, Paolo
2015-01-01
This study is an impact analysis of European Union (EU) energy efficiency policy that employs both top-down energy consumption data and bottom-up energy efficiency statistics or indicators. As such, it may be considered a contribution to the effort called for in the EU's 2006 Energy Services Directive (ESD) to develop a harmonized calculation model. Although this study does not estimate the realized savings from individual policy measures, it does provide estimates of realized energy savings for energy efficiency policy measures in aggregate. Using fixed effects panel models, the annual cumulative savings in 2011 of combined household and manufacturing sector electricity and natural gas usage attributed to EU energy efficiency policies since 2000 is estimated to be 1136 PJ; the savings attributed to energy efficiency policies since 2006 is estimated to be 807 PJ, or the equivalent of 5.6% of 2011 EU energy consumption. As well as its contribution to energy efficiency policy analysis, this study adds to the development of methods that can improve the quality of information provided by standardized energy efficiency and sustainable resource indexes. - Highlights: • Impact analysis of European Union energy efficiency policy. • Harmonization of top-down energy consumption and bottom-up energy efficiency indicators. • Fixed effects models for Member States for household and manufacturing sectors and combined electricity and natural gas usage. • EU energy efficiency policies since 2000 are estimated to have saved 1136 Petajoules. • Energy savings attributed to energy efficiency policies since 2006 are 5.6 percent of 2011 combined electricity and natural gas usage.
An improved routine for the fast estimate of ion cyclotron heating efficiency in tokamak plasmas
International Nuclear Information System (INIS)
Brambilla, M.
1992-02-01
The subroutine ICEVAL for the rapid simulation of Ion Cyclotron Heating in tokamak plasmas is based on analytic estimates of the wave behaviour near resonances, and on drastic but reasonable simplifications of the real geometry. The subroutine has been rewritten to improve the model and to facilitate its use as input in transport codes. In the new version the influence of quasilinear minority heating on the damping efficiency is taken into account using the well-known Stix analytic approximation. Among other improvements are: a) the possibility of considering plasmas with more than two ion species; b) inclusion of Landau, Transit Time and collisional damping on the electrons non localised at resonances; c) better models for the antenna spectrum and for the construction of the power deposition profiles. The results of ICEVAL are compared in detail with those of the full-wave code FELICE for the case of Hydrogen minority heating in a Deuterium plasma; except for details which depend on the excitation of global eigenmodes, agreement is excellent. ICEVAL is also used to investigate the enhancement of the absorption efficiency due to quasilinear heating of the minority ions. The effect is a strongly non-linear function of the available power, and decreases rapidly with increasing concentration. For parameters typical of Asdex Upgrade plasmas, about 4 MW are required to produce a significant increase of the single-pass absorption at concentrations between 10 and 20%. (orig.)
FAST LABEL: Easy and efficient solution of joint multi-label and estimation problems
Sundaramoorthi, Ganesh
2014-06-01
We derive an easy-to-implement and efficient algorithm for solving multi-label image partitioning problems in the form of the problem addressed by Region Competition. These problems jointly determine a parameter for each of the regions in the partition. Given an estimate of the parameters, a fast approximate solution to the multi-label sub-problem is derived by a global update that uses smoothing and thresholding. The method is empirically validated to be robust to fine details of the image that plague local solutions. Further, in comparison to global methods for the multi-label problem, the method is more efficient and it is easy for a non-specialist to implement. We give sample Matlab code for the multi-label Chan-Vese problem in this paper! Experimental comparison to the state-of-the-art in multi-label solutions to Region Competition shows that our method achieves equal or better accuracy, with the main advantage being speed and ease of implementation.
Directory of Open Access Journals (Sweden)
Hossein Jafari Mansoorian
2017-01-01
Full Text Available Background & Aims of the Study: A feed forward artificial neural network (FFANN was developed to predict the efficiency of total petroleum hydrocarbon (TPH removal from a contaminated soil, using soil washing process with Tween 80. The main objective of this study was to assess the performance of developed FFANN model for the estimation of TPH removal. Materials and Methods: Several independent repressors including pH, shaking speed, surfactant concentration and contact time were used to describe the removal of TPH as a dependent variable in a FFANN model. 85% of data set observations were used for training the model and remaining 15% were used for model testing, approximately. The performance of the model was compared with linear regression and assessed, using Root of Mean Square Error (RMSE as goodness-of-fit measure Results: For the prediction of TPH removal efficiency, a FANN model with a three-hidden-layer structure of 4-3-1 and a learning rate of 0.01 showed the best predictive results. The RMSE and R2 for the training and testing steps of the model were obtained to be 2.596, 0.966, 10.70 and 0.78, respectively. Conclusion: For about 80% of the TPH removal efficiency can be described by the assessed regressors the developed model. Thus, focusing on the optimization of soil washing process regarding to shaking speed, contact time, surfactant concentration and pH can improve the TPH removal performance from polluted soils. The results of this study could be the basis for the application of FANN for the assessment of soil washing process and the control of petroleum hydrocarbon emission into the environments.
Towards the Estimation of an Efficient Benchmark Portfolio: The Case of Croatian Emerging Market
Directory of Open Access Journals (Sweden)
Dolinar Denis
2017-04-01
Full Text Available The fact that cap-weighted indices provide an inefficient risk-return trade-off is well known today. Various research approaches evolved suggesting alternative to cap-weighting in an effort to come up with a more efficient market index benchmark. In this paper we aim to use such an approach and focus on the Croatian capital market. We apply statistical shrinkage method suggested by Ledoit and Wolf (2004 to estimate the covariance matrix and follow the work of Amenc et al. (2011 to obtain estimates of expected returns that rely on risk-return trade-off. Empirical findings for the proposed portfolio optimization include out-of-sample and robustness testing. This way we compare the performance of the capital-weighted benchmark to the alternative and ensure that consistency is achieved in different volatility environments. Research findings do not seem to support relevant research results for the developed markets but rather complement earlier research (Zoričić et al., 2014.
Betowski, Don; Bevington, Charles; Allison, Thomas C
2016-01-19
Halogenated chemical substances are used in a broad array of applications, and new chemical substances are continually being developed and introduced into commerce. While recent research has considerably increased our understanding of the global warming potentials (GWPs) of multiple individual chemical substances, this research inevitably lags behind the development of new chemical substances. There are currently over 200 substances known to have high GWP. Evaluation of schemes to estimate radiative efficiency (RE) based on computational chemistry are useful where no measured IR spectrum is available. This study assesses the reliability of values of RE calculated using computational chemistry techniques for 235 chemical substances against the best available values. Computed vibrational frequency data is used to estimate RE values using several Pinnock-type models, and reasonable agreement with reported values is found. Significant improvement is obtained through scaling of both vibrational frequencies and intensities. The effect of varying the computational method and basis set used to calculate the frequency data is discussed. It is found that the vibrational intensities have a strong dependence on basis set and are largely responsible for differences in computed RE values.
International Nuclear Information System (INIS)
Mohan, V.; Chudalayandi, K.; Sundaram, M.; Krishnamony, S.
1996-01-01
Estimation of gaseous activity forms an important component of air monitoring at Madras Atomic Power Station (MAPS). The gases of importance are argon 41 an air activation product and fission product noble gas xenon 133. For estimating the concentration, the experimental method is used in which a grab sample is collected in a 100 ml volumetric standard flask. The activity of gas is then computed by gamma spectrometry using a predetermined efficiency estimated experimentally. An attempt is made using theoretical approach to validate the experimental method of efficiency estimation. Two analytical models named relative flux model and absolute activity model were developed independently of each other. Attention is focussed on the efficiencies for 41 Ar and 133 Xe. Results show that the present method of sampling and analysis using 100 ml volumetric flask is adequate and acceptable. (author). 5 refs., 2 tabs
Rosenblum, Michael; van der Laan, Mark J.
2010-01-01
Models, such as logistic regression and Poisson regression models, are often used to estimate treatment effects in randomized trials. These models leverage information in variables collected before randomization, in order to obtain more precise estimates of treatment effects. However, there is the danger that model misspecification will lead to bias. We show that certain easy to compute, model-based estimators are asymptotically unbiased even when the working model used is arbitrarily misspecified. Furthermore, these estimators are locally efficient. As a special case of our main result, we consider a simple Poisson working model containing only main terms; in this case, we prove the maximum likelihood estimate of the coefficient corresponding to the treatment variable is an asymptotically unbiased estimator of the marginal log rate ratio, even when the working model is arbitrarily misspecified. This is the log-linear analog of ANCOVA for linear models. Our results demonstrate one application of targeted maximum likelihood estimation. PMID:20628636
Rosenblum, Michael; van der Laan, Mark J
2010-04-01
Models, such as logistic regression and Poisson regression models, are often used to estimate treatment effects in randomized trials. These models leverage information in variables collected before randomization, in order to obtain more precise estimates of treatment effects. However, there is the danger that model misspecification will lead to bias. We show that certain easy to compute, model-based estimators are asymptotically unbiased even when the working model used is arbitrarily misspecified. Furthermore, these estimators are locally efficient. As a special case of our main result, we consider a simple Poisson working model containing only main terms; in this case, we prove the maximum likelihood estimate of the coefficient corresponding to the treatment variable is an asymptotically unbiased estimator of the marginal log rate ratio, even when the working model is arbitrarily misspecified. This is the log-linear analog of ANCOVA for linear models. Our results demonstrate one application of targeted maximum likelihood estimation.
Liu, Y.; Pau, G. S. H.; Finsterle, S.
2015-12-01
Parameter inversion involves inferring the model parameter values based on sparse observations of some observables. To infer the posterior probability distributions of the parameters, Markov chain Monte Carlo (MCMC) methods are typically used. However, the large number of forward simulations needed and limited computational resources limit the complexity of the hydrological model we can use in these methods. In view of this, we studied the implicit sampling (IS) method, an efficient importance sampling technique that generates samples in the high-probability region of the posterior distribution and thus reduces the number of forward simulations that we need to run. For a pilot-point inversion of a heterogeneous permeability field based on a synthetic ponded infiltration experiment simulated with TOUGH2 (a subsurface modeling code), we showed that IS with linear map provides an accurate Bayesian description of the parameterized permeability field at the pilot points with just approximately 500 forward simulations. We further studied the use of surrogate models to improve the computational efficiency of parameter inversion. We implemented two reduced-order models (ROMs) for the TOUGH2 forward model. One is based on polynomial chaos expansion (PCE), of which the coefficients are obtained using the sparse Bayesian learning technique to mitigate the "curse of dimensionality" of the PCE terms. The other model is Gaussian process regression (GPR) for which different covariance, likelihood and inference models are considered. Preliminary results indicate that ROMs constructed based on the prior parameter space perform poorly. It is thus impractical to replace this hydrological model by a ROM directly in a MCMC method. However, the IS method can work with a ROM constructed for parameters in the close vicinity of the maximum a posteriori probability (MAP) estimate. We will discuss the accuracy and computational efficiency of using ROMs in the implicit sampling procedure
Energy Technology Data Exchange (ETDEWEB)
Bootsma, G. J., E-mail: Gregory.Bootsma@rmp.uhn.on.ca [Radiation Medicine Program, Princess Margaret Cancer Centre, Toronto, Ontario M5G 2M9 (Canada); Verhaegen, F. [Department of Radiation Oncology - MAASTRO, GROW—School for Oncology and Developmental Biology, Maastricht University Medical Center, Maastricht 6201 BN (Netherlands); Medical Physics Unit, Department of Oncology, McGill University, Montreal, Quebec H3G 1A4 (Canada); Jaffray, D. A. [Radiation Medicine Program, Princess Margaret Cancer Centre, Toronto, Ontario M5G 2M9 (Canada); Ontario Cancer Institute, Princess Margaret Cancer Centre, Toronto, Ontario M5G 2M9 (Canada); Department of Radiation Oncology, University of Toronto, Toronto, Ontario M5G 2M9 (Canada)
2015-01-15
suitable GOF metric with strong correlation with the actual error of the scatter fit, S{sub F}. Fitting the scatter distribution to a limited sum of sine and cosine functions using a low-pass filtered fast Fourier transform provided a computationally efficient and accurate fit. The CMCF algorithm reduces the number of photon histories required by over four orders of magnitude. The simulated experiments showed that using a compensator reduced the computational time by a factor between 1.5 and 1.75. The scatter estimates for the simulated and measured data were computed between 35–93 s and 114–122 s, respectively, using 16 Intel Xeon cores (3.0 GHz). The CMCF scatter correction improved the contrast-to-noise ratio by 10%–50% and reduced the reconstruction error to under 3% for the simulated phantoms. Conclusions: The novel CMCF algorithm significantly reduces the computation time required to estimate the scatter distribution by reducing the statistical noise in the MC scatter estimate and limiting the number of projection angles that must be simulated. Using the scatter estimate provided by the CMCF algorithm to correct both simulated and real projection data showed improved reconstruction image quality.
International Nuclear Information System (INIS)
Cizelj, L.
1994-10-01
In this report, an original probabilistic model aimed to assess the efficiency of particular maintenance strategy in terms of tube failure probability is proposed. The model concentrates on axial through wall cracks in the residual stress dominated tube expansion transition zone. It is based on the recent developments in probabilistic fracture mechanics and accounts for scatter in material, geometry and crack propagation data. Special attention has been paid to model the uncertainties connected to non-destructive examination technique (e.g., measurement errors, non-detection probability). First and second order reliability methods (FORM and SORM) have been implemented to calculate the failure probabilities. This is the first time that those methods are applied to the reliability analysis of components containing stress-corrosion cracks. In order to predict the time development of the tube failure probabilities, an original linear elastic fracture mechanics based crack propagation model has been developed. It accounts for the residual and operating stresses together. Also, the model accounts for scatter in residual and operational stresses due to the random variations in tube geometry and material data. Due to the lack of reliable crack velocity vs load data, the non-destructive examination records of the crack propagation have been employed to estimate the velocities at the crack tips. (orig./GL) [de
Efficient Maximum Likelihood Estimation for Pedigree Data with the Sum-Product Algorithm.
Engelhardt, Alexander; Rieger, Anna; Tresch, Achim; Mansmann, Ulrich
2016-01-01
We analyze data sets consisting of pedigrees with age at onset of colorectal cancer (CRC) as phenotype. The occurrence of familial clusters of CRC suggests the existence of a latent, inheritable risk factor. We aimed to compute the probability of a family possessing this risk factor as well as the hazard rate increase for these risk factor carriers. Due to the inheritability of this risk factor, the estimation necessitates a costly marginalization of the likelihood. We propose an improved EM algorithm by applying factor graphs and the sum-product algorithm in the E-step. This reduces the computational complexity from exponential to linear in the number of family members. Our algorithm is as precise as a direct likelihood maximization in a simulation study and a real family study on CRC risk. For 250 simulated families of size 19 and 21, the runtime of our algorithm is faster by a factor of 4 and 29, respectively. On the largest family (23 members) in the real data, our algorithm is 6 times faster. We introduce a flexible and runtime-efficient tool for statistical inference in biomedical event data with latent variables that opens the door for advanced analyses of pedigree data. © 2017 S. Karger AG, Basel.
A laboratory method to estimate the efficiency of plant extract to neutralize soil acidity
Directory of Open Access Journals (Sweden)
Marcelo E. Cassiolato
2002-06-01
Full Text Available Water-soluble plant organic compounds have been proposed to be efficient in alleviating soil acidity. Laboratory methods were evaluated to estimate the efficiency of plant extracts to neutralize soil acidity. Plant samples were dried at 65ºC for 48 h and ground to pass 1 mm sieve. Plant extraction procedure was: transfer 3.0 g of plant sample to a becker, add 150 ml of deionized water, shake for 8 h at 175 rpm and filter. Three laboratory methods were evaluated: sigma (Ca+Mg+K of the plant extracts; electrical conductivity of the plant extracts and titration of plant extracts with NaOH solution between pH 3 to 7. These methods were compared with the effect of the plant extracts on acid soil chemistry. All laboratory methods were related with soil reaction. Increasing sigma (Ca+Mg+K, electrical conductivity and the volume of NaOH solution spent to neutralize H+ ion of the plant extracts were correlated with the effect of plant extract on increasing soil pH and exchangeable Ca and decreasing exchangeable Al. It is proposed the electrical conductivity method for estimating the efficiency of plant extract to neutralize soil acidity because it is easily adapted for routine analysis and uses simple instrumentations and materials.Tem sido proposto que os compostos orgânicos de plantas solúveis em água são eficientes na amenização da acidez do solo. Foram avaliados métodos de laboratório para estimar a eficiência dos extratos de plantas na neutralização da acidez do solo. Os materiais de plantas foram secos a 65º C por 48 horas, moídos e passados em peneira de 1mm. Utilizou-se o seguinte procedimento para obtenção do extrato de plantas: transferir 3.0 g da amostra de planta para um becker, adicionar 150 ml de água deionizada, agitar por 8h a 175 rpm e filtrar. Avaliaram-se três métodos de laboratório: sigma (Ca + Mg + K do extrato de planta, condutividade elétrica (CE do extrato de planta e titulação do extrato de planta com solu
International Nuclear Information System (INIS)
Naqvi, S.A.A.; Ashfaq, M.
2014-01-01
High yielding crop like maize is very important for countries like Pakistan, which is third cereal crop after wheat and rice. Maize accounts for 4.8 percent of the total cropped area and 4.82 percent of the value of agricultural production. It is grown all over the country but major areas are Sahiwal, Okara and Faisalabad. Chiniot is one of the distinct agroecological domains of central Punjab for the maize cultivation, that's why this district was selected for the study and the technical efficiency of hybrid maize farmers was estimated. The primary data of 120 farmers, 40 farmers from each of the three tehsils of Chiniot were collected in the year 2011. Causes of low yields for various farmers than the others, while using the same input bundle were estimated. The managerial factors causing the inefficiency of production were also measured. The average technical efficiency was estimated to be 91 percent, while it was found to be 94.8, 92.7 and 90.8 for large, medium and small farmers, respectively. Stochastic frontier production model was used to measure technical efficiency. Statistical software Frontier 4.1 was used to analyse the data to generate inferences because the estimates of efficiency were produced as a direct output from package. It was concluded that the efficiency can be enhanced by covering the inefficiency from the environmental variables, farmers personal characteristics and farming conditions. (author)
Rosati, A; Dejong, T M
2003-06-01
It has been theorized that photosynthetic radiation use efficiency (PhRUE) over the course of a day is constant for leaves throughout a canopy if leaf nitrogen content and photosynthetic properties are adapted to local light so that canopy photosynthesis over a day is optimized. To test this hypothesis, 'daily' photosynthesis of individual leaves of Solanum melongena plants was calculated from instantaneous rates of photosynthesis integrated over the daylight hours. Instantaneous photosynthesis was estimated from the photosynthetic responses to photosynthetically active radiation (PAR) and from the incident PAR measured on individual leaves during clear and overcast days. Plants were grown with either abundant or scarce N fertilization. Both net and gross daily photosynthesis of leaves were linearly related to daily incident PAR exposure of individual leaves, which implies constant PhRUE over a day throughout the canopy. The slope of these relationships (i.e. PhRUE) increased with N fertilization. When the relationship was calculated for hourly instead of daily periods, the regressions were curvilinear, implying that PhRUE changed with time of the day and incident radiation. Thus, linearity (i.e. constant PhRUE) was achieved only when data were integrated over the entire day. Using average PAR in place of instantaneous incident PAR increased the slope of the relationship between daily photosynthesis and incident PAR of individual leaves, and the regression became curvilinear. The slope of the relationship between daily gross photosynthesis and incident PAR of individual leaves increased for an overcast compared with a clear day, but the slope remained constant for net photosynthesis. This suggests that net PhRUE of all leaves (and thus of the whole canopy) may be constant when integrated over a day, not only when the incident PAR changes with depth in the canopy, but also when it varies on the same leaf owing to changes in daily incident PAR above the canopy. The
Efficient Levenberg-Marquardt minimization of the maximum likelihood estimator for Poisson deviates
International Nuclear Information System (INIS)
Laurence, T.; Chromy, B.
2010-01-01
Histograms of counted events are Poisson distributed, but are typically fitted without justification using nonlinear least squares fitting. The more appropriate maximum likelihood estimator (MLE) for Poisson distributed data is seldom used. We extend the use of the Levenberg-Marquardt algorithm commonly used for nonlinear least squares minimization for use with the MLE for Poisson distributed data. In so doing, we remove any excuse for not using this more appropriate MLE. We demonstrate the use of the algorithm and the superior performance of the MLE using simulations and experiments in the context of fluorescence lifetime imaging. Scientists commonly form histograms of counted events from their data, and extract parameters by fitting to a specified model. Assuming that the probability of occurrence for each bin is small, event counts in the histogram bins will be distributed according to the Poisson distribution. We develop here an efficient algorithm for fitting event counting histograms using the maximum likelihood estimator (MLE) for Poisson distributed data, rather than the non-linear least squares measure. This algorithm is a simple extension of the common Levenberg-Marquardt (L-M) algorithm, is simple to implement, quick and robust. Fitting using a least squares measure is most common, but it is the maximum likelihood estimator only for Gaussian-distributed data. Non-linear least squares methods may be applied to event counting histograms in cases where the number of events is very large, so that the Poisson distribution is well approximated by a Gaussian. However, it is not easy to satisfy this criterion in practice - which requires a large number of events. It has been well-known for years that least squares procedures lead to biased results when applied to Poisson-distributed data; a recent paper providing extensive characterization of these biases in exponential fitting is given. The more appropriate measure based on the maximum likelihood estimator (MLE
The efficiency of modified jackknife and ridge type regression estimators: a comparison
Directory of Open Access Journals (Sweden)
Sharad Damodar Gore
2008-09-01
Full Text Available A common problem in multiple regression models is multicollinearity, which produces undesirable effects on the least squares estimator. To circumvent this problem, two well known estimation procedures are often suggested in the literature. They are Generalized Ridge Regression (GRR estimation suggested by Hoerl and Kennard iteb8 and the Jackknifed Ridge Regression (JRR estimation suggested by Singh et al. iteb13. The GRR estimation leads to a reduction in the sampling variance, whereas, JRR leads to a reduction in the bias. In this paper, we propose a new estimator namely, Modified Jackknife Ridge Regression Estimator (MJR. It is based on the criterion that combines the ideas underlying both the GRR and JRR estimators. We have investigated standard properties of this new estimator. From a simulation study, we find that the new estimator often outperforms the LASSO, and it is superior to both GRR and JRR estimators, using the mean squared error criterion. The conditions under which the MJR estimator is better than the other two competing estimators have been investigated.
Efficient estimation of the robustness region of biological models with oscillatory behavior.
Directory of Open Access Journals (Sweden)
Mochamad Apri
Full Text Available Robustness is an essential feature of biological systems, and any mathematical model that describes such a system should reflect this feature. Especially, persistence of oscillatory behavior is an important issue. A benchmark model for this phenomenon is the Laub-Loomis model, a nonlinear model for cAMP oscillations in Dictyostelium discoideum. This model captures the most important features of biomolecular networks oscillating at constant frequencies. Nevertheless, the robustness of its oscillatory behavior is not yet fully understood. Given a system that exhibits oscillating behavior for some set of parameters, the central question of robustness is how far the parameters may be changed, such that the qualitative behavior does not change. The determination of such a "robustness region" in parameter space is an intricate task. If the number of parameters is high, it may be also time consuming. In the literature, several methods are proposed that partially tackle this problem. For example, some methods only detect particular bifurcations, or only find a relatively small box-shaped estimate for an irregularly shaped robustness region. Here, we present an approach that is much more general, and is especially designed to be efficient for systems with a large number of parameters. As an illustration, we apply the method first to a well understood low-dimensional system, the Rosenzweig-MacArthur model. This is a predator-prey model featuring satiation of the predator. It has only two parameters and its bifurcation diagram is available in the literature. We find a good agreement with the existing knowledge about this model. When we apply the new method to the high dimensional Laub-Loomis model, we obtain a much larger robustness region than reported earlier in the literature. This clearly demonstrates the power of our method. From the results, we conclude that the biological system underlying is much more robust than was realized until now.
Effect of LET on the efficiency of dose re-estimation in LiF using uv photo-transfer
Energy Technology Data Exchange (ETDEWEB)
Douglas, J A; Baker, D M; Marshall, M; Budd, T [UKAEA Atomic Energy Research Establishment, Harwell. Environmental and Medical Sciences Div.
1980-09-01
Glow curves from TLD600 and TLD700 extruded rods exposed to ..gamma..-, X- and neutron radiations have been compared before and after uv photo-transfer. Re-estimation efficiency increases with LET by an amount which varies from batch to batch.
DEFF Research Database (Denmark)
Kock, Anders Bredahl; Callot, Laurent
We show that the adaptive Lasso (aLasso) and the adaptive group Lasso (agLasso) are oracle efficient in stationary vector autoregressions where the number of parameters per equation is smaller than the number of observations. In particular, this means that the parameters are estimated consistently...
To Estimation of Efficient Usage of Organic Fuel in the Cycle of Steam Power Installations
Directory of Open Access Journals (Sweden)
A. P. Nesenchuk
2013-01-01
Full Text Available Tendencies of power engineering development in the world were shown in this article. There were carried out the thermodynamic Analysis of efficient usage of different types of fuel. This article shows the obtained result, which reflects that low-calorie fuel (from the point of thermodynamics is more efficient to use at steam power stations then high-energy fuel.
An Integrated Approach for Estimating the Energy Efficiency of Seventeen Countries
Directory of Open Access Journals (Sweden)
Chia-Nan Wang
2017-10-01
Full Text Available Increased energy efficiency is one of the most effective ways to achieve climate change mitigation. This study aims to evaluate the energy efficiency of seventeen countries. The evaluation is based on an integrated method that combines the super slack-based (super SBM model and the Malmquist productivity index (MPI to investigate the energy efficiency of seventeen countries during the period of 2010–2015. The results in this study are that the United States, Columbia, Japan, China, and Saudi Arabia perform the best in energy efficiency, whereas Brazil, Russia, Indonesia, and India perform the worst during the entire sample period. The energy efficiency of these countries arrived mainly from technological improvement. The study provides suggestions for the seventeen countries’ government to control the energy consumption and contribute to environmental protection.
Validation of an efficient visual method for estimating leaf area index ...
African Journals Online (AJOL)
This study aimed to evaluate the accuracy and applicability of a visual method for estimating LAI in clonal Eucalyptus grandis × E. urophylla plantations and to compare it with hemispherical photography, ceptometer and LAI-2000® estimates. Destructive sampling for direct determination of the actual LAI was performed in ...
Pishravian, Arash; Aghabozorgi Sahaf, Masoud Reza
2012-12-01
In this paper speech-music separation using Blind Source Separation is discussed. The separating algorithm is based on the mutual information minimization where the natural gradient algorithm is used for minimization. In order to do that, score function estimation from observation signals (combination of speech and music) samples is needed. The accuracy and the speed of the mentioned estimation will affect on the quality of the separated signals and the processing time of the algorithm. The score function estimation in the presented algorithm is based on Gaussian mixture based kernel density estimation method. The experimental results of the presented algorithm on the speech-music separation and comparing to the separating algorithm which is based on the Minimum Mean Square Error estimator, indicate that it can cause better performance and less processing time
Carroll, Raymond; Maity, Arnab; Mammen, Enno; Yu, Kyusang
2009-01-01
We consider the efficient estimation of a regression parameter in a partially linear additive nonparametric regression model from repeated measures data when the covariates are multivariate. To date, while there is some literature in the scalar covariate case, the problem has not been addressed in the multivariate additive model case. Ours represents a first contribution in this direction. As part of this work, we first describe the behavior of nonparametric estimators for additive models with repeated measures when the underlying model is not additive. These results are critical when one considers variants of the basic additive model. We apply them to the partially linear additive repeated-measures model, deriving an explicit consistent estimator of the parametric component; if the errors are in addition Gaussian, the estimator is semiparametric efficient. We also apply our basic methods to a unique testing problem that arises in genetic epidemiology; in combination with a projection argument we develop an efficient and easily computed testing scheme. Simulations and an empirical example from nutritional epidemiology illustrate our methods.
Carroll, Raymond
2009-04-23
We consider the efficient estimation of a regression parameter in a partially linear additive nonparametric regression model from repeated measures data when the covariates are multivariate. To date, while there is some literature in the scalar covariate case, the problem has not been addressed in the multivariate additive model case. Ours represents a first contribution in this direction. As part of this work, we first describe the behavior of nonparametric estimators for additive models with repeated measures when the underlying model is not additive. These results are critical when one considers variants of the basic additive model. We apply them to the partially linear additive repeated-measures model, deriving an explicit consistent estimator of the parametric component; if the errors are in addition Gaussian, the estimator is semiparametric efficient. We also apply our basic methods to a unique testing problem that arises in genetic epidemiology; in combination with a projection argument we develop an efficient and easily computed testing scheme. Simulations and an empirical example from nutritional epidemiology illustrate our methods.
The Approach to an Estimation of a Local Area Network Functioning Efficiency
Directory of Open Access Journals (Sweden)
M. M. Taraskin
2010-09-01
Full Text Available In the article authors call attention to a choice of system of metrics, which permits to take a qualitative assessment of local area network functioning efficiency in condition of computer attacks.
International Nuclear Information System (INIS)
Saidur, R.; Sattar, M.A.; Masjuki, H.H.; Ahmed, S.; Hashim, U.
2007-01-01
The purpose of this work is to apply the useful energy and exergy analysis models for different modes of transport in Malaysia and to compare the result with a few countries. In this paper, energy and exergy efficiencies of the various sub-sectors are presented by considering the energy and exergy flows from 1995 to 2003. Respective flow diagrams to find the overall energy and exergy efficiencies of Malaysian transportation sector are also presented. The estimated overall energy efficiency ranges from 22.74% (1999) to 22.98% (1998) with a mean of 22.82+/-0.06% and that of overall exergy efficiency ranges from 22.44% (2000) to 22.82% (1998) with a mean of 22.55+/-0.12%. The results are compared with respect to present energy and exergy efficiencies in each sub-sector. The transportation sector used about 40% of the total energy consumed in 2002. Therefore, it is important to identify the energy and exergy flows and the pertinent losses. The road sub-sector has appeared to be the most efficient one compared to the air and marine sub-sectors. Also found that the energy and exergy efficiencies of Malaysian transportation sector are lower than that of Turkey but higher than Norway
DEREGULATION, FINANCIAL CRISIS, AND BANK EFFICIENCY IN TAIWAN: AN ESTIMATION OF UNDESIRABLE OUTPUTS
Liao, Chang-Sheng
2018-01-01
Purpose- This study investigates the undesirable impacts of outputson bank efficiency and contributes to the literature by assessing howregulation policies and other events impact bank efficiency in Taiwan inregards to deregulation, financial crisis, and financial reform from 1993 to2011. Methodology- In order to effectively deal with both undesirableand desirable outputs, this study follows Seiford and Zhu (2002), who recommendusing the standard data envelopment analysis model to measure per...
An Efficient Code-Timing Estimator for DS-CDMA Systems over Resolvable Multipath Channels
Directory of Open Access Journals (Sweden)
Jian Li
2005-04-01
Full Text Available We consider the problem of training-based code-timing estimation for the asynchronous direct-sequence code-division multiple-access (DS-CDMA system. We propose a modified large-sample maximum-likelihood (MLSML estimator that can be used for the code-timing estimation for the DS-CDMA systems over the resolvable multipath channels in closed form. Simulation results show that MLSML can be used to provide a high correct acquisition probability and a high estimation accuracy. Simulation results also show that MLSML can have very good near-far resistant capability due to employing a data model similar to that for adaptive array processing where strong interferences can be suppressed.
Efficient Estimation of Dynamic Density Functions with Applications in Streaming Data
Qahtan, Abdulhakim Ali Ali
2016-01-01
application is to detect outliers in data streams from sensor networks based on the estimated PDF. The method detects outliers accurately and outperforms baseline methods designed for detecting and cleaning outliers in sensor data. The third application
Efficient Estimation for Diffusions Sampled at High Frequency Over a Fixed Time Interval
DEFF Research Database (Denmark)
Jakobsen, Nina Munkholt; Sørensen, Michael
Parametric estimation for diffusion processes is considered for high frequency observations over a fixed time interval. The processes solve stochastic differential equations with an unknown parameter in the diffusion coefficient. We find easily verified conditions on approximate martingale...
Yin, Xinyou; Belay, Daniel W; van der Putten, Peter E L; Struik, Paul C
2014-12-01
Maximum quantum yield for leaf CO2 assimilation under limiting light conditions (Φ CO2LL) is commonly estimated as the slope of the linear regression of net photosynthetic rate against absorbed irradiance over a range of low-irradiance conditions. Methodological errors associated with this estimation have often been attributed either to light absorptance by non-photosynthetic pigments or to some data points being beyond the linear range of the irradiance response, both causing an underestimation of Φ CO2LL. We demonstrate here that a decrease in photosystem (PS) photochemical efficiency with increasing irradiance, even at very low levels, is another source of error that causes a systematic underestimation of Φ CO2LL. A model method accounting for this error was developed, and was used to estimate Φ CO2LL from simultaneous measurements of gas exchange and chlorophyll fluorescence on leaves using various combinations of species, CO2, O2, or leaf temperature levels. The conventional linear regression method under-estimated Φ CO2LL by ca. 10-15%. Differences in the estimated Φ CO2LL among measurement conditions were generally accounted for by different levels of photorespiration as described by the Farquhar-von Caemmerer-Berry model. However, our data revealed that the temperature dependence of PSII photochemical efficiency under low light was an additional factor that should be accounted for in the model.
Vahidi, Vahid; Saberinia, Ebrahim; Regentova, Emma E.
2017-10-01
A channel estimation (CE) method based on compressed sensing (CS) is proposed to estimate the sparse and doubly selective (DS) channel for hyperspectral image transmission from unmanned aircraft vehicles to ground stations. The proposed method contains three steps: (1) the priori estimate of the channel by orthogonal matching pursuit (OMP), (2) calculation of the linear minimum mean square error (LMMSE) estimate of the received pilots given the estimated channel, and (3) estimate of the complex amplitudes and Doppler shifts of the channel using the enhanced received pilot data applying a second round of a CS algorithm. The proposed method is named DS-LMMSE-OMP, and its performance is evaluated by simulating transmission of AVIRIS hyperspectral data via the communication channel and assessing their fidelity for the automated analysis after demodulation. The performance of the DS-LMMSE-OMP approach is compared with that of two other state-of-the-art CE methods. The simulation results exhibit up to 8-dB figure of merit in the bit error rate and 50% improvement in the hyperspectral image classification accuracy.
A Modularized Efficient Framework for Non-Markov Time Series Estimation
Schamberg, Gabriel; Ba, Demba; Coleman, Todd P.
2018-06-01
We present a compartmentalized approach to finding the maximum a-posteriori (MAP) estimate of a latent time series that obeys a dynamic stochastic model and is observed through noisy measurements. We specifically consider modern signal processing problems with non-Markov signal dynamics (e.g. group sparsity) and/or non-Gaussian measurement models (e.g. point process observation models used in neuroscience). Through the use of auxiliary variables in the MAP estimation problem, we show that a consensus formulation of the alternating direction method of multipliers (ADMM) enables iteratively computing separate estimates based on the likelihood and prior and subsequently "averaging" them in an appropriate sense using a Kalman smoother. As such, this can be applied to a broad class of problem settings and only requires modular adjustments when interchanging various aspects of the statistical model. Under broad log-concavity assumptions, we show that the separate estimation problems are convex optimization problems and that the iterative algorithm converges to the MAP estimate. As such, this framework can capture non-Markov latent time series models and non-Gaussian measurement models. We provide example applications involving (i) group-sparsity priors, within the context of electrophysiologic specrotemporal estimation, and (ii) non-Gaussian measurement models, within the context of dynamic analyses of learning with neural spiking and behavioral observations.
Directory of Open Access Journals (Sweden)
Pin-Chih Wang
2014-09-01
Full Text Available This study is intended to conduct an extended evaluation of sustainability based on the material flow analysis of resource productivity. We first present updated information on the material flow analysis (MFA database in Taiwan. Essential indicators are selected to quantify resource productivity associated with the economy-wide MFA of Taiwan. The study also applies the IPAT (impact-population-affluence-technology master equation to measure trends of material use efficiency in Taiwan and to compare them with those of other Asia-Pacific countries. An extended evaluation of efficiency, in comparison with selected economies by applying data envelopment analysis (DEA, is conducted accordingly. The Malmquist Productivity Index (MPI is thereby adopted to quantify the patterns and the associated changes of efficiency. Observations and summaries can be described as follows. Based on the MFA of the Taiwanese economy, the average growth rates of domestic material input (DMI; 2.83% and domestic material consumption (DMC; 2.13% in the past two decades were both less than that of gross domestic product (GDP; 4.95%. The decoupling of environmental pressures from economic growth can be observed. In terms of the decomposition analysis of the IPAT equation and in comparison with 38 other economies, the material use efficiency of Taiwan did not perform as well as its economic growth. The DEA comparisons of resource productivity show that Denmark, Germany, Luxembourg, Malta, Netherlands, United Kingdom and Japan performed the best in 2008. Since the MPI consists of technological change (frontier-shift or innovation and efficiency change (catch-up, the change in efficiency (catch-up of Taiwan has not been accomplished as expected in spite of the increase in its technological efficiency.
DEFF Research Database (Denmark)
Henningsen, Arne; Fabricius, Ole; Olsen, Jakob Vesterlund
2014-01-01
Based on a theoretical microeconomic model, we econometrically estimate investment utilization, adjustment costs, and technical efficiency in Danish pig farms based on a large unbalanced panel dataset. As our theoretical model indicates that adjustment costs are caused both by increased inputs...... of investment activities by the maximum likelihood method so that we can estimate the adjustment costs that occur in the year of the investment and the three following years. Our results show that investments are associated with significant adjustment costs, especially in the year in which the investment...
Bellier, Edwige; Grøtan, Vidar; Engen, Steinar; Schartau, Ann Kristin; Diserud, Ola H; Finstad, Anders G
2012-10-01
Obtaining accurate estimates of diversity indices is difficult because the number of species encountered in a sample increases with sampling intensity. We introduce a novel method that requires that the presence of species in a sample to be assessed while the counts of the number of individuals per species are only required for just a small part of the sample. To account for species included as incidence data in the species abundance distribution, we modify the likelihood function of the classical Poisson log-normal distribution. Using simulated community assemblages, we contrast diversity estimates based on a community sample, a subsample randomly extracted from the community sample, and a mixture sample where incidence data are added to a subsample. We show that the mixture sampling approach provides more accurate estimates than the subsample and at little extra cost. Diversity indices estimated from a freshwater zooplankton community sampled using the mixture approach show the same pattern of results as the simulation study. Our method efficiently increases the accuracy of diversity estimates and comprehension of the left tail of the species abundance distribution. We show how to choose the scale of sample size needed for a compromise between information gained, accuracy of the estimates and cost expended when assessing biological diversity. The sample size estimates are obtained from key community characteristics, such as the expected number of species in the community, the expected number of individuals in a sample and the evenness of the community.
International Nuclear Information System (INIS)
Hoffman, Ian M.; Goldman, Charles A.; Rybka, Gregory; Leventis, Greg; Schwartz, Lisa; Sanstad, Alan H.; Schiller, Steven
2017-01-01
The program administrator and total cost of saved energy allow comparison of the cost of efficiency across utilities, states, and program types, and can identify potential performance improvements. Comparing program administrator cost with the total cost of saved energy can indicate the degree to which programs leverage investment by participants. Based on reported total costs and savings information for U.S. utility efficiency programs from 2009 to 2013, we estimate the savings-weighted average total cost of saved electricity across 20 states at $0.046 per kilowatt-hour (kW h), comparing favorably with energy supply costs and retail rates. Programs targeted on the residential market averaged $0.030 per kW h compared to $0.053 per kW h for non-residential programs. Lighting programs, with an average total cost of $0.018 per kW h, drove lower savings costs in the residential market. We provide estimates for the most common program types and find that program administrators and participants on average are splitting the costs of efficiency in half. More consistent, standardized and complete reporting on efficiency programs is needed. Differing definitions and quantification of costs, savings and savings lifetimes pose challenges for comparing program results. Reducing these uncertainties could increase confidence in efficiency as a resource among planners and policymakers. - Highlights: • The cost of saved energy allows comparisons among energy resource investments. • Findings from the most expansive collection yet of total energy efficiency program costs. • The weighted average total cost of saved electricity was $0.046 for 20 states in 2009–2013. • Averages in the residential and non-residential sectors were $0.030 and $0.053 per kW h, respectively. • Results strongly indicate need for more consistent, reliable and complete reporting on efficiency programs.
Valid and efficient manual estimates of intracranial volume from magnetic resonance images
International Nuclear Information System (INIS)
Klasson, Niklas; Olsson, Erik; Rudemo, Mats; Eckerström, Carl; Malmgren, Helge; Wallin, Anders
2015-01-01
Manual segmentations of the whole intracranial vault in high-resolution magnetic resonance images are often regarded as very time-consuming. Therefore it is common to only segment a few linearly spaced intracranial areas to estimate the whole volume. The purpose of the present study was to evaluate how the validity of intracranial volume estimates is affected by the chosen interpolation method, orientation of the intracranial areas and the linear spacing between them. Intracranial volumes were manually segmented on 62 participants from the Gothenburg MCI study using 1.5 T, T 1 -weighted magnetic resonance images. Estimates of the intracranial volumes were then derived using subsamples of linearly spaced coronal, sagittal or transversal intracranial areas from the same volumes. The subsamples of intracranial areas were interpolated into volume estimates by three different interpolation methods. The linear spacing between the intracranial areas ranged from 2 to 50 mm and the validity of the estimates was determined by comparison with the entire intracranial volumes. A progressive decrease in intra-class correlation and an increase in percentage error could be seen with increased linear spacing between intracranial areas. With small linear spacing (≤15 mm), orientation of the intracranial areas and interpolation method had negligible effects on the validity. With larger linear spacing, the best validity was achieved using cubic spline interpolation with either coronal or sagittal intracranial areas. Even at a linear spacing of 50 mm, cubic spline interpolation on either coronal or sagittal intracranial areas had a mean absolute agreement intra-class correlation with the entire intracranial volumes above 0.97. Cubic spline interpolation in combination with linearly spaced sagittal or coronal intracranial areas overall resulted in the most valid and robust estimates of intracranial volume. Using this method, valid ICV estimates could be obtained in less than five
International Nuclear Information System (INIS)
Knoche, H.W.; Parkhurst, A.M.; Tam, S.W.
1979-01-01
The effect of volume on the liquid scintillation counting performance of 14 C-samples has been investigated. A decrease in counting efficiency was observed for samples with volumes below about 6 ml and those above about 18 ml when unquenched samples were assayed. Two quench-correction methods, sample channels ratio and external standard channels ratio, and three different liquid scintillation counters, were used in an investigation to determine the magnitude of the error in predicting counting efficiencies when small volume samples (2 ml) with different levels of quenching were assayed. The 2 ml samples exhibited slightly greater standard deviations of the difference between predicted and determined counting efficiencies than did 15 ml samples. Nevertheless, the magnitude of the errors indicate that if the sample channels ratio method of quench correction is employed, 2 ml samples may be counted in conventional counting vials with little loss in counting precision. (author)
Energy Technology Data Exchange (ETDEWEB)
Messenger, Mike; Bharvirkar, Ranjit; Golemboski, Bill; Goldman, Charles A.; Schiller, Steven R.
2010-04-14
Public and private funding for end-use energy efficiency actions is expected to increase significantly in the United States over the next decade. For example, Barbose et al (2009) estimate that spending on ratepayer-funded energy efficiency programs in the U.S. could increase from $3.1 billion in 2008 to $7.5 and 12.4 billion by 2020 under their medium and high scenarios. This increase in spending could yield annual electric energy savings ranging from 0.58% - 0.93% of total U.S. retail sales in 2020, up from 0.34% of retail sales in 2008. Interest in and support for energy efficiency has broadened among national and state policymakers. Prominent examples include {approx}$18 billion in new funding for energy efficiency programs (e.g., State Energy Program, Weatherization, and Energy Efficiency and Conservation Block Grants) in the 2009 American Recovery and Reinvestment Act (ARRA). Increased funding for energy efficiency should result in more benefits as well as more scrutiny of these results. As energy efficiency becomes a more prominent component of the U.S. national energy strategy and policies, assessing the effectiveness and energy saving impacts of energy efficiency programs is likely to become increasingly important for policymakers and private and public funders of efficiency actions. Thus, it is critical that evaluation, measurement, and verification (EM&V) is carried out effectively and efficiently, which implies that: (1) Effective program evaluation, measurement, and verification (EM&V) methodologies and tools are available to key stakeholders (e.g., regulatory agencies, program administrators, consumers, and evaluation consultants); and (2) Capacity (people and infrastructure resources) is available to conduct EM&V activities and report results in ways that support program improvement and provide data that reliably compares achieved results against goals and similar programs in other jurisdictions (benchmarking). The National Action Plan for Energy
Application of Artificial Neural Networks for Efficient High-Resolution 2D DOA Estimation
Directory of Open Access Journals (Sweden)
M. Agatonović
2012-12-01
Full Text Available A novel method to provide high-resolution Two-Dimensional Direction of Arrival (2D DOA estimation employing Artificial Neural Networks (ANNs is presented in this paper. The observed space is divided into azimuth and elevation sectors. Multilayer Perceptron (MLP neural networks are employed to detect the presence of a source in a sector while Radial Basis Function (RBF neural networks are utilized for DOA estimation. It is shown that a number of appropriately trained neural networks can be successfully used for the high-resolution DOA estimation of narrowband sources in both azimuth and elevation. The training time of each smaller network is significantly re¬duced as different training sets are used for networks in detection and estimation stage. By avoiding the spectral search, the proposed method is suitable for real-time ap¬plications as it provides DOA estimates in a matter of seconds. At the same time, it demonstrates the accuracy comparable to that of the super-resolution 2D MUSIC algorithm.
Directory of Open Access Journals (Sweden)
Francesco Sarracino
2017-04-01
Full Text Available Recent studies documented that survey data contain duplicate records. We assess how duplicate records affect regression estimates, and we evaluate the effectiveness of solutions to deal with duplicate records. Results show that the chances of obtaining unbiased estimates when data contain 40 doublets (about 5% of the sample range between 3.5% and 11.5% depending on the distribution of duplicates. If 7 quintuplets are present in the data (2% of the sample, then the probability of obtaining biased estimates ranges between 11% and 20%. Weighting the duplicate records by the inverse of their multiplicity, or dropping superfluous duplicates outperform other solutions in all considered scenarios. Our results illustrate the risk of using data in presence of duplicate records and call for further research on strategies to analyze affected data.
Efficient spectral estimation by MUSIC and ESPRIT with application to sparse FFT
Directory of Open Access Journals (Sweden)
Daniel ePotts
2016-02-01
Full Text Available In spectral estimation, one has to determine all parameters of an exponential sum for finitely many (noisysampled data of this exponential sum.Frequently used methods for spectral estimation are MUSIC (MUltiple SIgnal Classification and ESPRIT (Estimation of Signal Parameters viaRotational Invariance Technique.For a trigonometric polynomial of large sparsity, we present a new sparse fast Fourier transform byshifted sampling and using MUSIC resp. ESPRIT, where the ESPRIT based method has lower computational cost.Later this technique is extended to a new reconstruction of a multivariate trigonometric polynomial of large sparsity for given (noisy values sampled on a reconstructing rank-1 lattice. Numerical experiments illustrate thehigh performance of these procedures.
Quantum tomography via compressed sensing: error bounds, sample complexity and efficient estimators
International Nuclear Information System (INIS)
Flammia, Steven T; Gross, David; Liu, Yi-Kai; Eisert, Jens
2012-01-01
Intuitively, if a density operator has small rank, then it should be easier to estimate from experimental data, since in this case only a few eigenvectors need to be learned. We prove two complementary results that confirm this intuition. Firstly, we show that a low-rank density matrix can be estimated using fewer copies of the state, i.e. the sample complexity of tomography decreases with the rank. Secondly, we show that unknown low-rank states can be reconstructed from an incomplete set of measurements, using techniques from compressed sensing and matrix completion. These techniques use simple Pauli measurements, and their output can be certified without making any assumptions about the unknown state. In this paper, we present a new theoretical analysis of compressed tomography, based on the restricted isometry property for low-rank matrices. Using these tools, we obtain near-optimal error bounds for the realistic situation where the data contain noise due to finite statistics, and the density matrix is full-rank with decaying eigenvalues. We also obtain upper bounds on the sample complexity of compressed tomography, and almost-matching lower bounds on the sample complexity of any procedure using adaptive sequences of Pauli measurements. Using numerical simulations, we compare the performance of two compressed sensing estimators—the matrix Dantzig selector and the matrix Lasso—with standard maximum-likelihood estimation (MLE). We find that, given comparable experimental resources, the compressed sensing estimators consistently produce higher fidelity state reconstructions than MLE. In addition, the use of an incomplete set of measurements leads to faster classical processing with no loss of accuracy. Finally, we show how to certify the accuracy of a low-rank estimate using direct fidelity estimation, and describe a method for compressed quantum process tomography that works for processes with small Kraus rank and requires only Pauli eigenstate preparations
Estimating the Efficiency of Therapy Groups in a College Counseling Center
Weatherford, Ryan D.
2017-01-01
College counseling centers are facing rapidly increasing demands for services and are tasked to find efficient ways of providing adequate services while managing limited space. The use of therapy groups has been proposed as a method of managing demand. This brief report examines the clinical time savings of a traditional group therapy program in a…
Satyavada, Harish; Baldi, S.
2018-01-01
The operating principle of condensing boilers is based on exploiting heat from flue gases to pre-heat cold water at the inlet of the boiler: by condensing into liquid form, flue gases recover their latent heat of vaporization, leading to 10–12% increased efficiency with respect to traditional
de Graaf, C.S.L.; Kandhai, D.; Sloot, P.M.A.
According to Basel III, financial institutions have to charge a credit valuation adjustment (CVA) to account for a possible counterparty default. Calculating this measure and its sensitivities is one of the biggest challenges in risk management. Here, we introduce an efficient method for the
C.S.L. de Graaf (Kees); B.D. Kandhai; P.M.A. Sloot
2017-01-01
htmlabstractAccording to Basel III, financial institutions have to charge a credit valuation adjustment (CVA) to account for a possible counterparty default. Calculating this measure and its sensitivities is one of the biggest challenges in risk management. Here, we introduce an efficient method
Estimating the Efficiency of Michigan's Rural and Urban Public School Districts
Maranowski, Rita
2012-01-01
This study examined student achievement in Michigan public school districts to determine if rural school districts are demonstrating greater financial efficiency by producing higher levels of student achievement than school districts in other geographic locations with similar socioeconomics. Three models were developed using multiple regression…
Estimation of efficiency of dust suppressing works at 30-km zone near the Chernobyl' NPP
International Nuclear Information System (INIS)
Bakin, R.I.; Tkachenko, A.V.; Sukhoruchkin, A.K.
1989-01-01
Data on efficiency of dust suppressing works at 30-km zone near NPP are analyzed. It is necessary: to reduce radionuclide content in the air in the spring, when the weather is dry, to conduct dust suppressing works on roads and sections of surface with nonfixed ground; in the summer, to wash roads every day. 3 figs
FAST LABEL: Easy and efficient solution of joint multi-label and estimation problems
Sundaramoorthi, Ganesh; Hong, Byungwoo
2014-01-01
that plague local solutions. Further, in comparison to global methods for the multi-label problem, the method is more efficient and it is easy for a non-specialist to implement. We give sample Matlab code for the multi-label Chan-Vese problem in this paper
Estimation of the economical and ecological efficiency of the solar heat supply in Russia
International Nuclear Information System (INIS)
Marchenko, O.V.; Solomin, S.V.
2001-01-01
One carried out numerical study of application efficiency of solar heat supply systems in the climatic conditions of Russia with regard to their economical competitiveness with organic fuel heat conventional sources and role in reduction of greenhouse gas releases. One defined the regions where (under certain conditions) application of solar energy to generate low-potential heat may be reasonable [ru
Dimmick, R. L.; Boyd, A.; Wolochow, H.
1975-01-01
Aerosols of KBr and AgNO3 were mixed, exposed to light in a glass tube and collected in the dark. About 15% of the collected material was reduced to silver upon development. Thus, two aerosols of particles that react to form a photo-reducible compound can be used to measure coagulation efficiency.
ESTIMATION OF EFFICIENCY OF MACHINERY FOR PRE-SOWING TREATMENT OF RADISH SEEDS FOR SEED PRODUCTION
Directory of Open Access Journals (Sweden)
S. M. Sirota
2016-01-01
Full Text Available The results of pre-sowing treatment of radish seeds aimed at increasing of seed production, yield and its productivity in protected area are presented. The density fractionation of radish seeds by gravity separator is recommended for improvement of planting material quality and increasing of utilization efficiency of frame area.
Estimating the Value of Price Risk Reduction in Energy Efficiency Investments in Buildings
Directory of Open Access Journals (Sweden)
Pekka Tuominen
2017-10-01
Full Text Available This paper presents a method for calculating the value of price risk reduction to a consumer that can be achieved with investments in energy efficiency. The value of price risk reduction is discussed to some length in general terms in the literature reviewed but, so far, no methodology for calculating the value has been presented. Here we suggest such a method. The problem of valuating price risk reduction is approached using a variation of the Black–Scholes model by considering a hypothetical financial instrument that a consumer would purchase to insure herself against unexpected price hikes. This hypothetical instrument is then compared with an actual energy efficiency investment that reaches the same level of price risk reduction. To demonstrate the usability of the method, case examples are calculated for typical single-family houses in Finland. The results show that the price risk entailed in household energy consumption can be reduced by a meaningful amount with energy efficiency investments, and that the monetary value of this reduction can be calculated. It is argued that this often-overlooked benefit of energy efficiency investments merits more consideration in future studies.
Czech Academy of Sciences Publication Activity Database
Krupková, Lenka; Marková, I.; Havránková, Kateřina; Pokorný, Radek; Urban, Otmar; Šigut, Ladislav; Pavelka, Marian; Cienciala, E.; Marek, Michal V.
2017-01-01
Roč. 31, č. 1 (2017), s. 325-337 ISSN 0931-1890 R&D Projects: GA MŠk(CZ) LO1415; GA MŠk(CZ) LM2015061 Institutional support: RVO:67179843 Keywords : Solar radiation * Biomass increment * Carbon flux * light use efficiency Subject RIV: GK - Forestry OBOR OECD: Forestry Impact factor: 1.842, year: 2016
Estimation of Transpiration and Water Use Efficiency Using Satellite and Field Observations
Choudhury, Bhaskar J.; Quick, B. E.
2003-01-01
Structure and function of terrestrial plant communities bring about intimate relations between water, energy, and carbon exchange between land surface and atmosphere. Total evaporation, which is the sum of transpiration, soil evaporation and evaporation of intercepted water, couples water and energy balance equations. The rate of transpiration, which is the major fraction of total evaporation over most of the terrestrial land surface, is linked to the rate of carbon accumulation because functioning of stomata is optimized by both of these processes. Thus, quantifying the spatial and temporal variations of the transpiration efficiency (which is defined as the ratio of the rate of carbon accumulation and transpiration), and water use efficiency (defined as the ratio of the rate of carbon accumulation and total evaporation), and evaluation of modeling results against observations, are of significant importance in developing a better understanding of land surface processes. An approach has been developed for quantifying spatial and temporal variations of transpiration, and water-use efficiency based on biophysical process-based models, satellite and field observations. Calculations have been done using concurrent meteorological data derived from satellite observations and four dimensional data assimilation for four consecutive years (1987-1990) over an agricultural area in the Northern Great Plains of the US, and compared with field observations within and outside the study area. The paper provides substantive new information about interannual variation, particularly the effect of drought, on the efficiency values at a regional scale.
Jiménez-Martínez, J.; Molinero-Huguet, J.; Candela, L.
2009-04-01
Water requirements for different crop types according to soil type and climate conditions play not only an important role in agricultural efficiency production, though also for water resources management and control of pollutants in drainage water. The key issue to attain these objectives is the irrigation efficiency. Application of computer codes for irrigation simulation constitutes a fast and inexpensive approach to study optimal agricultural management practices. To simulate daily water balance in the soil, vadose zone and aquifer the VisualBALAN V. 2.0 code was applied to an experimental area under irrigation characterized by its aridity. The test was carried out in three experimental plots for annual row crops (lettuce and melon), perennial vegetables (artichoke), and fruit trees (citrus) under common agricultural practices in open air for October 1999-September 2008. Drip irrigation was applied to crops production due to the scarcity of water resources and the need for water conservation. Water level change was monitored in the top unconfined aquifer for each experimental plot. Results of water balance modelling show a good agreement between observed and estimated water level values. For the study period, mean drainage obtained values were 343 mm, 261 mm and 205 mm for lettuce and melon, artichoke and citrus respectively. Assessment of water use efficiency was based on the IE indicator proposed by the ASCE Task Committee. For the modelled period, water use efficiency was estimated as 73, 71 and 78 % of the applied dose (irrigation + precipitation) for lettuce and melon, artichoke and citrus, respectively.
Vitikainen, Kirsi; Street, Andrew; Linna, Miika
2009-02-01
Hospital efficiency has been the subject of numerous health economics studies, but there is little evidence on how the chosen output and casemix measures affect the efficiency results. The aim of this study is to examine the robustness of efficiency results due to these factors. Comparison is made between activities and episode output measures, and two different output grouping systems (Classic and FullDRG). Non-parametric data envelopment analysis is used as an analysis technique. The data consist of all public acute care hospitals in Finland in 2005 (n=40). Efficiency estimates were not found to be highly sensitive to the choice between episode and activity descriptions of output, but more so to the choice of DRG grouping system. Estimates are most sensitive to scale assumptions, with evidence of decreasing returns to scale in larger hospitals. Episode measures are generally to be preferred to activity measures because these better capture the patient pathway, while FullDRGs are preferred to Classic DRGs particularly because of the better description of outpatient output in the former grouping system. Attention should be paid to reducing the extent of scale inefficiency in Finland.
Rhee, Seung-Whee
2017-09-01
In order to separate aluminum from the base-cap of spent fluorescent lamp (SFL), the separation efficiency of hammer crusher unit is estimated by introducing a binary separation theory. The base-cap of SFL is composed by glass fragment, binder, ferrous metal, copper and aluminum. The hammer crusher unit to recover aluminum from the base-cap consists of 3stages of hammer crusher, magnetic separator and vibrating screen. The optimal conditions of rotating speed and operating time in the hammer crusher unit are decided at each stage. At the optimal conditions, the aluminum yield and the separation efficiency of hammer crusher unit are estimated by applying a sequential binary separation theory at each stage. And the separation efficiency between hammer crusher unit and roll crush system is compared to show the performance of aluminum recovery from the base-cap of SFL. Since the separation efficiency can be increased to 99% at stage 3, from the experimental results, it is found that aluminum from the base-cap can be sufficiently recovered by the hammer crusher unit. Copyright © 2017. Published by Elsevier Ltd.
How to efficiently obtain accurate estimates of flower visitation rates by pollinators
Fijen, Thijs P.M.; Kleijn, David
2017-01-01
Regional declines in insect pollinators have raised concerns about crop pollination. Many pollinator studies use visitation rate (pollinators/time) as a proxy for the quality of crop pollination. Visitation rate estimates are based on observation durations that vary significantly between studies.
Economical efficiency estimation of the power system with an accelerator breeder
International Nuclear Information System (INIS)
Rublev, O.V.; Komin, A.V.
1990-01-01
The review deals with economical indices of nuclear power system with an accelerator breeder producing secondary nuclear fuel. Electric power cost was estimated by the method of discounted cost. Power system with accelerator breeder compares unfavourably with traditional nuclear power systems with respect to its capitalized cost
Jiang, George J.; Sluis, Pieter J. van der
1999-01-01
While the stochastic volatility (SV) generalization has been shown to improve the explanatory power over the Black-Scholes model, empirical implications of SV models on option pricing have not yet been adequately tested. The purpose of this paper is to first estimate a multivariate SV model using
Derrick, Timothy R; Edwards, W Brent; Fellin, Rebecca E; Seay, Joseph F
2016-02-08
The purpose of this research was to utilize a series of models to estimate the stress in a cross section of the tibia, located 62% from the proximal end, during walking. Twenty-eight male, active duty soldiers walked on an instrumented treadmill while external force data and kinematics were recorded. A rigid body model was used to estimate joint moments and reaction forces. A musculoskeletal model was used to gather muscle length, muscle velocity, moment arm and orientation information. Optimization procedures were used to estimate muscle forces and finally internal bone forces and moments were applied to an inhomogeneous, subject specific bone model obtained from CT scans to estimate stress in the bone cross section. Validity was assessed by comparison to stresses calculated from strain gage data in the literature and sensitivity was investigated using two simplified versions of the bone model-a homogeneous model and an ellipse approximation. Peak compressive stress occurred on the posterior aspect of the cross section (-47.5 ± 14.9 MPa). Peak tensile stress occurred on the anterior aspect (27.0 ± 11.7 MPa) while the location of peak shear was variable between subjects (7.2 ± 2.4 MPa). Peak compressive, tensile and shear stresses were within 0.52 MPa, 0.36 MPa and 3.02 MPa respectively of those calculated from the converted strain gage data. Peak values from a inhomogeneous model of the bone correlated well with homogeneous model (normal: 0.99; shear: 0.94) as did the normal ellipse model (r=0.89-0.96). However, the relationship between shear stress in the inhomogeneous model and ellipse model was less accurate (r=0.64). The procedures detailed in this paper provide a non-invasive and relatively quick method of estimating cross sectional stress that holds promise for assessing injury and osteogenic stimulus in bone during normal physical activity. Copyright © 2016 Elsevier Ltd. All rights reserved.
SU-E-I-65: Estimation of Tagging Efficiency in Pseudo-Continuous Arterial Spin Labeling (pCASL) MRI
Energy Technology Data Exchange (ETDEWEB)
Jen, M [Chang Gung University, Taoyuan City, Taiwan (China); Yan, F; Tseng, Y; Chen, C [Taipei Medical University - Shuang Ho Hospital, Ministry of Health and Welf, New Taipei City, Taiwan (China); Lin, C [GE Healthcare, Taiwan (China); GE Healthcare China, Beijing (China); Liu, H [UT MD Anderson Cancer Center, Houston, TX (United States)
2015-06-15
Purpose: pCASL was recommended as a potent approach for absolute cerebral blood flow (CBF) quantification in clinical practice. However, uncertainties of tagging efficiency in pCASL remain an issue. This study aimed to estimate tagging efficiency by using short quantitative pulsed ASL scan (FAIR-QUIPSSII) and compare resultant CBF values with those calibrated by using 2D Phase Contrast (PC) MRI. Methods: Fourteen normal volunteers participated in this study. All images, including whole brain (WB) pCASL, WB FAIR-QUIPSSII and single-slice 2D PC, were collected on a 3T clinical MRI scanner with a 8-channel head coil. DeltaM map was calculated by averaging the subtraction of tag/control pairs in pCASL and FAIR-QUIPSSII images and used for CBF calculation. Tagging efficiency was then calculated by the ratio of mean gray matter CBF obtained from pCASL and FAIR-QUIPSSII. For comparison, tagging efficiency was also estimated with 2D PC, a previously established method, by contrast WB CBF in pCASL and 2D PC. Feasibility of estimation from a short FAIR-QUIPSSII scan was evaluated by number of averages required for obtaining a stable deltaM value. Setting deltaM calculated by maximum number of averaging (50 pairs) as reference, stable results were defined within ±10% variation. Results: Tagging efficiencies obtained by 2D PC MRI (0.732±0.092) were significantly lower than which obtained by FAIRQUIPPSSII (0.846±0.097) (P<0.05). Feasibility results revealed that four pairs of images in FAIR-QUIPPSSII scan were sufficient to obtain a robust calibration of less than 10% differences from using 50 pairs. Conclusion: This study found that reliable estimation of tagging efficiency could be obtained by a few pairs of FAIR-QUIPSSII images, which suggested that calibration scan in a short duration (within 30s) was feasible. Considering recent reports concerning variability of PC MRI-based calibration, this study proposed an effective alternative for CBF quantification with pCASL.
A Methodology for the Estimation of the Wind Generator Economic Efficiency
Zaleskis, G.
2017-12-01
Integration of renewable energy sources and the improvement of the technological base may not only reduce the consumption of fossil fuel and environmental load, but also ensure the power supply in regions with difficult fuel delivery or power failures. The main goal of the research is to develop the methodology of evaluation of the wind turbine economic efficiency. The research has demonstrated that the electricity produced from renewable sources may be much more expensive than the electricity purchased from the conventional grid.
Efficient quality-eactor estimation of a vertical cavity employing a high-contrast grating
DEFF Research Database (Denmark)
Taghizadeh, Alireza; Mørk, Jesper; Chung, Il-Sug
2017-01-01
Hybrid vertical cavity lasers employing high-contrast grating reflectors are attractive for Si-integrated light source applications. Here, a method for reducing a three-dimensional (3D) optical simulation of this laser structure to lower-dimensional simulations is suggested, which allows for very...... fast and approximate analysis of the quality-factor of the 3D cavity. This approach enables us to efficiently optimize the laser cavity design without performing cumbersome 3D simulations....
ROSATI, A.; DEJONG, T. M.
2003-01-01
It has been theorized that photosynthetic radiation use efficiency (PhRUE) over the course of a day is constant for leaves throughout a canopy if leaf nitrogen content and photosynthetic properties are adapted to local light so that canopy photosynthesis over a day is optimized. To test this hypothesis, ‘daily’ photosynthesis of individual leaves of Solanum melongena plants was calculated from instantaneous rates of photosynthesis integrated over the daylight hours. Instantaneous photosynthes...
Estimation of the drift eliminator efficiency using numerical and experimental methods
Directory of Open Access Journals (Sweden)
Stodůlka Jiří
2016-01-01
Full Text Available The purpose of the drift eliminators is to prevent water from escaping in significant amounts the cooling tower. They are designed to catch the droplets dragged by the tower draft and the efficiency given by the shape of the eliminator is the main evaluation criteria. The ability to eliminate the escaping water droplets is studied using CFD and using the experimental IPI method.
International Nuclear Information System (INIS)
Hovhannisyan, L.S.; Harutyunyan, N.R.
2013-01-01
The power efficiency of the thermal power plant (TPP) modernization by using combined-cycle technologies is introduced. It is shown that it is possible to achieve the greatest decrease in the specific fuel consumption at modernizing the TPP at the expense of introducing progressive 'know-how' of the electric power generation: for TPP on gas, it is combined-cycle, gas-turbine superstructures of steam-power plants and gas-turbines with heat utilization
Thermal efficiency and particulate pollution estimation of four biomass fuels grown on wasteland
Energy Technology Data Exchange (ETDEWEB)
Kandpal, J.B.; Madan, M. [Indian Inst. of Tech., New Delhi (India). Centre for Rural Development and Technology
1996-10-01
The thermal performance and concentration of suspended particulate matter were studied for 1-hour combustion of four biomass fuels, namely Acacia nilotica, Leucaena leucocepholea, Jatropha curcus, and Morus alba grown in wasteland. Among the four biomass fuels, the highest thermal efficiency was achieved with Acacia nilotica. The suspended particulate matter concentration for 1-hour combustion of four biomass fuels ranged between 850 and 2,360 {micro}g/m{sup 3}.
An Efficient Power Estimation Methodology for Complex RISC Processor-based Platforms
Rethinagiri , Santhosh Kumar; Ben Atitallah , Rabie; Dekeyser , Jean-Luc; Niar , Smail; Senn , Eric
2012-01-01
International audience; In this contribution, we propose an efficient power estima- tion methodology for complex RISC processor-based plat- forms. In this methodology, the Functional Level Power Analysis (FLPA) is used to set up generic power models for the different parts of the system. Then, a simulation framework based on virtual platform is developed to evalu- ate accurately the activities used in the related power mod- els. The combination of the two parts above leads to a het- erogeneou...
Li, Zhigang; Yu, Xiaohua; Zeng, Yinchu
2011-01-01
Using a unique survey data on agricultural traders in China in 2004, this study provides direct evidence on significance of interregional transport costs and their key determinants. Our major findings are as follows: (1) the trade barriers within China are dominated by transport-related costs but not artificial barriers, approximated by tolls and fines; (2) Labor and fuels costs are the most significant component of transport costs; (3) road quality is very important for transportation effici...
El Gharamti, M.; Bethke, I.; Tjiputra, J.; Bertino, L.
2016-02-01
Given the recent strong international focus on developing new data assimilation systems for biological models, we present in this comparative study the application of newly developed state-parameters estimation tools to an ocean ecosystem model. It is quite known that the available physical models are still too simple compared to the complexity of the ocean biology. Furthermore, various biological parameters remain poorly unknown and hence wrong specifications of such parameters can lead to large model errors. Standard joint state-parameters augmentation technique using the ensemble Kalman filter (Stochastic EnKF) has been extensively tested in many geophysical applications. Some of these assimilation studies reported that jointly updating the state and the parameters might introduce significant inconsistency especially for strongly nonlinear models. This is usually the case for ecosystem models particularly during the period of the spring bloom. A better handling of the estimation problem is often carried out by separating the update of the state and the parameters using the so-called Dual EnKF. The dual filter is computationally more expensive than the Joint EnKF but is expected to perform more accurately. Using a similar separation strategy, we propose a new EnKF estimation algorithm in which we apply a one-step-ahead smoothing to the state. The new state-parameters estimation scheme is derived in a consistent Bayesian filtering framework and results in separate update steps for the state and the parameters. Unlike the classical filtering path, the new scheme starts with an update step and later a model propagation step is performed. We test the performance of the new smoothing-based schemes against the standard EnKF in a one-dimensional configuration of the Norwegian Earth System Model (NorESM) in the North Atlantic. We use nutrients profile (up to 2000 m deep) data and surface partial CO2 measurements from Mike weather station (66o N, 2o E) to estimate
Directory of Open Access Journals (Sweden)
Hakan Gunes
2016-12-01
Full Text Available This study aims to investigate the cost efficiency of Turkish commercial banks over the restructuring period of the Turkish banking system, which coincides with the 2008 financial global crisis and the 2010 European sovereign debt crisis. To this end, within the stochastic frontier framework, we employ true fixed effects model, where the unobserved bank heterogeneity is integrated in the inefficiency distribution at a mean level. To select the cost function with the most appropriate inefficiency correlates, we first adopt a search algorithm and then utilize the model averaging approach to verify that our results are not exposed to model selection bias. Overall, our empirical results reveal that cost efficiencies of Turkish banks have improved over time, with the effects of the 2008 and 2010 crises remaining rather limited. Furthermore, not only the cost efficiency scores but also impacts of the crises on those scores appear to vary with regard to bank size and ownership structure, in accordance with much of the existing literature.
Directory of Open Access Journals (Sweden)
Northcutt Sally L
2010-04-01
Full Text Available Abstract Background Molecular estimates of breeding value are expected to increase selection response due to improvements in the accuracy of selection and a reduction in generation interval, particularly for traits that are difficult or expensive to record or are measured late in life. Several statistical methods for incorporating molecular data into breeding value estimation have been proposed, however, most studies have utilized simulated data in which the generated linkage disequilibrium may not represent the targeted livestock population. A genomic relationship matrix was developed for 698 Angus steers and 1,707 Angus sires using 41,028 single nucleotide polymorphisms and breeding values were estimated using feed efficiency phenotypes (average daily feed intake, residual feed intake, and average daily gain recorded on the steers. The number of SNPs needed to accurately estimate a genomic relationship matrix was evaluated in this population. Results Results were compared to estimates produced from pedigree-based mixed model analysis of 862 Angus steers with 34,864 identified paternal relatives but no female ancestors. Estimates of additive genetic variance and breeding value accuracies were similar for AFI and RFI using the numerator and genomic relationship matrices despite fewer animals in the genomic analysis. Bootstrap analyses indicated that 2,500-10,000 markers are required for robust estimation of genomic relationship matrices in cattle. Conclusions This research shows that breeding values and their accuracies may be estimated for commercially important sires for traits recorded in experimental populations without the need for pedigree data to establish identity by descent between members of the commercial and experimental populations when at least 2,500 SNPs are available for the generation of a genomic relationship matrix.
Estimation of the energy efficiency of cryogenic filled tank use in different systems and devices
International Nuclear Information System (INIS)
Blagin, E.V.; Dovgyallo, A.I.; Nekrasova, S.O.; Sarmin, D.V.; Uglanov, D.A.
2016-01-01
Highlights: • The cryogenic fueling tank is a device for storage and gasification of working fluid. • Potential energy of pressure can be converted to electricity by circuit of turbines. • It is possible to compensate up to 8% of energy consumed for liquefaction. - Abstract: This article presents a device for storage and gasification of cryogenic working fluid. This device is called cryogenic fueling tank. Working fluid pressure increases during the gasification and potential energy of this pressure can be used in different ways. The ways of integrating the cryogenic fueling tank into existing energy plants are described in this article. The estimation of the cryogenic fueling tank application in the gasification facility as well as in the onboard power system was carried out. This estimation shows that application of such tank as well as a circuit of turbines allows generating up to near 8% of energy which was consumed during gas liquefaction. The estimation of the additionally generated electric energy value was also carried out for each of the cases.
An Improved Weise’s Rule for Efficient Estimation of Stand Quadratic Mean Diameter
Directory of Open Access Journals (Sweden)
Róbert Sedmák
2015-07-01
Full Text Available The main objective of this study was to explore the accuracy of Weise’s rule of thumb applied to an estimation of the quadratic mean diameter of a forest stand. Virtual stands of European beech (Fagus sylvatica L. across a range of structure types were stochastically generated and random sampling was simulated. We compared the bias and accuracy of stand quadratic mean diameter estimates, employing different ranks of measured stems from a set of the 10 trees nearest to the sampling point. We proposed several modifications of the original Weise’s rule based on the measurement and averaging of two different ranks centered to a target rank. In accordance with the original formulation of the empirical rule, we recommend the application of the measurement of the 6th stem in rank corresponding to the 55% sample percentile of diameter distribution, irrespective of mean diameter size and degree of diameter dispersion. The study also revealed that the application of appropriate two-measurement modifications of Weise’s method, the 4th and 8th ranks or 3rd and 9th ranks averaged to the 6th central rank, should be preferred over the classic one-measurement estimation. The modified versions are characterised by an improved accuracy (about 25% without statistically significant bias and measurement costs comparable to the classic Weise method.
Directory of Open Access Journals (Sweden)
Klitynska Oksana V.
2016-01-01
Full Text Available Background: Caries is a pathological process that occurs in the hard tissues of the teeth after eruption and reduced quality of life due to significant complications, especially in children. An extremely high incidence of dental caries among children living permanently in Transcarpathian region requires a comprehensive prevention program. The aim of this study was to determine the efficiency of complex caries prevention program among children permanently living in the area of biogeochemical fluorine deficiency. Aim of the study: To evaluate efficiency level of using phased program of caries prevention among children of different age groups domiciled in Transcarpathian region. Material and Methods: On examination of 346 children aged 3-8 years, among which 163 (46.9% boys and 183 (53.1% girls, a phased program of complex prophylaxis was created, covering the basic dental diseases in children living permanently in deficiency conditions. The program included: hygienic education of preschool children and their parents; exogenous medicament prevention; early identification and treatment of caries using conventional methods according to treatment protocols; endogenous non-medical prevention, nutrition correction have proved its effectiveness. Results: The indicator of caries prevention efficiency of the proposed scheme for children 5-7 (3-5 years is 69.5%; for children 8-10 age group (6-8 years - 66.9%. Conclusion: The main strategy of pediatric dental services in Ukraine should be created for the children population (aged up 18 years through national and regional programs for the primary prevention of main dental diseases with providing adequate financing in sufficient volume to preserve the nation's dental health for the next 20 years.
Directory of Open Access Journals (Sweden)
Sergey Kharitonov
2015-06-01
Full Text Available Optimum transport infrastructure usage is an important aspect of the development of the national economy of the Russian Federation. Thus, development of instruments for assessing the efficiency of infrastructure is impossible without constant monitoring of a number of significant indicators. This work is devoted to the selection of indicators and the method of their calculation in relation to the transport subsystem as airport infrastructure. The work also reflects aspects of the evaluation of the possibilities of algorithmic computational mechanisms to improve the tools of public administration transport subsystems.
Directory of Open Access Journals (Sweden)
Borsukiewicz-Gozdur Aleksandra
2007-01-01
Full Text Available In the work presented are the results of investigations regarding the effectiveness of operation of power plant fed by geothermal water with the flow rate of 100, 150, and 200 m3/h and temperatures of 70, 80, and 90 °C, i. e. geothermal water with the parameters available in some towns of West Pomeranian region as well as in Stargard Szczecinski (86.4 °C, Poland. The results of calculations regard the system of geothermal power plant with possibility of utilization of heat for technological purposes. Analyzed are possibilities of application of different working fluids with respect to the most efficient utilization of geothermal energy. .
Pioz, Maryline; Guis, Hélène; Calavas, Didier; Durand, Benoît; Abrial, David; Ducrot, Christian
2011-04-20
Understanding the spatial dynamics of an infectious disease is critical when attempting to predict where and how fast the disease will spread. We illustrate an approach using a trend-surface analysis (TSA) model combined with a spatial error simultaneous autoregressive model (SAR(err) model) to estimate the speed of diffusion of bluetongue (BT), an infectious disease of ruminants caused by bluetongue virus (BTV) and transmitted by Culicoides. In a first step to gain further insight into the spatial transmission characteristics of BTV serotype 8, we used 2007-2008 clinical case reports in France and TSA modelling to identify the major directions and speed of disease diffusion. We accounted for spatial autocorrelation by combining TSA with a SAR(err) model, which led to a trend SAR(err) model. Overall, BT spread from north-eastern to south-western France. The average trend SAR(err)-estimated velocity across the country was 5.6 km/day. However, velocities differed between areas and time periods, varying between 2.1 and 9.3 km/day. For more than 83% of the contaminated municipalities, the trend SAR(err)-estimated velocity was less than 7 km/day. Our study was a first step in describing the diffusion process for BT in France. To our knowledge, it is the first to show that BT spread in France was primarily local and consistent with the active flight of Culicoides and local movements of farm animals. Models such as the trend SAR(err) models are powerful tools to provide information on direction and speed of disease diffusion when the only data available are date and location of cases.
Liénard, Jean; Lynn, Kendra; Strigul, Nikolay; Norris, Benjamin K.; Gatziolis, Demetrios; Mullarney, Julia C.; Bryan, Karin, R.; Henderson, Stephen M.
2016-09-01
Aquatic vegetation can shelter coastlines from energetic waves and tidal currents, sometimes enabling accretion of fine sediments. Simulation of flow and sediment transport within submerged canopies requires quantification of vegetation geometry. However, field surveys used to determine vegetation geometry can be limited by the time required to obtain conventional caliper and ruler measurements. Building on recent progress in photogrammetry and computer vision, we present a method for reconstructing three-dimensional canopy geometry. The method was used to survey a dense canopy of aerial mangrove roots, called pneumatophores, in Vietnam's Mekong River Delta. Photogrammetric estimation of geometry required 1) taking numerous photographs at low tide from multiple viewpoints around 1 m2 quadrats, 2) computing relative camera locations and orientations by triangulation of key features present in multiple images and reconstructing a dense 3D point cloud, and 3) extracting pneumatophore locations and diameters from the point cloud data. Step 3) was accomplished by a new 'sector-slice' algorithm, yielding geometric parameters every 5 mm along a vertical profile. Photogrammetric analysis was compared with manual caliper measurements. In all 5 quadrats considered, agreement was found between manual and photogrammetric estimates of stem number, and of number × mean diameter, which is a key parameter appearing in hydrodynamic models. In two quadrats, pneumatophores were encrusted with numerous barnacles, generating a complex geometry not resolved by hand measurements. In remaining cases, moderate agreement between manual and photogrammetric estimates of stem diameter and solid volume fraction was found. By substantially reducing measurement time in the field while capturing in greater detail the 3D structure, photogrammetry has potential to improve input to hydrodynamic models, particularly for simulations of flow through large-scale, heterogenous canopies.
International Nuclear Information System (INIS)
Boyd, Gale A.
2014-01-01
This paper describes the EPA's voluntary ENERGY STAR program and the results of the automobile manufacturing industry's efforts to advance energy management as measured by the updated ENERGY STAR Energy Performance Indicator (EPI). A stochastic single-factor input frontier estimation using the gamma error distribution is applied to separately estimate the distribution of the electricity and fossil fuel efficiency of assembly plants using data from 2003 to 2005 and then compared to model results from a prior analysis conducted for the 1997–2000 time period. This comparison provides an assessment of how the industry has changed over time. The frontier analysis shows a modest improvement (reduction) in “best practice” for electricity use and a larger one for fossil fuels. This is accompanied by a large reduction in the variance of fossil fuel efficiency distribution. The results provide evidence of a shift in the frontier, in addition to some “catching up” of poor performing plants over time. - Highlights: • A non-public dataset of U.S. auto manufacturing plants is compiled. • A stochastic frontier with a gamma distribution is applied to plant level data. • Electricity and fuel use are modeled separately. • Comparison to prior analysis reveals a shift in the frontier and “catching up”. • Results are used by ENERGY STAR to award energy efficiency plant certifications
Estimating the efficiency from Brazilian banks: a bootstrapped Data Envelopment Analysis (DEA
Directory of Open Access Journals (Sweden)
Ana Elisa Périco
2016-01-01
Full Text Available Abstract The Brazilian banking sector went through several changes in its structure over the past few years. Such changes are related to fusions and acquisitions, as well as the largest market opening to foreign banks. The objective of this paper is to analyze, by applying the bootstrap DEA, the efficiency of banks in Brazil in 2010-2013. The methodology was applied to 30 largest banking organizations in a financial intermediation approach. In that model, the resources entering a bank in the form of deposits and total assets are classified as inputs and besides these manual labor is also considered as a resource capable of generating results. For the output variable, credit operations represent the most appropriate alternative, considering the role of the bank as a financial intermediary. In this work, the matter of the best classification among retail banks and banks specialized in credit has little relevance. The low relevance in this type of comparison is a result of analysis by segments (segments were analyzed separately. The results presented here point to an average level of efficiency for the large Brazilian banks in the period. This scenario requires efforts to reduce expenses but also to increase revenues.
International Nuclear Information System (INIS)
Bourassa, A.E.; Degenstein, D.A.; Llewellyn, E.J.
2008-01-01
The inversion of satellite-based observations of limb scattered sunlight for the retrieval of constituent species requires an efficient and accurate modelling of the measurement. We present the development of the SASKTRAN radiative transfer model for the prediction of limb scatter measurements at optical wavelengths by method of successive orders along rays traced in a spherical atmosphere. The component of the signal due to the first two scattering events of the solar beam is accounted for directly along rays traced in the three-dimensional geometry. Simplifying assumptions in successive scattering orders provide computational optimizations without severely compromising the accuracy of the solution. SASKTRAN is designed for the analysis of measurements from the OSIRIS instrument and the implementation of the algorithm is efficient such that the code is suitable for the inversion of OSIRIS profiles on desktop computers. SASKTRAN total limb radiance profiles generally compare better with Monte-Carlo reference models over a large range of solar conditions than the approximate spherical and plane-parallel models typically used for inversions
Directory of Open Access Journals (Sweden)
Xianglin Meng
2018-03-01
Full Text Available The normal vector estimation of the large-scale scattered point cloud (LSSPC plays an important role in point-based shape editing. However, the normal vector estimation for LSSPC cannot meet the great challenge of the sharp increase of the point cloud that is mainly attributed to its low computational efficiency. In this paper, a novel, fast method-based on bi-linear interpolation is reported on the normal vector estimation for LSSPC. We divide the point sets into many small cubes to speed up the local point search and construct interpolation nodes on the isosurface expressed by the point cloud. On the premise of calculating the normal vectors of these interpolated nodes, a normal vector bi-linear interpolation of the points in the cube is realized. The proposed approach has the merits of accurate, simple, and high efficiency, because the algorithm only needs to search neighbor and calculates normal vectors for interpolation nodes that are usually far less than the point cloud. The experimental results of several real and simulated point sets show that our method is over three times faster than the Elliptic Gabriel Graph-based method, and the average deviation is less than 0.01 mm.
Unsupervised Learning for Efficient Texture Estimation From Limited Discrete Orientation Data
Niezgoda, Stephen R.; Glover, Jared
2013-11-01
The estimation of orientation distribution functions (ODFs) from discrete orientation data, as produced by electron backscatter diffraction or crystal plasticity micromechanical simulations, is typically achieved via techniques such as the Williams-Imhof-Matthies-Vinel (WIMV) algorithm or generalized spherical harmonic expansions, which were originally developed for computing an ODF from pole figures measured by X-ray or neutron diffraction. These techniques rely on ad-hoc methods for choosing parameters, such as smoothing half-width and bandwidth, and for enforcing positivity constraints and appropriate normalization. In general, such approaches provide little or no information-theoretic guarantees as to their optimality in describing the given dataset. In the current study, an unsupervised learning algorithm is proposed which uses a finite mixture of Bingham distributions for the estimation of ODFs from discrete orientation data. The Bingham distribution is an antipodally-symmetric, max-entropy distribution on the unit quaternion hypersphere. The proposed algorithm also introduces a minimum message length criterion, a common tool in information theory for balancing data likelihood with model complexity, to determine the number of components in the Bingham mixture. This criterion leads to ODFs which are less likely to overfit (or underfit) the data, eliminating the need for a priori parameter choices.
Efficiency estimation method of three-wired AC to DC line transfer
Solovev, S. V.; Bardanov, A. I.
2018-05-01
The development of power semiconductor converters technology expands the scope of their application to medium voltage distribution networks (6-35 kV). Particularly rectifiers and inverters of appropriate power capacity complement the topology of such voltage level networks with the DC links and lines. The article presents a coefficient that allows taking into account the increase of transmission line capacity depending on the parameters of it. The application of the coefficient is presented by the example of transfer three-wired AC line to DC in various methods. Dependences of the change in the capacity from the load power factor of the line and the reactive component of the resistance of the transmission line are obtained. Conclusions are drawn about the most efficient ways of converting a three-wired AC line to direct current.
DEFF Research Database (Denmark)
Løhndorf, Petar Durdevic; Pedersen, Simon; Yang, Zhenyu
2016-01-01
Offshore oil and gas industry has been active in the North Sea for more than half a century, contributing to the economy and facilitating a low oil import rate in the producing countries. The peak production was reached in the early 2000s, and since then the oil production has been decreasing while...... to reach the desired oil production capacity, consequently the discharged amount of oil increases.This leads to oceanic pollution, which has been linked to various negative effects in the marine life. The current legislation requires a maximum oil discharge of 30 parts per million (PPM). The oil in water...... a novel control technology which is based on online and dynamic OiW measurements. This article evaluates some currently available on- line measuring technologies to measure OiW, and the possibility to use these techniques for hydrocyclone efficiency evaluation, model development and as a feedback...
Mora Cordova, Angel
2018-05-22
Carbon nanotubes (CNTs) have been widely used to improve the electrical conductivity of polymers. However, not all CNTs actively participate in the conduction of electricity since they have to be close to each other to form a conductive network. The amount of active CNTs is rarely discussed as it is not captured by percolation theory. However, this amount is a very important information that could be used in a definition of loading efficiency for CNTs (and, in general, for any nanofiller). Thus, we develop a computational tool to quantify the amount of CNTs that actively participates in the conductive network. We then use this quantity to propose a definition of loading efficiency. We compare our results with an expression presented in the literature for the fraction of percolated CNTs (although not presented as a definition of efficiency). We found that this expression underestimates the fraction of percolated CNTs. We thus propose an improved estimation. We also study how efficiency changes with CNT loading and the CNT aspect ratio. We use this concept to study the size of the representative volume element (RVE) for polymers loaded with CNTs, which has received little attention in the past. Here, we find the size of RVE based on both loading efficiency and electrical conductivity such that the scales of “morphological” and “functional” RVEs can be compared. Additionally, we study the relations between particle and network properties (such as efficiency, CNT conductivity and junction resistance) and the conductivity of CNT/polymer composites. We present a series of recommendations to improve the conductivity of a composite based on our simulation results.
Mora Cordova, Angel; Han, Fei; Lubineau, Gilles
2018-01-01
Carbon nanotubes (CNTs) have been widely used to improve the electrical conductivity of polymers. However, not all CNTs actively participate in the conduction of electricity since they have to be close to each other to form a conductive network. The amount of active CNTs is rarely discussed as it is not captured by percolation theory. However, this amount is a very important information that could be used in a definition of loading efficiency for CNTs (and, in general, for any nanofiller). Thus, we develop a computational tool to quantify the amount of CNTs that actively participates in the conductive network. We then use this quantity to propose a definition of loading efficiency. We compare our results with an expression presented in the literature for the fraction of percolated CNTs (although not presented as a definition of efficiency). We found that this expression underestimates the fraction of percolated CNTs. We thus propose an improved estimation. We also study how efficiency changes with CNT loading and the CNT aspect ratio. We use this concept to study the size of the representative volume element (RVE) for polymers loaded with CNTs, which has received little attention in the past. Here, we find the size of RVE based on both loading efficiency and electrical conductivity such that the scales of “morphological” and “functional” RVEs can be compared. Additionally, we study the relations between particle and network properties (such as efficiency, CNT conductivity and junction resistance) and the conductivity of CNT/polymer composites. We present a series of recommendations to improve the conductivity of a composite based on our simulation results.
Energy Technology Data Exchange (ETDEWEB)
Ng, J.C.; Hertle, A.; Seawright, A.A. [Queensland Univ., Brisbane (Australia). National Research Centre for Environmental Toxicology; Mcdougall, K.W. [Wollongbar Agricultural Institute (Australia)
1997-12-31
With the view of estimating bioavailability of metals from contaminated sites and risk assessment, a rat model is used for a comparative bioavailability test in which groups of rats were given via the oral route a slurry of arsenic contaminated soils, a solution of sodium arsenate or sodium arsenite, or calcium arsenite spiked wheat flour. Blood samples are collected 96 hours after dosing for the arsenic determination. The comparative bioavailability (CBA) is calculated from the ratio of arsenic results obtained from the soil group and arsenic control group dosed with sodium arsenate or arsenite. CBA results show a good correlation with 0.5 M HCl and 1.0 M HCl acid leachates. The rat model process to be a sensitive indicator using the blood for the study of bioavailability of arsenic in soils
Rajabi, Mohammad Mahdi; Ataie-Ashtiani, Behzad
2016-05-01
Bayesian inference has traditionally been conceived as the proper framework for the formal incorporation of expert knowledge in parameter estimation of groundwater models. However, conventional Bayesian inference is incapable of taking into account the imprecision essentially embedded in expert provided information. In order to solve this problem, a number of extensions to conventional Bayesian inference have been introduced in recent years. One of these extensions is 'fuzzy Bayesian inference' which is the result of integrating fuzzy techniques into Bayesian statistics. Fuzzy Bayesian inference has a number of desirable features which makes it an attractive approach for incorporating expert knowledge in the parameter estimation process of groundwater models: (1) it is well adapted to the nature of expert provided information, (2) it allows to distinguishably model both uncertainty and imprecision, and (3) it presents a framework for fusing expert provided information regarding the various inputs of the Bayesian inference algorithm. However an important obstacle in employing fuzzy Bayesian inference in groundwater numerical modeling applications is the computational burden, as the required number of numerical model simulations often becomes extremely exhaustive and often computationally infeasible. In this paper, a novel approach of accelerating the fuzzy Bayesian inference algorithm is proposed which is based on using approximate posterior distributions derived from surrogate modeling, as a screening tool in the computations. The proposed approach is first applied to a synthetic test case of seawater intrusion (SWI) in a coastal aquifer. It is shown that for this synthetic test case, the proposed approach decreases the number of required numerical simulations by an order of magnitude. Then the proposed approach is applied to a real-world test case involving three-dimensional numerical modeling of SWI in Kish Island, located in the Persian Gulf. An expert
DEFF Research Database (Denmark)
Jensen, Jørgen Juncher
2007-01-01
In on-board decision support systems efficient procedures are needed for real-time estimation of the maximum ship responses to be expected within the next few hours, given on-line information on the sea state and user defined ranges of possible headings and speeds. For linear responses standard...... frequency domain methods can be applied. To non-linear responses like the roll motion, standard methods like direct time domain simulations are not feasible due to the required computational time. However, the statistical distribution of non-linear ship responses can be estimated very accurately using...... the first-order reliability method (FORM), well-known from structural reliability problems. To illustrate the proposed procedure, the roll motion is modelled by a simplified non-linear procedure taking into account non-linear hydrodynamic damping, time-varying restoring and wave excitation moments...
DEFF Research Database (Denmark)
Mühlfeld, Christian; Papadakis, Tamara; Krasteva, Gabriela
2010-01-01
Quantitative information about the innervation is essential to analyze the structure-function relationships of organs. So far, there has been no unbiased stereological tool for this purpose. This study presents a new unbiased and efficient method to quantify the total length of axons in a given...... reference volume, illustrated on the left ventricle of the mouse heart. The method is based on the following steps: 1) estimation of the reference volume; 2) randomization of location and orientation using appropriate sampling techniques; 3) counting of nerve fiber profiles hit by a defined test area within...
Liu, Jingxia; Colditz, Graham A
2018-05-01
There is growing interest in conducting cluster randomized trials (CRTs). For simplicity in sample size calculation, the cluster sizes are assumed to be identical across all clusters. However, equal cluster sizes are not guaranteed in practice. Therefore, the relative efficiency (RE) of unequal versus equal cluster sizes has been investigated when testing the treatment effect. One of the most important approaches to analyze a set of correlated data is the generalized estimating equation (GEE) proposed by Liang and Zeger, in which the "working correlation structure" is introduced and the association pattern depends on a vector of association parameters denoted by ρ. In this paper, we utilize GEE models to test the treatment effect in a two-group comparison for continuous, binary, or count data in CRTs. The variances of the estimator of the treatment effect are derived for the different types of outcome. RE is defined as the ratio of variance of the estimator of the treatment effect for equal to unequal cluster sizes. We discuss a commonly used structure in CRTs-exchangeable, and derive the simpler formula of RE with continuous, binary, and count outcomes. Finally, REs are investigated for several scenarios of cluster size distributions through simulation studies. We propose an adjusted sample size due to efficiency loss. Additionally, we also propose an optimal sample size estimation based on the GEE models under a fixed budget for known and unknown association parameter (ρ) in the working correlation structure within the cluster. © 2018 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Directory of Open Access Journals (Sweden)
Müller Kai F
2005-10-01
Full Text Available Abstract Background For parsimony analyses, the most common way to estimate confidence is by resampling plans (nonparametric bootstrap, jackknife, and Bremer support (Decay indices. The recent literature reveals that parameter settings that are quite commonly employed are not those that are recommended by theoretical considerations and by previous empirical studies. The optimal search strategy to be applied during resampling was previously addressed solely via standard search strategies available in PAUP*. The question of a compromise between search extensiveness and improved support accuracy for Bremer support received even less attention. A set of experiments was conducted on different datasets to find an empirical cut-off point at which increased search extensiveness does not significantly change Bremer support and jackknife or bootstrap proportions any more. Results For the number of replicates needed for accurate estimates of support in resampling plans, a diagram is provided that helps to address the question whether apparently different support values really differ significantly. It is shown that the use of random addition cycles and parsimony ratchet iterations during bootstrapping does not translate into higher support, nor does any extension of the search extensiveness beyond the rather moderate effort of TBR (tree bisection and reconnection branch swapping plus saving one tree per replicate. Instead, in case of very large matrices, saving more than one shortest tree per iteration and using a strict consensus tree of these yields decreased support compared to saving only one tree. This can be interpreted as a small risk of overestimating support but should be more than compensated by other factors that counteract an enhanced type I error. With regard to Bremer support, a rule of thumb can be derived stating that not much is gained relative to the surplus computational effort when searches are extended beyond 20 ratchet iterations per
Efficient synthesis of tension modulation in strings and membranes based on energy estimation.
Avanzini, Federico; Marogna, Riccardo; Bank, Balázs
2012-01-01
String and membrane vibrations cannot be considered as linear above a certain amplitude due to the variation in string or membrane tension. A relevant special case is when the tension is spatially constant and varies in time only in dependence of the overall string length or membrane surface. The most apparent perceptual effect of this tension modulation phenomenon is the exponential decay of pitch in time. Pitch glides due to tension modulation are an important timbral characteristic of several musical instruments, including the electric guitar and tom-tom drum, and many ethnic instruments. This paper presents a unified formulation to the tension modulation problem for one-dimensional (1-D) (string) and two-dimensional (2-D) (membrane) cases. In addition, it shows that the short-time average of the tension variation, which is responsible for pitch glides, is approximately proportional to the system energy. This proportionality allows the efficient physics-based sound synthesis of pitch glides. The proposed models require only slightly more computational resources than linear models as opposed to earlier tension-modulated models of higher complexity. © 2012 Acoustical Society of America.
A rapid, sensitive, and cost-efficient assay to estimate viability of potato cyst nematodes.
van den Elsen, Sven; Ave, Maaike; Schoenmakers, Niels; Landeweert, Renske; Bakker, Jaap; Helder, Johannes
2012-02-01
Potato cyst nematodes (PCNs) are quarantine organisms, and they belong to the economically most relevant pathogens of potato worldwide. Methodologies to assess the viability of their cysts, which can contain 200 to 500 eggs protected by the hardened cuticle of a dead female, are either time and labor intensive or lack robustness. We present a robust and cost-efficient viability assay based on loss of membrane integrity upon death. This assay uses trehalose, a disaccharide present at a high concentration in the perivitelline fluid of PCN eggs, as a viability marker. Although this assay can detect a single viable egg, the limit of detection for regular field samples was higher, ≈10 viable eggs, due to background signals produced by other soil components. On the basis of 30 nonviable PCN samples from The Netherlands, a threshold level was defined (ΔA(trehalose) = 0.0094) below which the presence of >10 viable eggs is highly unlikely (true for ≈99.7% of the observations). This assay can easily be combined with a subsequent DNA-based species determination. The presence of trehalose is a general phenomenon among cyst nematodes; therefore, this method can probably be used for (for example) soybean, sugar beet, and cereal cyst nematodes as well.
Aleksandrov, V. I.; Vasilyeva, M. A.; Pomeranets, I. B.
2017-10-01
The paper presents analytical calculations of specific pressure loss in hydraulic transport of the Kachkanarsky GOK iron ore processing tailing slurry. The calculations are based on the results of the experimental studies on specific pressure loss dependence upon hydraulic roughness of pipelines internal surface lined with polyurethane coating. The experiments proved that hydraulic roughness of polyurethane coating is by the factor of four smaller than that of steel pipelines, resulting in a decrease of hydraulic resistance coefficients entered into calculating formula of specific pressure loss - the Darcy-Weisbach formula. Relative and equivalent roughness coefficients are calculated for pipelines with polyurethane coating and without it. Comparative calculations show that hydrotransport pipelines polyurethane coating application is conductive to a specific energy consumption decrease in hydraulic transport of the Kachkanarsky GOC iron ore processing tailings slurry by the factor of 1.5. The experiments were performed on a laboratory hydraulic test rig with a view to estimate the character and rate of physical roughness change in pipe samples with polyurethane coating. The experiments showed that during the following 484 hours of operation, roughness changed in all pipe samples inappreciably. As a result of processing of the experimental data by the mathematical statistics methods, an empirical formula was obtained for the calculation of operating roughness of polyurethane coating surface, depending on the pipeline operating duration with iron ore processing tailings slurry.
A simplified model of natural and mechanical removal to estimate cleanup equipment efficiency
International Nuclear Information System (INIS)
Lehr, W.
2001-01-01
Oil spill response organizations rely on modelling to make decisions in offshore response operations. Models are used to test different cleanup strategies and to measure the expected cost of cleanup and the reduction in environmental impact. The oil spill response community has traditionally used the concept of worst case scenario in developing contingency plans for spill response. However, there are many drawbacks to this approach. The Hazardous Materials Response Division of the National Oceanic and Atmospheric Administration in Cooperation with the U.S. Navy Supervisor of Salvage and Diving has developed a Trajectory Analysis Planner (TAP) which will give planners the tool to try out different cleanup strategies and equipment configurations based upon historical wind and current conditions instead of worst-case scenarios. The spill trajectory model is a classic example in oil spill modelling that uses advanced non-linear three-dimensional hydrodynamical sub-models to estimate surface currents under conditions where oceanographic initial conditions are not accurately known and forecasts of wind stress are unreliable. In order to get better answers, it is often necessary to refine input values rather than increasing the sophistication of the hydrodynamics. This paper described another spill example where the level of complexity of the algorithms needs to be evaluated with regard to the reliability of the input, the sensitivity of the answers to input and model parameters, and the comparative reliability of other algorithms in the model. 9 refs., 1 fig
On the Usage of GPUs for Efficient Motion Estimation in Medical Image Sequences
Directory of Open Access Journals (Sweden)
Jeyarajan Thiyagalingam
2011-01-01
Full Text Available Images are ubiquitous in biomedical applications from basic research to clinical practice. With the rapid increase in resolution, dimensionality of the images and the need for real-time performance in many applications, computational requirements demand proper exploitation of multicore architectures. Towards this, GPU-specific implementations of image analysis algorithms are particularly promising. In this paper, we investigate the mapping of an enhanced motion estimation algorithm to novel GPU-specific architectures, the resulting challenges and benefits therein. Using a database of three-dimensional image sequences, we show that the mapping leads to substantial performance gains, up to a factor of 60, and can provide near-real-time experience. We also show how architectural peculiarities of these devices can be best exploited in the benefit of algorithms, most specifically for addressing the challenges related to their access patterns and different memory configurations. Finally, we evaluate the performance of the algorithm on three different GPU architectures and perform a comprehensive analysis of the results.
El Gharamti, Mohamad
2012-04-01
Accurate knowledge of the movement of contaminants in porous media is essential to track their trajectory and later extract them from the aquifer. A two-dimensional flow model is implemented and then applied on a linear contaminant transport model in the same porous medium. Because of different sources of uncertainties, this coupled model might not be able to accurately track the contaminant state. Incorporating observations through the process of data assimilation can guide the model toward the true trajectory of the system. The Kalman filter (KF), or its nonlinear invariants, can be used to tackle this problem. To overcome the prohibitive computational cost of the KF, the singular evolutive Kalman filter (SEKF) and the singular fixed Kalman filter (SFKF) are used, which are variants of the KF operating with low-rank covariance matrices. Experimental results suggest that under perfect and imperfect model setups, the low-rank filters can provide estimates as accurate as the full KF but at much lower computational effort. Low-rank filters are demonstrated to significantly reduce the computational effort of the KF to almost 3%. © 2012 American Society of Civil Engineers.
Chen, Siyuan; Epps, Julien
2014-12-01
Monitoring pupil and blink dynamics has applications in cognitive load measurement during human-machine interaction. However, accurate, efficient, and robust pupil size and blink estimation pose significant challenges to the efficacy of real-time applications due to the variability of eye images, hence to date, require manual intervention for fine tuning of parameters. In this paper, a novel self-tuning threshold method, which is applicable to any infrared-illuminated eye images without a tuning parameter, is proposed for segmenting the pupil from the background images recorded by a low cost webcam placed near the eye. A convex hull and a dual-ellipse fitting method are also proposed to select pupil boundary points and to detect the eyelid occlusion state. Experimental results on a realistic video dataset show that the measurement accuracy using the proposed methods is higher than that of widely used manually tuned parameter methods or fixed parameter methods. Importantly, it demonstrates convenience and robustness for an accurate and fast estimate of eye activity in the presence of variations due to different users, task types, load, and environments. Cognitive load measurement in human-machine interaction can benefit from this computationally efficient implementation without requiring a threshold calibration beforehand. Thus, one can envisage a mini IR camera embedded in a lightweight glasses frame, like Google Glass, for convenient applications of real-time adaptive aiding and task management in the future.
Directory of Open Access Journals (Sweden)
Andrew E. Suyker
2013-11-01
Full Text Available Remote sensing techniques that provide synoptic and repetitive observations over large geographic areas have become increasingly important in studying the role of agriculture in global carbon cycles. However, it is still challenging to model crop yields based on remotely sensed data due to the variation in radiation use efficiency (RUE across crop types and the effects of spatial heterogeneity. In this paper, we propose a production efficiency model-based method to estimate corn and soybean yields with MODerate Resolution Imaging Spectroradiometer (MODIS data by explicitly handling the following two issues: (1 field-measured RUE values for corn and soybean are applied to relatively pure pixels instead of the biome-wide RUE value prescribed in the MODIS vegetation productivity product (MOD17; and (2 contributions to productivity from vegetation other than crops in mixed pixels are deducted at the level of MODIS resolution. Our estimated yields statistically correlate with the national survey data for rainfed counties in the Midwestern US with low errors for both corn (R2 = 0.77; RMSE = 0.89 MT/ha and soybeans (R2 = 0.66; RMSE = 0.38 MT/ha. Because the proposed algorithm does not require any retrospective analysis that constructs empirical relationships between the reported yields and remotely sensed data, it could monitor crop yields over large areas.
International Nuclear Information System (INIS)
Tanaka, Yohei; Momma, Akihiko; Kato, Ken; Negishi, Akira; Takano, Kiyonami; Nozaki, Ken; Kato, Tohru
2009-01-01
Uncertainty of electrical efficiency measurement was investigated for a 10 kW-class SOFC system using town gas. Uncertainty of heating value measured by the gas chromatography method on a mole base was estimated as ±0.12% at 95% level of confidence. Micro-gas chromatography with/without CH 4 quantification may be able to reduce uncertainty of measurement. Calibration and uncertainty estimation methods are proposed for flow-rate measurement of town gas with thermal mass-flow meters or controllers. By adequate calibrations for flowmeters, flow rate of town gas or natural gas at 35 standard litters per minute can be measured within relative uncertainty ±1.0% at 95 % level of confidence. Uncertainty of power measurement can be as low as ±0.14% when a precise wattmeter is used and calibrated properly. It is clarified that electrical efficiency for non-pressurized 10 kW-class SOFC systems can be measured within ±1.0% relative uncertainty at 95% level of confidence with the developed techniques when the SOFC systems are operated relatively stably
Estimation of Efficiency of the Cooling Channel of the Nozzle Blade of Gas-Turbine Engines
Vikulin, A. V.; Yaroslavtsev, N. L.; Zemlyanaya, V. A.
2018-02-01
The main direction of improvement of gas-turbine plants (GTP) and gas-turbine engines (GTE) is increasing the gas temperature at the turbine inlet. For the solution of this problem, promising systems of intensification of heat exchange in cooled turbine blades are developed. With this purpose, studies of the efficiency of the cooling channel of the nozzle blade in the basic modification and of the channel after constructive measures for improvement of the cooling system by the method of calorimetry in a liquid-metal thermostat were conducted. The combined system of heat-exchange intensification with the complicated scheme of branched channels is developed; it consists of a vortex matrix and three rows of inclined intermittent trip strips. The maximum value of hydraulic resistance ξ is observed at the first row of the trip strips, which is connected with the effect of dynamic impact of airflow on the channel walls, its turbulence, and rotation by 117° at the inlet to the channels formed by the trip strips. These factors explain the high value of hydraulic resistance equal to 3.7-3.4 for the first row of the trip strips. The obtained effect was also confirmed by the results of thermal tests, i.e., the unevenness of heat transfer on the back and on the trough of the blade is observed at the first row of the trip strips, which amounts 8-12%. This unevenness has a fading character; at the second row of the trip strips, it amounts to 3-7%, and it is almost absent at the third row. At the area of vortex matrix, the intensity of heat exchange on the blade back is higher as compared to the trough, which is explained by the different height of the matrix ribs on its opposite sides. The design changes in the nozzle blade of basic modification made it possible to increase the intensity of heat exchange by 20-50% in the area of the vortex matrix and by 15-30% on the section of inclined intermittent trip strips. As a result of research, new criteria dependences for the
Energy Technology Data Exchange (ETDEWEB)
Mintairov, M. A., E-mail: mamint@mail.ioffe.ru; Evstropov, V. V.; Mintairov, S. A.; Shvarts, M. Z.; Timoshina, N. Kh.; Kalyuzhnyy, N. A. [Russian Academy of Sciences, Ioffe Physical-Technical Institute (Russian Federation)
2015-05-15
A method is proposed for estimating the potential efficiency which can be achieved in an initially unbalanced multijunction solar cell by the mutual convergence of photogenerated currents: to extract this current from a relatively narrow band-gap cell and to add it to a relatively wide-gap cell. It is already known that the properties facilitating relative convergence are inherent to such objects as bound excitons, quantum dots, donor-acceptor pairs, and others located in relatively wide-gap cells. In fact, the proposed method is reduced to the problem of obtaining such a required light current-voltage (I–V) characteristic which corresponds to the equality of all photogenerated short-circuit currents. Two methods for obtaining the required light I–V characteristic are used. The first one is selection of the spectral composition of the radiation incident on the multijunction solar cell from an illuminator. The second method is a double shift of the dark I–V characteristic: a current shift J{sub g} (common set photogenerated current) and a voltage shift (−J{sub g}R{sub s}), where R{sub s} is the series resistance. For the light and dark I–V characteristics, a general analytical expression is derived, which considers the effect of so-called luminescence coupling in multijunction solar cells. The experimental I–V characteristics are compared with the calculated ones for a three-junction InGaP/GaAs/Ge solar cell with R{sub s} = 0.019 Ω cm{sup 2} and a maximum factual efficiency of 36.9%. Its maximum potential efficiency is estimated as 41.2%.
Directory of Open Access Journals (Sweden)
Imran Khan
2018-01-01
Full Text Available Uplink and Downlink channel estimation in massive Multiple Input Multiple Output (MIMO systems is an intricate issue because of the increasing channel matrix dimensions. The channel feedback overhead using traditional codebook schemes is very large, which consumes more bandwidth and decreases the overall system efficiency. The purpose of this paper is to decrease the channel estimation overhead by taking the advantage of sparse attributes and also to optimize the Energy Efficiency (EE of the system. To cope with this issue, we propose a novel approach by using Compressed-Sensing (CS, Block Iterative-Support-Detection (Block-ISD, Angle-of-Departure (AoD and Structured Compressive Sampling Matching Pursuit (S-CoSaMP algorithms to reduce the channel estimation overhead and compare them with the traditional algorithms. The CS uses temporal-correlation of time-varying channels to produce Differential-Channel Impulse Response (DCIR among two CIRs that are adjacent in time-slots. DCIR has greater sparsity than the conventional CIRs as it can be easily compressed. The Block-ISD uses spatial-correlation of the channels to obtain the block-sparsity which results in lower pilot-overhead. AoD quantizes the channels whose path-AoDs variation is slower than path-gains and such information is utilized for reducing the overhead. S-CoSaMP deploys structured-sparsity to obtain reliable Channel-State-Information (CSI. MATLAB simulation results show that the proposed CS based algorithms reduce the feedback and pilot-overhead by a significant percentage and also improve the system capacity as compared with the traditional algorithms. Moreover, the EE level increases with increasing Base Station (BS density, UE density and lowering hardware impairments level.
Wang, Yikai; Kang, Jian; Kemmer, Phebe B; Guo, Ying
2016-01-01
Currently, network-oriented analysis of fMRI data has become an important tool for understanding brain organization and brain networks. Among the range of network modeling methods, partial correlation has shown great promises in accurately detecting true brain network connections. However, the application of partial correlation in investigating brain connectivity, especially in large-scale brain networks, has been limited so far due to the technical challenges in its estimation. In this paper, we propose an efficient and reliable statistical method for estimating partial correlation in large-scale brain network modeling. Our method derives partial correlation based on the precision matrix estimated via Constrained L1-minimization Approach (CLIME), which is a recently developed statistical method that is more efficient and demonstrates better performance than the existing methods. To help select an appropriate tuning parameter for sparsity control in the network estimation, we propose a new Dens-based selection method that provides a more informative and flexible tool to allow the users to select the tuning parameter based on the desired sparsity level. Another appealing feature of the Dens-based method is that it is much faster than the existing methods, which provides an important advantage in neuroimaging applications. Simulation studies show that the Dens-based method demonstrates comparable or better performance with respect to the existing methods in network estimation. We applied the proposed partial correlation method to investigate resting state functional connectivity using rs-fMRI data from the Philadelphia Neurodevelopmental Cohort (PNC) study. Our results show that partial correlation analysis removed considerable between-module marginal connections identified by full correlation analysis, suggesting these connections were likely caused by global effects or common connection to other nodes. Based on partial correlation, we find that the most significant
Directory of Open Access Journals (Sweden)
Jaewook Lee
2015-06-01
Full Text Available This paper presents an efficient method for estimating capacity-fade uncertainty in lithium-ion batteries (LIBs in order to integrate them into the battery-management system (BMS of electric vehicles, which requires simple and inexpensive computation for successful application. The study uses the pseudo-two-dimensional (P2D electrochemical model, which simulates the battery state by solving a system of coupled nonlinear partial differential equations (PDEs. The model parameters that are responsible for electrode degradation are identified and estimated, based on battery data obtained from the charge cycles. The Bayesian approach, with parameters estimated by probability distributions, is employed to account for uncertainties arising in the model and battery data. The Markov Chain Monte Carlo (MCMC technique is used to draw samples from the distributions. The complex computations that solve a PDE system for each sample are avoided by employing a polynomial-based metamodel. As a result, the computational cost is reduced from 5.5 h to a few seconds, enabling the integration of the method into the vehicle BMS. Using this approach, the conservative bound of capacity fade can be determined for the vehicle in service, which represents the safety margin reflecting the uncertainty.
El-Serehy, Hamed A; Bahgat, Magdy M; Al-Rasheid, Khaled; Al-Misned, Fahad; Mortuza, Golam; Shafik, Hesham
2014-07-01
Interest has increased over the last several years in using different methods for treating sewage. The rapid population growth in developing countries (Egypt, for example, with a population of more than 87 millions) has created significant sewage disposal problems. There is therefore a growing need for sewage treatment solutions with low energy requirements and using indigenous materials and skills. Gravel Bed Hydroponics (GBH) as a constructed wetland system for sewage treatment has been proved effective for sewage treatment in several Egyptian villages. The system provided an excellent environment for a wide range of species of ciliates (23 species) and these organisms were potentially very useful as biological indicators for various saprobic conditions. Moreover, the ciliates provided excellent means for estimating the efficiency of the system for sewage purification. Results affirmed the ability of this system to produce high quality effluent with sufficient microbial reduction to enable the production of irrigation quality water.
Energy Technology Data Exchange (ETDEWEB)
Kurnik, Charles W [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Violette, Daniel M. [Navigant, Boulder, CO (United States); Rathbun, Pamela [Tetra Tech, Madison, WI (United States)
2017-11-02
This chapter focuses on the methods used to estimate net energy savings in evaluation, measurement, and verification (EM and V) studies for energy efficiency (EE) programs. The chapter provides a definition of net savings, which remains an unsettled topic both within the EE evaluation community and across the broader public policy evaluation community, particularly in the context of attribution of savings to a program. The chapter differs from the measure-specific Uniform Methods Project (UMP) chapters in both its approach and work product. Unlike other UMP resources that provide recommended protocols for determining gross energy savings, this chapter describes and compares the current industry practices for determining net energy savings but does not prescribe methods.
Directory of Open Access Journals (Sweden)
Wiktor Jakowluk
2014-11-01
Full Text Available System identification, in practice, is carried out by perturbing processes or plants under operation. That is why in many industrial applications a plant-friendly input signal would be preferred for system identification. The goal of the study is to design the optimal input signal which is then employed in the identification experiment and to examine the relationships between the index of friendliness of this input signal and the accuracy of parameter estimation when the measured output signal is significantly affected by noise. In this case, the objective function was formulated through maximisation of the Fisher information matrix determinant (D-optimality expressed in conventional Bolza form. As setting such conditions of the identification experiment we can only talk about the D-suboptimality, we quantify the plant trajectories using the D-efficiency measure. An additional constraint, imposed on D-efficiency of the solution, should allow one to attain the most adequate information content from the plant which operating point is perturbed in the least invasive (most friendly way. A simple numerical example, which clearly demonstrates the idea presented in the paper, is included and discussed.
Directory of Open Access Journals (Sweden)
Jonathon Taylor
2018-05-01
Full Text Available Mortality rates rise during hot weather in England, and projected future increases in heatwave frequency and intensity require the development of heat protection measures such as the adaptation of housing to reduce indoor overheating. We apply a combined building physics and health model to dwellings in the West Midlands, UK, using an English Housing Survey (EHS-derived stock model. Regional temperature exposures, heat-related mortality risk, and space heating energy consumption were estimated for 2030s, 2050s, and 2080s medium emissions climates prior to and following heat mitigating, energy-efficiency, and occupant behaviour adaptations. Risk variation across adaptations, dwellings, and occupant types were assessed. Indoor temperatures were greatest in converted flats, while heat mortality rates were highest in bungalows due to the occupant age profiles. Full energy efficiency retrofit reduced regional domestic space heating energy use by 26% but increased summertime heat mortality 3–4%, while reduced façade absorptance decreased heat mortality 12–15% but increased energy consumption by 4%. External shutters provided the largest reduction in heat mortality (37–43%, while closed windows caused a large increase in risk (29–64%. Ensuring adequate post-retrofit ventilation, targeted installation of shutters, and ensuring operable windows in dwellings with heat-vulnerable occupants may save energy and significantly reduce heat-related mortality.
El Gharamti, Mohamad; Valstar, Johan R.; Hoteit, Ibrahim
2014-01-01
Reactive contaminant transport models are used by hydrologists to simulate and study the migration and fate of industrial waste in subsurface aquifers. Accurate transport modeling of such waste requires clear understanding of the system's parameters, such as sorption and biodegradation. In this study, we present an efficient sequential data assimilation scheme that computes accurate estimates of aquifer contamination and spatially variable sorption coefficients. This assimilation scheme is based on a hybrid formulation of the ensemble Kalman filter (EnKF) and optimal interpolation (OI) in which solute concentration measurements are assimilated via a recursive dual estimation of sorption coefficients and contaminant state variables. This hybrid EnKF-OI scheme is used to mitigate background covariance limitations due to ensemble under-sampling and neglected model errors. Numerical experiments are conducted with a two-dimensional synthetic aquifer in which cobalt-60, a radioactive contaminant, is leached in a saturated heterogeneous clayey sandstone zone. Assimilation experiments are investigated under different settings and sources of model and observational errors. Simulation results demonstrate that the proposed hybrid EnKF-OI scheme successfully recovers both the contaminant and the sorption rate and reduces their uncertainties. Sensitivity analyses also suggest that the adaptive hybrid scheme remains effective with small ensembles, allowing to reduce the ensemble size by up to 80% with respect to the standard EnKF scheme. © 2014 Elsevier Ltd.
El Gharamti, Mohamad
2014-09-01
Reactive contaminant transport models are used by hydrologists to simulate and study the migration and fate of industrial waste in subsurface aquifers. Accurate transport modeling of such waste requires clear understanding of the system\\'s parameters, such as sorption and biodegradation. In this study, we present an efficient sequential data assimilation scheme that computes accurate estimates of aquifer contamination and spatially variable sorption coefficients. This assimilation scheme is based on a hybrid formulation of the ensemble Kalman filter (EnKF) and optimal interpolation (OI) in which solute concentration measurements are assimilated via a recursive dual estimation of sorption coefficients and contaminant state variables. This hybrid EnKF-OI scheme is used to mitigate background covariance limitations due to ensemble under-sampling and neglected model errors. Numerical experiments are conducted with a two-dimensional synthetic aquifer in which cobalt-60, a radioactive contaminant, is leached in a saturated heterogeneous clayey sandstone zone. Assimilation experiments are investigated under different settings and sources of model and observational errors. Simulation results demonstrate that the proposed hybrid EnKF-OI scheme successfully recovers both the contaminant and the sorption rate and reduces their uncertainties. Sensitivity analyses also suggest that the adaptive hybrid scheme remains effective with small ensembles, allowing to reduce the ensemble size by up to 80% with respect to the standard EnKF scheme. © 2014 Elsevier Ltd.
Sadeghifar, Hamidreza
2015-10-01
Developing general methods that rely on column data for the efficiency estimation of operating (existing) distillation columns has been overlooked in the literature. Most of the available methods are based on empirical mass transfer and hydraulic relations correlated to laboratory data. Therefore, these methods may not be sufficiently accurate when applied to industrial columns. In this paper, an applicable and accurate method was developed for the efficiency estimation of distillation columns filled with trays. This method can calculate efficiency as well as mass and heat transfer coefficients without using any empirical mass transfer or hydraulic correlations and without the need to estimate operational or hydraulic parameters of the column. E.g., the method does not need to estimate tray interfacial area, which can be its most important advantage over all the available methods. The method can be used for the efficiency prediction of any trays in distillation columns. For the efficiency calculation, the method employs the column data and uses the true rates of the mass and heat transfers occurring inside the operating column. It is highly emphasized that estimating efficiency of an operating column has to be distinguished from that of a column being designed.
Directory of Open Access Journals (Sweden)
Mohammad Manir Hossain Mollah
Full Text Available Identifying genes that are differentially expressed (DE between two or more conditions with multiple patterns of expression is one of the primary objectives of gene expression data analysis. Several statistical approaches, including one-way analysis of variance (ANOVA, are used to identify DE genes. However, most of these methods provide misleading results for two or more conditions with multiple patterns of expression in the presence of outlying genes. In this paper, an attempt is made to develop a hybrid one-way ANOVA approach that unifies the robustness and efficiency of estimation using the minimum β-divergence method to overcome some problems that arise in the existing robust methods for both small- and large-sample cases with multiple patterns of expression.The proposed method relies on a β-weight function, which produces values between 0 and 1. The β-weight function with β = 0.2 is used as a measure of outlier detection. It assigns smaller weights (≥ 0 to outlying expressions and larger weights (≤ 1 to typical expressions. The distribution of the β-weights is used to calculate the cut-off point, which is compared to the observed β-weight of an expression to determine whether that gene expression is an outlier. This weight function plays a key role in unifying the robustness and efficiency of estimation in one-way ANOVA.Analyses of simulated gene expression profiles revealed that all eight methods (ANOVA, SAM, LIMMA, EBarrays, eLNN, KW, robust BetaEB and proposed perform almost identically for m = 2 conditions in the absence of outliers. However, the robust BetaEB method and the proposed method exhibited considerably better performance than the other six methods in the presence of outliers. In this case, the BetaEB method exhibited slightly better performance than the proposed method for the small-sample cases, but the the proposed method exhibited much better performance than the BetaEB method for both the small- and large
Directory of Open Access Journals (Sweden)
Bazhenov Viktor Ivanovich
2015-09-01
Full Text Available The starting stage of the tender procedures in Russia with the participation of foreign suppliers dictates the feasibility of the developments for economical methods directed to comparison of technical solutions on the construction field. The article describes the example of practical Life Cycle Cost (LCC evaluations under respect of Present Value (PV determination. These create a possibility for investor to estimate long-term projects (indicated as 25 years as commercially profitable, taking into account inflation rate, interest rate, real discount rate (indicated as 5 %. For economic analysis air-blower station of WWTP was selected as a significant energy consumer. Technical variants for the comparison of blower types are: 1 - multistage without control, 2 - multistage with VFD control, 3 - single stage double vane control. The result of LCC estimation shows the last variant as most attractive or cost-effective for investments with economy of 17,2 % (variant 1 and 21,0 % (variant 2 under adopted duty conditions and evaluations of capital costs (Cic + Cin with annual expenditure related (Ce+Co+Cm. The adopted duty conditions include daily and seasonal fluctuations of air flow. This was the reason for the adopted energy consumption as, kW∙h: 2158 (variant 1,1743...2201 (variant 2, 1058...1951 (variant 3. The article refers to Europump guide tables in order to simplify sophisticated factors search (Cp /Cn, df, which can be useful for economical analyses in Russia. Example of evaluations connected with energy-efficient solutions is given, but this reference involves the use of materials for the cases with resource savings, such as all types of fuel. In conclusion follows the assent to use LCC indicator jointly with the method of determining discounted cash flows, that will satisfy the investor’s need for interest source due to technical and economical comparisons.
Directory of Open Access Journals (Sweden)
Andreas Tewes
2018-02-01
Full Text Available Radiation Use Efficiency (RUE defines the productivity with which absorbed photosynthetically active radiation (APAR is converted to plant biomass. Readily used in crop growth models to predict dry matter accumulation, RUE is commonly determined by elaborate static sensor measurements in the field. Different definitions are used, based on total absorbed PAR (RUEtotal or PAR absorbed by the photosynthetically active leaf tissue only (RUEgreen. Previous studies have shown that the fraction of PAR absorbed (fAPAR, which supports the assessment of RUE, can be reliably estimated via remote sensing (RS, but unfortunately at spatial resolutions too coarse for experimental agriculture. UAV-based RS offers the possibility to cover plant reflectance at very high spatial and temporal resolution, possibly covering several experimental plots in little time. We investigated if (a UAV-based low-cost camera imagery allowed estimating RUEs in different experimental plots where maize was cultivated in the growing season of 2016, (b those values were different from the ones previously reported in literature and (c there was a difference between RUEtotal and RUEgreen. We determined fractional cover and canopy reflectance based on the RS imagery. Our study found that RUEtotal ranges between 4.05 and 4.59, and RUEgreen between 4.11 and 4.65. These values are higher than those published in other research articles, but not outside the range of plausibility. The difference between RUEtotal and RUEgreen was minimal, possibly due to prolonged canopy greenness induced by the stay-green trait of the cultivar grown. The procedure presented here makes time-consuming APAR measurements for determining RUE especially in large experiments superfluous.
Madenjian, Charles P.; Rediske, Richard R.; O'Keefe, James P.; David, Solomon R.
2014-01-01
A technique for laboratory estimation of net trophic transfer efficiency (γ) of polychlorinated biphenyl (PCB) congeners to piscivorous fish from their prey is described herein. During a 135-day laboratory experiment, we fed bloater (Coregonus hoyi) that had been caught in Lake Michigan to lake trout (Salvelinus namaycush) kept in eight laboratory tanks. Bloater is a natural prey for lake trout. In four of the tanks, a relatively high flow rate was used to ensure relatively high activity by the lake trout, whereas a low flow rate was used in the other four tanks, allowing for low lake trout activity. On a tank-by-tank basis, the amount of food eaten by the lake trout on each day of the experiment was recorded. Each lake trout was weighed at the start and end of the experiment. Four to nine lake trout from each of the eight tanks were sacrificed at the start of the experiment, and all 10 lake trout remaining in each of the tanks were euthanized at the end of the experiment. We determined concentrations of 75 PCB congeners in the lake trout at the start of the experiment, in the lake trout at the end of the experiment, and in bloaters fed to the lake trout during the experiment. Based on these measurements, γ was calculated for each of 75 PCB congeners in each of the eight tanks. Mean γ was calculated for each of the 75 PCB congeners for both active and inactive lake trout. Because the experiment was replicated in eight tanks, the standard error about mean γ could be estimated. Results from this type of experiment are useful in risk assessment models to predict future risk to humans and wildlife eating contaminated fish under various scenarios of environmental contamination.
Directory of Open Access Journals (Sweden)
Zoltán Endre Rákossy
2012-01-01
Full Text Available Due to the fast changing wireless communication standards coupled with strict performance constraints, the demand for flexible yet high-performance architectures is increasing. To tackle the flexibility requirement, software-defined radio (SDR is emerging as an obvious solution, where the underlying hardware implementation is tuned via software layers to the varied standards depending on power-performance and quality requirements leading to adaptable, cognitive radio. In this paper, we conduct a case study for representatives of two complexity classes of WCDMA channel estimation algorithms and explore the effect of flexibility on energy efficiency using different implementation options. Furthermore, we propose new design guidelines for both highly specialized architectures and highly flexible architectures using high-level synthesis, to enable the required performance and flexibility to support multiple applications. Our experiments with various design points show that the resulting architectures meet the performance constraints of WCDMA and a wide range of options are offered for tuning such architectures depending on power/performance/area constraints of SDR.
Syrejshchikova, T. I.; Gryzunov, Yu. A.; Smolina, N. V.; Komar, A. A.; Uzbekov, M. G.; Misionzhnik, E. J.; Maksimova, N. M.
2010-05-01
The efficiency of the therapy of psychiatric diseases is estimated using the fluorescence measurements of the conformational changes of human serum albumin in the course of medical treatment. The fluorescence decay curves of the CAPIDAN probe (N-carboxyphenylimide of the dimethylaminonaphthalic acid) in the blood serum are measured. The probe is specifically bound to the albumin drug binding sites and exhibits fluorescence as a reporter ligand. A variation in the conformation of the albumin molecule substantially affects the CAPIDAN fluorescence decay curve on the subnanosecond time scale. A subnanosecond pulsed laser or a Pico-Quant LED excitation source and a fast photon detector with a time resolution of about 50 ps are used for the kinetic measurements. The blood sera of ten patients suffering from depression and treated at the Institute of Psychiatry were preliminary clinically tested. Blood for analysis was taken from each patient prior to the treatment and on the third week of treatment. For ten patients, the analysis of the fluorescence decay curves of the probe in the blood serum using the three-exponential fitting shows that the difference between the amplitudes of the decay function corresponding to the long-lived (9 ns) fluorescence of the probe prior to and after the therapeutic procedure reliably differs from zero at a significance level of 1% ( p < 0.01).
Garnier, Romain; Odunlami, Marc; Le Bris, Vincent; Bégué, Didier; Baraille, Isabelle; Coulaud, Olivier
2016-05-01
A new variational algorithm called adaptive vibrational configuration interaction (A-VCI) intended for the resolution of the vibrational Schrödinger equation was developed. The main advantage of this approach is to efficiently reduce the dimension of the active space generated into the configuration interaction (CI) process. Here, we assume that the Hamiltonian writes as a sum of products of operators. This adaptive algorithm was developed with the use of three correlated conditions, i.e., a suitable starting space, a criterion for convergence, and a procedure to expand the approximate space. The velocity of the algorithm was increased with the use of a posteriori error estimator (residue) to select the most relevant direction to increase the space. Two examples have been selected for benchmark. In the case of H2CO, we mainly study the performance of A-VCI algorithm: comparison with the variation-perturbation method, choice of the initial space, and residual contributions. For CH3CN, we compare the A-VCI results with a computed reference spectrum using the same potential energy surface and for an active space reduced by about 90%.
International Nuclear Information System (INIS)
Waring, R.H.; Law, B.E.; Goulden, M.L.; Bassow, S.L.; McCreight, R.W.; Wofsy, S.C.; Bazzaz, F.A.
1995-01-01
Two independent methods of estimating gross ecosystem production (GEP) were compared over a period of 2 years at monthly integrals for a mixed forest of conifers and deciduous hardwoods at Harvard Forest in central Massachusetts. Continuous eddy flux measurements of net ecosystem exchange (NEE) provided one estimate of GEP by taking day to night temperature differences into account to estimate autotrophic and heterotrophic respiration. GEP was also estimated with a quantum efficiency model based on measurements of maximum quantum efficiency (Qmax), seasonal variation in canopy phenology and chlorophyll content, incident PAR, and the constraints of freezing temperatures and vapour pressure deficits on stomatal conductance. Quantum efficiency model estimates of GEP and those derived from eddy flux measurements compared well at monthly integrals over two consecutive years (R 2 = 0–98). Remotely sensed data were acquired seasonally with an ultralight aircraft to provide a means of scaling the leaf area and leaf pigmentation changes that affected the light absorption of photosynthetically active radiation to larger areas. A linear correlation between chlorophyll concentrations in the upper canopy leaves of four hardwood species and their quantum efficiencies (R 2 = 0–99) suggested that seasonal changes in quantum efficiency for the entire canopy can be quantified with remotely sensed indices of chlorophyll. Analysis of video data collected from the ultralight aircraft indicated that the fraction of conifer cover varied from < 7% near the instrument tower to about 25% for a larger sized area. At 25% conifer cover, the quantum efficiency model predicted an increase in the estimate of annual GEP of < 5% because unfavourable environmental conditions limited conifer photosynthesis in much of the non-growing season when hardwoods lacked leaves
Energy Technology Data Exchange (ETDEWEB)
Kurnik, Charles W [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Baumgartner, Robert [Tetra Tech, Madison, WI (United States)
2017-10-05
This chapter presents an overview of best practices for designing and executing survey research to estimate gross energy savings in energy efficiency evaluations. A detailed description of the specific techniques and strategies for designing questions, implementing a survey, and analyzing and reporting the survey procedures and results is beyond the scope of this chapter. So for each topic covered below, readers are encouraged to consult articles and books cited in References, as well as other sources that cover the specific topics in greater depth. This chapter focuses on the use of survey methods to collect data for estimating gross savings from energy efficiency programs.
Marshall, M.; Tu, K. P.
2015-12-01
Large-area crop yield models (LACMs) are commonly employed to address climate-driven changes in crop yield and inform policy makers concerned with climate change adaptation. Production efficiency models (PEMs), a class of LACMs that rely on the conservative response of carbon assimilation to incoming solar radiation absorbed by a crop contingent on environmental conditions, have increasingly been used over large areas with remote sensing spectral information to improve the spatial resolution of crop yield estimates and address important data gaps. Here, we present a new PEM that combines model principles from the remote sensing-based crop yield and evapotranspiration (ET) model literature. One of the major limitations of PEMs is that they are evaluated using data restricted in both space and time. To overcome this obstacle, we first validated the model using 2009-2014 eddy covariance flux tower Gross Primary Production data in a rice field in the Central Valley of California- a critical agro-ecosystem of the United States. This evaluation yielded a Willmot's D and mean absolute error of 0.81 and 5.24 g CO2/d, respectively, using CO2, leaf area, temperature, and moisture constraints from the MOD16 ET model, Priestley-Taylor ET model, and the Global Production Efficiency Model (GLOPEM). A Monte Carlo simulation revealed that the model was most sensitive to the Enhanced Vegetation Index (EVI) input, followed by Photosynthetically Active Radiation, vapor pressure deficit, and air temperature. The model will now be evaluated using 30 x 30m (Landsat resolution) biomass transects developed in 2011 and 2012 from spectroradiometric and other non-destructive in situ metrics for several cotton, maize, and rice fields across the Central Valley. Finally, the model will be driven by Daymet and MODIS data over the entire State of California and compared with county-level crop yield statistics. It is anticipated that the new model will facilitate agro-climatic decision-making in
Directory of Open Access Journals (Sweden)
A. V. Pasechnik
2017-01-01
Summary Despite of success of dental implantation, there are often complications at the early post-operative period of implant placing associated with wound damage and aseptic inflammation. Purpose of the work is studying clinical efficiency of combined local application of new mucosal gel “Apior” and magnetotherapy at the early post-operative period after dental implantation. Combined local application of the mucosal gel “Apior” and pulsating low-frequency electromagnetic field in the complex medical treatment of patients after conducting an operation of setting dental implants favourably affects the common state of patients and clinical symptoms of inflammation in the area of operating wound. As compared with patients who had traditional anti-inflammatory therapy, the patients treated with local application of apigel and magnetoterapy had decline of edema incidence, of gingival mucosa hyperemia, of discomfort in the area of conducted operation. There occurred more rapid improvement of inflammation painfulness, which correlated with the improvement of hygienic state of oral cavity and promoted to prevention of bacterial content of damaged mucous surfaces. Estimation of microvasculatory blood stream by the method of ultrasonic doppler flowmetry revealed more rapid normalization of volume and linear high systole speed of blood stream in the periimplant tissues in case of use of new complex local rehabilitation method, that testified to the less pronounced inflammation of oral mucosa after the operation. The authors came to conclusion that the local application of the offered method of medical treatment of early post-operative complications of dental implantation reduces terms of renewal of structural-functional integrity of oral mucosa, helps in preventing development of inflammatory complications and strengthening endosseus implant. The inclusion in the treatment management of a new combined method of application of mucosal gel “Apior” and
Parthiban, S; Govindaraj, P; Senthilkumar, S
2018-03-01
Twenty-five primer pairs developed from genomic simple sequence repeats (SSR) were compared with 25 expressed sequence tags (EST) SSRs to evaluate the efficiency of these two sets of primers using 59 sugarcane genetic stocks. The mean polymorphism information content (PIC) of genomic SSR was higher (0.72) compared to the PIC value recorded by EST-SSR marker (0.62). The relatively low level of polymorphism in EST-SSR markers may be due to the location of these markers in more conserved and expressed sequences compared to genomic sequences which are spread throughout the genome. Dendrogram based on the genomic SSR and EST-SSR marker data showed differences in grouping of genotypes. A total of 59 sugarcane accessions were grouped into 6 and 4 clusters using genomic SSR and EST-SSR, respectively. The highly efficient genomic SSR could subcluster the genotypes of some of the clusters formed by EST-SSR markers. The difference in dendrogram observed was probably due to the variation in number of markers produced by genomic SSR and EST-SSR and different portion of genome amplified by both the markers. The combined dendrogram (genomic SSR and EST-SSR) more clearly showed the genetic relationship among the sugarcane genotypes by forming four clusters. The mean genetic similarity (GS) value obtained using EST-SSR among 59 sugarcane accessions was 0.70, whereas the mean GS obtained using genomic SSR was 0.63. Although relatively lower level of polymorphism was displayed by the EST-SSR markers, genetic diversity shown by the EST-SSR was found to be promising as they were functional marker. High level of PIC and low genetic similarity values of genomic SSR may be more useful in DNA fingerprinting, selection of true hybrids, identification of variety specific markers and genetic diversity analysis. Identification of diverse parents based on cluster analysis can be effectively done with EST-SSR as the genetic similarity estimates are based on functional attributes related to
Furlotte, Nicholas A; Eskin, Eleazar
2015-05-01
Multiple-trait association mapping, in which multiple traits are used simultaneously in the identification of genetic variants affecting those traits, has recently attracted interest. One class of approaches for this problem builds on classical variance component methodology, utilizing a multitrait version of a linear mixed model. These approaches both increase power and provide insights into the genetic architecture of multiple traits. In particular, it is possible to estimate the genetic correlation, which is a measure of the portion of the total correlation between traits that is due to additive genetic effects. Unfortunately, the practical utility of these methods is limited since they are computationally intractable for large sample sizes. In this article, we introduce a reformulation of the multiple-trait association mapping approach by defining the matrix-variate linear mixed model. Our approach reduces the computational time necessary to perform maximum-likelihood inference in a multiple-trait model by utilizing a data transformation. By utilizing a well-studied human cohort, we show that our approach provides more than a 10-fold speedup, making multiple-trait association feasible in a large population cohort on the genome-wide scale. We take advantage of the efficiency of our approach to analyze gene expression data. By decomposing gene coexpression into a genetic and environmental component, we show that our method provides fundamental insights into the nature of coexpressed genes. An implementation of this method is available at http://genetics.cs.ucla.edu/mvLMM. Copyright © 2015 by the Genetics Society of America.
Energy Technology Data Exchange (ETDEWEB)
Letschert, Virginie E. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Bojda, Nicholas [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Ke, Jing [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); McNeil, Michael A. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)
2012-07-01
This study analyzes the financial impacts on consumers of minimum efficiency performance standards (MEPS) for appliances that could be implemented in 13 major economies around the world. We use the Bottom-Up Energy Analysis System (BUENAS), developed at Lawrence Berkeley National Laboratory (LBNL), to analyze various appliance efficiency target levels to estimate the net present value (NPV) of policies designed to provide maximum energy savings while not penalizing consumers financially. These policies constitute what we call the “cost-effective potential” (CEP) scenario. The CEP scenario is designed to answer the question: How high can we raise the efficiency bar in mandatory programs while still saving consumers money?
Directory of Open Access Journals (Sweden)
A. P. Kurilo
2010-06-01
Full Text Available The theorem of system of indicators for an estimation of the security of information processes in the computer systems is formulated and proved. A number of the signs is proved, allowing to consider set of the indicators of efficiency of counteraction to the threats of information safety of the computer systems as the system.
Directory of Open Access Journals (Sweden)
Daud Jones Kachamba
2017-06-01
Full Text Available Applications of unmanned aircraft systems (UASs to assist in forest inventories have provided promising results in biomass estimation for different forest types. Recent studies demonstrating use of different types of remotely sensed data to assist in biomass estimation have shown that accuracy and precision of estimates are influenced by the size of field sample plots used to obtain reference values for biomass. The objective of this case study was to assess the influence of sample plot size on efficiency of UAS-assisted biomass estimates in the dry tropical miombo woodlands of Malawi. The results of a design-based field sample inventory assisted by three-dimensional point clouds obtained from aerial imagery acquired with a UAS showed that the root mean square errors as well as the standard error estimates of mean biomass decreased as sample plot sizes increased. Furthermore, relative efficiency values over different sample plot sizes were above 1.0 in a design-based and model-assisted inferential framework, indicating that UAS-assisted inventories were more efficient than purely field-based inventories. The results on relative costs for UAS-assisted and pure field-based sample plot inventories revealed that there is a trade-off between inventory costs and required precision. For example, in our study if a standard error of less than approximately 3 Mg ha−1 was targeted, then a UAS-assisted forest inventory should be applied to ensure more cost effective and precise estimates. Future studies should therefore focus on finding optimum plot sizes for particular applications, like for example in projects under the Reducing Emissions from Deforestation and Forest Degradation, plus forest conservation, sustainable management of forest and enhancement of carbon stocks (REDD+ mechanism with different geographical scales.
International Nuclear Information System (INIS)
Takai, Toshihide; Nakagiri, Toshio; Inagaki, Yoshiyuki
2008-10-01
A new experimental apparatus by the thermo-chemical and electrolytic Hybrid-Hydrogen production in Lower Temperature range (HHLT) was developed and hydrogen production experiment was performed to confirm the system operability. Hydrogen production efficiency was estimated and technical problems were clarified through the experimental results. Stable operation of the SO 3 electrolysis cell and the sulfur dioxide solution electrolysis cell were confirmed during experimental operation and any damage which would be affected solid operation was not detected under post operation inspection. To improve hydrogen production efficiency, it was found that the reduction of sulfuric acid circulation and the decrease in the cell voltage were key issues. (author)
Rao, D V; Brunetti, A; Gigante, G E; Takeda, T; Itai, Y; Akatsuka, T
2002-01-01
A new approach is developed to estimate the geometrical factors, solid angle approximation and geometrical efficiency for a system with experimental arrangements using X-ray tube and secondary target as an excitation source in order to produce the nearly monoenergetic K alpha radiation to excite the sample. The variation of the solid angle is studied by changing the radius and length of the collimators towards and away from the source and sample. From these values the variation of the total solid angle and geometrical efficiency is deduced and the optimum value is used for the experimental work. (authors)
Energy Technology Data Exchange (ETDEWEB)
Rao, D.V.; Cesareo, R.; Brunetti, A. [Sassari University, Istituto di Matematica e Fisica (Italy); Gigante, G.E. [Roma Universita, Dipt. di Fisica (Italy); Takeda, T.; Itai, Y. [Tsukuba Univ., Ibaraki (Japan). Inst. of Clinical Medicine; Akatsuka, T. [Yamagata Univ., Yonezawa (Japan). Faculty of Engineering
2002-10-01
A new approach is developed to estimate the geometrical factors, solid angle approximation and geometrical efficiency for a system with experimental arrangements using X-ray tube and secondary target as an excitation source in order to produce the nearly monoenergetic K{alpha} radiation to excite the sample. The variation of the solid angle is studied by changing the radius and length of the collimators towards and away from the source and sample. From these values the variation of the total solid angle and geometrical efficiency is deduced and the optimum value is used for the experimental work. (authors)
Rao, D. V.; Cesareo, R.; Brunetti, A.; Gigante, G. E.; Takeda, T.; Itai, Y.; Akatsuka, T.
2002-10-01
A new approach is developed to estimate the geometrical factors, solid angle approximation and geometrical efficiency for a system with experimental arrangements using X-ray tube and secondary target as an excitation source in order to produce the nearly monoenergetic Kα radiation to excite the sample. The variation of the solid angle is studied by changing the radius and length of the collimators towards and away from the source and sample. From these values the variation of the total solid angle and geometrical efficiency is deduced and the optimum value is used for the experimental work.
DEFF Research Database (Denmark)
Wirenfeldt, Martin; Dalmau, Ishar; Finsen, Bente
2003-01-01
Stereology offers a set of unbiased principles to obtain precise estimates of total cell numbers in a defined region. In terms of microglia, which in the traumatized and diseased CNS is an extremely dynamic cell population, the strength of stereology is that the resultant estimate is unaffected...... of microglia, although with this thickness, the intensity of the staining is too high to distinguish single cells. Lectin histochemistry does not visualize microglia throughout the section and, accordingly, is not suited for the optical fractionator. The mean total number of Mac-1+ microglial cells...... in the unilateral dentate gyrus of the normal young adult male C57BL/6 mouse was estimated to be 12,300 (coefficient of variation (CV)=0.13) with a mean coefficient of error (CE) of 0.06. The perspective of estimating microglial cell numbers using stereology is to establish a solid basis for studying the dynamics...
Directory of Open Access Journals (Sweden)
Edinam Dope Setsoafia
2017-01-01
Full Text Available This study evaluated the profit efficiency of artisanal fishing in the Pru District of Ghana by explicitly computing profit efficiency level, identifying the sources of profit inefficiency, and examining the constraints of artisanal fisheries. Cross-sectional data was obtained from 120 small-scale fishing households using semistructured questionnaire. The stochastic profit frontier model was used to compute profit efficiency level and identify the determinants of profit inefficiency while Garrett ranking technique was used to rank the constraints. The average profit efficiency level was 81.66% which implies that about 82% of the prospective maximum profit was gained due to production efficiency. That is, only 18% of the potential profit was lost due to the fishers’ inefficiency. Also, the age of the household head and household size increase the inefficiency level while experience in artisanal fishing tends to decrease the inefficiency level. From the Garrett ranking, access to credit facility to fully operate the small-scale fishing business was ranked as the most pressing issue followed by unstable prices while perishability was ranked last among the constraints. The study, therefore, recommends that group formation should be encouraged to enable easy access to loans and contract sales to boost profitability.
Directory of Open Access Journals (Sweden)
Shahriar Shirvani Moghaddam
2012-01-01
Full Text Available Uniform linear array (ULA geometry does not perform well for direction of arrival (DOA estimation at directions close to the array endfires. Shirvani and Akbari solved this problem by displacing two elements from both ends of the ULA to the top and/or bottom of the array axis. Shirvani-Akbari array (SAA presents a considerable improvement in the DOA estimation of narrowband sources arriving at endfire directions in terms of DOA estimation accuracy and angular resolution. In this paper, all new proposed SAA configurations are modelled and also examined, numerically. In this paper, two well-known DOA estimation algorithms, multiple signal classification (MUSIC and minimum variance distortionless response (MVDR, are used to evaluate the effectiveness of proposed arrays using total root mean square error (RMSE criterion. In addition, two new scenarios are proposed which divide angular search to two parts, directions close to array endfires as well as middle angles. For middle angles, which belong to (−70∘≤≤70∘, ULA is considered, and for endfire angles, the angles which belong to (−90∘≤≤−70∘ and (70∘≤≤90∘, SAA is considered. Simulation results of new proposed scenarios for DOA estimation of narrowband signals show the better performance with lower computational load.
Gong, Jian; Lou, Shuntian; Guo, Yiduo
2016-04-01
An estimation of signal parameters via a rotational invariance techniques-like (ESPRIT-like) algorithm is proposed to estimate the direction of arrival and direction of departure for bistatic multiple-input multiple-output (MIMO) radar. The properties of a noncircular signal and Euler's formula are first exploited to establish a real-valued bistatic MIMO radar array data, which is composed of sine and cosine data. Then the receiving/transmitting selective matrices are constructed to obtain the receiving/transmitting rotational invariance factors. Since the rotational invariance factor is a cosine function, symmetrical mirror angle ambiguity may occur. Finally, a maximum likelihood function is used to avoid the estimation ambiguities. Compared with the existing ESPRIT, the proposed algorithm can save about 75% of computational load owing to the real-valued ESPRIT algorithm. Simulation results confirm the effectiveness of the ESPRIT-like algorithm.
Directory of Open Access Journals (Sweden)
Boris Grigorievich Saneev
2013-12-01
Full Text Available The authors consider a methodical approach which allows estimating energy efficiency parameters of the region’s economy using a fuel and energy balance (FEB. This approach was tested on the specific case of Baikal region. During the testing process the authors have developed ex ante and ex post FEBs and estimated energy efficiency parameters such as energy-, electro- and heat capacity of GRP, coefficients of useful utilization of fuel and energy resources and a monetary version of FEB. Forecast estimations are based on assumptions and limitations of technologically-intensive development scenario of the region. Authors show that the main factor of structural changes in the fuel and energy balance will be the large-scale development of hydrocarbon resources in Baikal region. It will cause structural changes in the composition of both the debit and credit of FEB (namely the structure of export and final consumption of fuel and energy resources. Authors assume that the forecast structural changes of the region’s FEB will significantly improve energy efficiency parameters of the economy: energy capacity of GRP will decrease by 1,5 times in 2010– 2030, electro and heat capacity – 1,9 times; coefficients of useful utilization of fuel and energy resources will increase by 3–5 p.p. This will save about 20 million tons of fuel equivalent (about 210 billion rubles in 2011 the prices until 2030
Solimun, Fernandes, Adji Achmad Rinaldo; Arisoesilaningsih, Endang
2017-12-01
Research in various fields generally investigates systems and involves latent variables. One method to analyze the model representing the system is path analysis. The data of latent variables measured using questionnaires by applying attitude scale model yields data in the form of score, before analyzed should be transformation so that it becomes data of scale. Path coefficient, is parameter estimator, calculated from scale data using method of successive interval (MSI) and summated rating scale (SRS). In this research will be identifying which data transformation method is better. Path coefficients have smaller varieties are said to be more efficient. The transformation method that produces scaled data and used in path analysis capable of producing path coefficients (parameter estimators) with smaller varieties is said to be better. The result of analysis using real data shows that on the influence of Attitude variable to Intention Entrepreneurship, has relative efficiency (ER) = 1, where it shows that the result of analysis using data transformation of MSI and SRS as efficient. On the other hand, for simulation data, at high correlation between items (0.7-0.9), MSI method is more efficient 1.3 times better than SRS method.
Lalam, N.; Jacob, C.; Jagers, P.
2004-01-01
We propose a stochastic modelling of the PCR amplification process by a size-dependent branching process starting as a supercritical Bienaymé-Galton-Watson transient phase and then having a saturation near-critical size-dependent phase. This model allows us to estimate the probability of replication
高效超分辨波达方向估计算法综述%Overview of efficient algorithms for super-resolution DOA estimates
Institute of Scientific and Technical Information of China (English)
闫锋刚; 沈毅; 刘帅; 金铭; 乔晓林
2015-01-01
Computationally efficient methods for super-resolution direction of arrival (DOA)estimation aim to reduce the complexity of conventional techniques,to economize on the costs of systems and to enhance the ro-bustness of DOA estimators against array geometries and other environmental restrictions,which has been an important topic in the field.According to the theory and elements of the multiple signal classification (MUSIC) algorithm and the primary derivations from MUSIC,state-of-the-art efficient super-resolution DOA estimators are classified into five different types.These five types of approaches reduce the complexity by real-valued com-putation,beam-space transformation,fast subspace estimation,rapid spectral search,and no spectral search, respectively.With such a classification,comprehensive overviews of each kind of efficient methods are given and numerical comparisons among these estimators are also conducted to illustrate their advantages.Future develop-ment trends of efficient algorithms for super-resolution DOA estimates are finally predicted with basic require-ments of real-world applications.%高效超分辨波达方向估计算法致力于降低超分辨算法的计算量、节约系统的实现成本、弱化算法对于阵列结构的依赖性，是推进超分辨理论工程化的一个重要研究课题。从多重信号分类（multiple signal classifi-cation，MUSIC）算法的原理和构成要素入手，以基于 MUSIC 派生高效超分辨算法的目的和方法为标准，将现存高效超分辨算法划分为实值运算、波束域变换、快速子空间估计、快速峰值搜索和免峰值搜索5大类。在此基础上，全面回顾总结了各类高效算法的发展历程和最新进展，对比分析了它们的主要优缺点。最后，结合空间谱估计实际工程化的应用需求，指出了高效超分辨算法的未来发展趋势。
Andreas Tewes; Jürgen Schellberg
2018-01-01
Radiation Use Efficiency (RUE) defines the productivity with which absorbed photosynthetically active radiation (APAR) is converted to plant biomass. Readily used in crop growth models to predict dry matter accumulation, RUE is commonly determined by elaborate static sensor measurements in the field. Different definitions are used, based on total absorbed PAR (RUEtotal) or PAR absorbed by the photosynthetically active leaf tissue only (RUEgreen). Previous studies have shown that the fraction ...
International Nuclear Information System (INIS)
Pop, O.M.; Stets, M.V.; Maslyuk, V.T.
2015-01-01
We consider a gamma-spectrometric complex of IEP of the NAS of Ukraine, where a passive multilayer external defense is used (complex has been made in 1989). We have developed and investigated a system of stability and lowering of background in the gamma-spectrometric complex. As metrological coefficients, the efficiency factor of defense are considered, the calculation and analysis of which show that their values are different for different energies of gamma-quanta and gamma-active nuclides
Total-Factor Energy Efficiency in BRI Countries: An Estimation Based on Three-Stage DEA Model
Directory of Open Access Journals (Sweden)
Changhong Zhao
2018-01-01
Full Text Available The Belt and Road Initiative (BRI is showing its great influence and leadership on the international energy cooperation. Based on the three-stage DEA model, total-factor energy efficiency (TFEE in 35 BRI countries in 2015 was measured in this article. It shows that the three-stage DEA model could eliminate errors of environment variable and random, which made the result better than traditional DEA model. When environment variable errors and random errors were eliminated, the mean value of TFEE was declined. It demonstrated that TFEE of the whole sample group was overestimated because of external environment impacts and random errors. The TFEE indicators of high-income countries like South Korea, Singapore, Israel and Turkey are 1, which is in the efficiency frontier. The TFEE indicators of Russia, Saudi Arabia, Poland and China are over 0.8. And the indicators of Uzbekistan, Ukraine, South Africa and Bulgaria are in a low level. The potential of energy-saving and emissions reduction is great in countries with low TFEE indicators. Because of the gap in energy efficiency, it is necessary to distinguish different countries in the energy technology options, development planning and regulation in BRI countries.
Siciliani, Luigi
2006-01-01
Policy makers are increasingly interested in developing performance indicators that measure hospital efficiency. These indicators may give the purchasers of health services an additional regulatory tool to contain health expenditure. Using panel data, this study compares different parametric (econometric) and non-parametric (linear programming) techniques for the measurement of a hospital's technical efficiency. This comparison was made using a sample of 17 Italian hospitals in the years 1996-9. Highest correlations are found in the efficiency scores between the non-parametric data envelopment analysis under the constant returns to scale assumption (DEA-CRS) and several parametric models. Correlation reduces markedly when using more flexible non-parametric specifications such as data envelopment analysis under the variable returns to scale assumption (DEA-VRS) and the free disposal hull (FDH) model. Correlation also generally reduces when moving from one output to two-output specifications. This analysis suggests that there is scope for developing performance indicators at hospital level using panel data, but it is important that extensive sensitivity analysis is carried out if purchasers wish to make use of these indicators in practice.
Kozai, Toyoki
2013-01-01
Extensive research has recently been conducted on plant factory with artificial light, which is one type of closed plant production system (CPPS) consisting of a thermally insulated and airtight structure, a multi-tier system with lighting devices, air conditioners and fans, a CO2 supply unit, a nutrient solution supply unit, and an environment control unit. One of the research outcomes is the concept of resource use efficiency (RUE) of CPPS.This paper reviews the characteristics of the CPPS compared with those of the greenhouse, mainly from the viewpoint of RUE, which is defined as the ratio of the amount of the resource fixed or held in plants to the amount of the resource supplied to the CPPS.It is shown that the use efficiencies of water, CO2 and light energy are considerably higher in the CPPS than those in the greenhouse. On the other hand, there is much more room for improving the light and electric energy use efficiencies of CPPS. Challenging issues for CPPS and RUE are also discussed.
International Nuclear Information System (INIS)
Bengel, F.M.; Nekolla, S.; Schwaiger, M.; Ungerer, M.
2000-01-01
We studied ten patients with idiopathic dilated cardiomyopathy (DCM) and 11 healthy normals by dynamic PET with 11 C-acetate and either tomographic radionuclide ventriculography or cine magnetic resonance imaging. A ''stroke work index'' (SWI) was calculated by: SWI = systolic blood pressure x stroke volume/body surface area. To estimate myocardial efficiency, a ''work-metabolic index'' (WMI) was then obtained as follows: WMI = SWI x heart rate/k(mono), where k(mono) is the washout constant for 11 C-acetate derived from mono-exponential fitting. In DCM patients, left ventricular ejection fraction was 19%±10% and end-diastolic volume was 92±28 ml/m 2 (vs 64%±7% and 55±8 ml/m 2 in normals, P 2 ; P 6 mmHg x ml/m 2 ; P<0.001) were lower in DCM patients, too. Overall, the WMI correlated positively with ejection parameters (r=0.73, P<0.001 for ejection fraction; r=0.93, P<0.001 for stroke volume), and inversely with systemic vascular resistance (r=-0.77; P<0.001). There was a weak positive correlation between WMI and end-diastolic volume in normals (r=0.45; P=0.17), while in DCM patients, a non-significant negative correlation coefficient (r=-0.21; P=0.57) was obtained. In conclusion non-invasive estimates of oxygen consumption and efficiency in the failing heart were reduced compared with those in normals. Estimates of efficiency increased with increasing contractile performance, and decreased with increasing ventricular afterload. In contrast to normals, the failing heart was not able to respond with an increase in efficiency to increasing ventricular volume.(orig./MG) (orig.)
Directory of Open Access Journals (Sweden)
Yury G. Odegov
2016-01-01
Full Text Available In conditions of increasing competition, the problems of efficiency increase of activity of the company are significantly actualized, which directly depends on efficiency of labour activity of every employee and the implemented business model of the organization. On this basis the aim of the research is to analyze existing indicators of performance evaluation of the labour activities of both the employee and the business model of the organization.The theoretical basis of the study consists of principles of the economic theory, the works of native and foreign experts in the field of job evaluation. The information base of the research consists of economic and legal literature dealing with problems of this study, the data published in periodicals, materials of Russian scientific conferences, seminars, and Internet resources.In this article I have used and found the application of scientific methods of data collection, methods of research and methods of assessing their credibility: quantitative, comparative, logical analysis and synthesis.The modern business concern about the accumulation of wealth of shareholders, giving the company stability, growth and efficiency inevitably leads to necessity of creation and development of technologies aimed at improving the productivity of employees. The paper presents a comparative analysis of different approaches to assessing the labour effectiveness.The performance of the work is the ratio of the four essential parameters that determine the measure of efficiency of persons’ activity: the quantity and quality of result of work (a service, material product or technology in relation to spend time and cost on its production. The use of employees («performance» should be in the following way that they could achieve the planned results in the workplace. The authors have noted that to develop of technologies for the measurement of productivity it is very important to use the procedures and indicators that are
Akhmetova, I. G.; Chichirova, N. D.
2017-11-01
When conducting an energy survey of heat supply enterprise operating several boilers located not far from each other, it is advisable to assess the degree of heat supply efficiency from individual boiler, the possibility of energy consumption reducing in the whole enterprise by switching consumers to a more efficient source, to close in effective boilers. It is necessary to consider the temporal dynamics of perspective load connection, conditions in the market changes. To solve this problem the radius calculation of the effective heat supply from the thermal energy source can be used. The disadvantage of existing methods is the high complexity, the need to collect large amounts of source data and conduct a significant amount of computational efforts. When conducting an energy survey of heat supply enterprise operating a large number of thermal energy sources, rapid assessment of the magnitude of the effective heating radius requires. Taking into account the specifics of conduct and objectives of the energy survey method of calculation of effective heating systems radius, to use while conducting the energy audit should be based on data available heat supply organization in open access, minimize efforts, but the result should be to match the results obtained by other methods. To determine the efficiency radius of Kazan heat supply system were determined share of cost for generation and transmission of thermal energy, capital investment to connect new consumers. The result were compared with the values obtained with the previously known methods. The suggested Express-method allows to determine the effective radius of the centralized heat supply from heat sources, in conducting energy audits with the effort minimum and the required accuracy.
Chatterjee, Sharmista; Seagrave, Richard C.
1993-01-01
The objective of this paper is to present an estimate of the second law thermodynamic efficiency of the various units comprising an Environmental Control and Life Support System (ECLSS). The technique adopted here is based on an evaluation of the 'lost work' within each functional unit of the subsystem. Pertinent information for our analysis is obtained from a user interactive integrated model of an ECLSS. The model was developed using ASPEN. A potential benefit of this analysis is the identification of subsystems with high entropy generation as the most likely candidates for engineering improvements. This work has been motivated by the fact that the design objective for a long term mission should be the evaluation of existing ECLSS technologies not only the basis of the quantity of work needed for or obtained from each subsystem but also on the quality of work. In a previous study Brandhorst showed that the power consumption for partially closed and completely closed regenerable life support systems was estimated as 3.5 kw/individual and 10-12 kw/individual respectively. With the increasing cost and scarcity of energy resources, our attention is drawn to evaluate the existing ECLSS technologies on the basis of their energy efficiency. In general the first law efficiency of a system is usually greater than 50 percent. From literature, the second law efficiency is usually about 10 percent. The estimation of second law efficiency of the system indicates the percentage of energy degraded as irreversibilities within the process. This estimate offers more room for improvement in the design of equipment. From another perspective, our objective is to keep the total entropy production of a life support system as low as possible and still ensure a positive entropy gradient between the system and the surroundings. The reason for doing so is as the entropy production of the system increases, the entropy gradient between the system and the surroundings decreases, and the
Dekkers, A L M; Slob, W
2012-10-01
In dietary exposure assessment, statistical methods exist for estimating the usual intake distribution from daily intake data. These methods transform the dietary intake data to normal observations, eliminate the within-person variance, and then back-transform the data to the original scale. We propose Gaussian Quadrature (GQ), a numerical integration method, as an efficient way of back-transformation. We compare GQ with six published methods. One method uses a log-transformation, while the other methods, including GQ, use a Box-Cox transformation. This study shows that, for various parameter choices, the methods with a Box-Cox transformation estimate the theoretical usual intake distributions quite well, although one method, a Taylor approximation, is less accurate. Two applications--on folate intake and fruit consumption--confirmed these results. In one extreme case, some methods, including GQ, could not be applied for low percentiles. We solved this problem by modifying GQ. One method is based on the assumption that the daily intakes are log-normally distributed. Even if this condition is not fulfilled, the log-transformation performs well as long as the within-individual variance is small compared to the mean. We conclude that the modified GQ is an efficient, fast and accurate method for estimating the usual intake distribution. Copyright © 2012 Elsevier Ltd. All rights reserved.
Directory of Open Access Journals (Sweden)
Musakhanov A.K.
2012-12-01
Full Text Available The questions of efficiency of mastering of technique of fight are considered for a capture for young judoists. Directions are selected the use of methods of the strictly regulated exercise and playing methods. In research 28 judoists took part in age 8-10 years. Duration of experiment two weeks. In one group of youths conducted game on snatching out of ribbons (clothes-pins and bandages, fastened on the kimono of opponent. In the second group work of taking of basic captures and educational meetings was conducted on a task on taking of capture. The training program contained playing methods and methods of the strictly regulated exercise. Comparison of the trainings programs defined specificity of their affecting development of different indexes of technique of fight for a capture. Recommended in training on the technique of fight for a capture the combined use of methods of the strictly regulated exercise and playing methods.
Directory of Open Access Journals (Sweden)
Irina N. Sorochinskaya
2012-11-01
Full Text Available Cardiovascular diseases are major reasons for population mortality in majority of countries, including Russia. Metabolic syndrome is considered to be one of the main pathologic states, leading to enhancement of atherogenesis, ischemic heart diseases and cerebrovascular diseases. Physical methods, including resort treatment play great role in metabolic syndrome prevention and treatment. Climate therapy depends on resort climate and season and is a major component of resort treatment. Psychological testing showed that combined resort treatment, using climate therapy of patients with stable effort angina at Sochi Health-resort is more efficient in autumn and of patients with metabolic syndrome in summer. The findings have been confirmed by clinic-functional indicators.
DEFF Research Database (Denmark)
Shirali, Mahmoud; Varley, Patrick Francis; Jensen, Just
2018-01-01
meat percentage (LMP) along with the derived traits of RFI and FCR; and (3) deriving Bayesian estimates of direct and correlated responses to selection on RFI, FCR, ADG, ADFI, and LMP. Response to selection was defined as the difference in additive genetic mean of the selected top individuals, expected......, respectively. Selection against RFIG showed a direct response of − 0.16 kg/d and correlated responses of − 0.16 kg/kg for FCR and − 0.15 kg/d for ADFI, with no effect on other production traits. Selection against FCR resulted in a direct response of − 0.17 kg/kg and correlated responses of − 0.14 kg/d for RFIG......, − 0.18 kg/d for ADFI, and 0.98% for LMP. Conclusions: The Bayesian methodology developed here enables prediction of breeding values for FCR and RFI from a single multi-variate model. In addition, we derived posterior distributions of direct and correlated responses to selection. Genetic parameter...
Directory of Open Access Journals (Sweden)
Kowal Robert
2016-12-01
Full Text Available A simple linear regression model is one of the pillars of classic econometrics. Multiple areas of research function within its scope. One of the many fundamental questions in the model concerns proving the efficiency of the most commonly used OLS estimators and examining their properties. In the literature of the subject one can find taking back to this scope and certain solutions in that regard. Methodically, they are borrowed from the multiple regression model or also from a boundary partial model. Not everything, however, is here complete and consistent. In the paper a completely new scheme is proposed, based on the implementation of the Cauchy-Schwarz inequality in the arrangement of the constraint aggregated from calibrated appropriately secondary constraints of unbiasedness which in a result of choice the appropriate calibrator for each variable directly leads to showing this property. A separate range-is a matter of choice of such a calibrator. These deliberations, on account of the volume and kinds of the calibration, were divided into a few parts. In the one the efficiency of OLS estimators is proven in a mixed scheme of the calibration by averages, that is preliminary, and in the most basic frames of the proposed methodology. In these frames the future outlines and general premises constituting the base of more distant generalizations are being created.
Ionkin, I. L.; Ragutkin, A. V.; Luning, B.; Zaichenko, M. N.
2016-06-01
For enhancement of the natural gas utilization efficiency in boilers, condensation heat utilizers of low-potential heat, which are constructed based on a contact heat exchanger, can be applied. A schematic of the contact heat exchanger with a humidifier for preheating and humidifying of air supplied in the boiler for combustion is given. Additional low-potential heat in this scheme is utilized for heating of the return delivery water supplied from a heating system. Preheating and humidifying of air supplied for combustion make it possible to use the condensation utilizer for heating of a heat-transfer agent to temperature exceeding the dewpoint temperature of water vapors contained in combustion products. The decision to mount the condensation heat utilizer on the boiler was taken based on the preliminary estimation of the additionally obtained heat. The operation efficiency of the condensation heat utilizer is determined by its structure and operation conditions of the boiler and the heating system. The software was developed for the thermal design of the condensation heat utilizer equipped by the humidifier. Computation investigations of its operation are carried out as a function of various operation parameters of the boiler and the heating system (temperature of the return delivery water and smoke fumes, air excess, air temperature at the inlet and outlet of the condensation heat utilizer, heating and humidifying of air in the humidifier, and portion of the circulating water). The heat recuperation efficiency is estimated for various operation conditions of the boiler and the condensation heat utilizer. Recommendations on the most effective application of the condensation heat utilizer are developed.
Directory of Open Access Journals (Sweden)
Davide Biagini
2011-04-01
Full Text Available To evaluate the effect of sexual neutering and age of castration on empty body weight (EBW components and estimated nitrogen excretion and efficiency, a trial was carried out on 3 groups of double-muscled Piemontese calves: early castrated (EC, 5th month of age, late castrated (LC, 12th month of age and intact males (IM, control group. Animals were fed at the same energy and protein level and slaughtered at 18th month of age. Live and slaughtering performances and EBW components were recorded, whereas N excretion was calculated by difference between diet and weight gain N content. In live and slaughtering performances, IM showed higher final, carcass and total meat weight than EC and LC (P<0.01. In EBW components, IM showed higher blood and head weight than EC and LC (P<0.01 and 0.05 respectively, and differences were found between EC and LC for head weights (P<0.01. IM showed higher body crude protein (BCP than EC and LC (P<0.01 and 0.05 respectively, but BCP/EBW ratio was higher only in IM than EC (P<0.05. Estimated N daily gain was higher in IM than EC and LC (P<0.01. Only LC showed higher excretion than IM (P<0.05, and N efficiency was higher in IM than EC and LC (P<0.05 and 0.01 respectively. In conclusion, for the Piemontese hypertrophied cattle castration significantly increases N excretion (+7% and reduces N efficiency (-15%, leading to a lower level of sustainability.
Directory of Open Access Journals (Sweden)
Constantin E Uhlig
Full Text Available To evaluate the relative efficiencies of five Internet-based digital and three paper-based scientific surveys and to estimate the costs for different-sized cohorts.Invitations to participate in a survey were distributed via e-mail to employees of two university hospitals (E1 and E2 and to members of a medical association (E3, as a link placed in a special text on the municipal homepage regularly read by the administrative employees of two cities (H1 and H2, and paper-based to workers at an automobile enterprise (P1 and college (P2 and senior (P3 students. The main parameters analyzed included the numbers of invited and actual participants, and the time and cost to complete the survey. Statistical analysis was descriptive, except for the Kruskal-Wallis-H-test, which was used to compare the three recruitment methods. Cost efficiencies were compared and extrapolated to different-sized cohorts.The ratios of completely answered questionnaires to distributed questionnaires were between 81.5% (E1 and 97.4% (P2. Between 6.4% (P1 and 57.0% (P2 of the invited participants completely answered the questionnaires. The costs per completely answered questionnaire were $0.57-$1.41 (E1-3, $1.70 and $0.80 for H1 and H2, respectively, and $3.36-$4.21 (P1-3. Based on our results, electronic surveys with 10, 20, 30, or 42 questions would be estimated to be most cost (and time efficient if more than 101.6-225.9 (128.2-391.7, 139.8-229.2 (93.8-193.6, 165.8-230.6 (68.7-115.7, or 188.2-231.5 (44.4-72.7 participants were required, respectively.The study efficiency depended on the technical modalities of the survey methods and engagement of the participants. Depending on our study design, our results suggest that in similar projects that will certainly have more than two to three hundred required participants, the most efficient way of conducting a questionnaire-based survey is likely via the Internet with a digital questionnaire, specifically via a centralized e-mail.
Directory of Open Access Journals (Sweden)
Tianxiang Cui
2017-12-01
Full Text Available Accurately quantifying gross primary production (GPP is of vital importance to understanding the global carbon cycle. Light-use efficiency (LUE models and process-based models have been widely used to estimate GPP at different spatial and temporal scales. However, large uncertainties remain in quantifying GPP, especially for croplands. Recently, remote measurements of solar-induced chlorophyll fluorescence (SIF have provided a new perspective to assess actual levels of plant photosynthesis. In the presented study, we evaluated the performance of three approaches, including the LUE-based multi-source data synergized quantitative (MuSyQ GPP algorithm, the process-based boreal ecosystem productivity simulator (BEPS model, and the SIF-based statistical model, in estimating the diurnal courses of GPP at a maize site in Zhangye, China. A field campaign was conducted to acquire synchronous far-red SIF (SIF760 observations and flux tower-based GPP measurements. Our results showed that both SIF760 and GPP were linearly correlated with APAR, and the SIF760-GPP relationship was adequately characterized using a linear function. The evaluation of the modeled GPP against the GPP measured from the tower demonstrated that all three approaches provided reasonable estimates, with R2 values of 0.702, 0.867, and 0.667 and RMSE values of 0.247, 0.153, and 0.236 mg m−2 s−1 for the MuSyQ-GPP, BEPS and SIF models, respectively. This study indicated that the BEPS model simulated the GPP best due to its efficiency in describing the underlying physiological processes of sunlit and shaded leaves. The MuSyQ-GPP model was limited by its simplification of some critical ecological processes and its weakness in characterizing the contribution of shaded leaves. The SIF760-based model demonstrated a relatively limited accuracy but showed its potential in modeling GPP without dependency on climate inputs in short-term studies.
Schubert, J. E.; Sanders, B. F.
2011-12-01
Urban landscapes are at the forefront of current research efforts in the field of flood inundation modeling for two major reasons. First, urban areas hold relatively large economic and social importance and as such it is imperative to avoid or minimize future damages. Secondly, urban flooding is becoming more frequent as a consequence of continued development of impervious surfaces, population growth in cities, climate change magnifying rainfall intensity, sea level rise threatening coastal communities, and decaying flood defense infrastructure. In reality urban landscapes are particularly challenging to model because they include a multitude of geometrically complex features. Advances in remote sensing technologies and geographical information systems (GIS) have promulgated fine resolution data layers that offer a site characterization suitable for urban inundation modeling including a description of preferential flow paths, drainage networks and surface dependent resistances to overland flow. Recent research has focused on two-dimensional modeling of overland flow including within-curb flows and over-curb flows across developed parcels. Studies have focused on mesh design and parameterization, and sub-grid models that promise improved performance relative to accuracy and/or computational efficiency. This presentation addresses how fine-resolution data, available in Los Angeles County, are used to parameterize, initialize and execute flood inundation models for the 1963 Baldwin Hills dam break. Several commonly used model parameterization strategies including building-resistance, building-block and building hole are compared with a novel sub-grid strategy based on building-porosity. Performance of the models is assessed based on the accuracy of depth and velocity predictions, execution time, and the time and expertise required for model set-up. The objective of this study is to assess field-scale applicability, and to obtain a better understanding of advantages
Zhang, Feng; Zhou, Guangsheng
2017-07-01
We estimated the light use efficiency ( LUE ) via vegetation canopy chlorophyll content ( CCC canopy ) based on in situ measurements of spectral reflectance, biophysical characteristics, ecosystem CO 2 fluxes and micrometeorological factors over a maize canopy in Northeast China. The results showed that among the common chlorophyll-related vegetation indices (VIs), CCC canopy had the most obviously exponential relationships with the red edge position (REP) ( R 2 = .97, p < .001) and normalized difference vegetation index (NDVI) ( R 2 = .91, p < .001). In a comparison of the indicating performances of NDVI, ratio vegetation index (RVI), wide dynamic range vegetation index (WDRVI), and 2-band enhanced vegetation index (EVI2) when estimating CCC canopy using all of the possible combinations of two separate wavelengths in the range 400-1300 nm, EVI2 [1214, 1259] and EVI2 [726, 1248] were better indicators, with R 2 values of .92 and .90 ( p < .001). Remotely monitoring LUE through estimating CCC canopy derived from field spectrometry data provided accurate prediction of midday gross primary productivity ( GPP ) in a rainfed maize agro-ecosystem ( R 2 = .95, p < .001). This study provides a new paradigm for monitoring vegetation GPP based on the combination of LUE models with plant physiological properties.
International Nuclear Information System (INIS)
Song, Myung Sub; Kim, Song Hyun; Kim, Jong Kyung; Noh, Jae Man
2014-01-01
The uncertainty with the sampling-based method is evaluated by repeating transport calculations with a number of cross section data sampled from the covariance uncertainty data. In the transport calculation with the sampling-based method, the transport equation is not modified; therefore, all uncertainties of the responses such as k eff , reaction rates, flux and power distribution can be directly obtained all at one time without code modification. However, a major drawback with the sampling-based method is that it requires expensive computational load for statistically reliable results (inside confidence level 0.95) in the uncertainty analysis. The purpose of this study is to develop a method for improving the computational efficiency and obtaining highly reliable uncertainty result in using the sampling-based method with Monte Carlo simulation. The proposed method is a method to reduce the convergence time of the response uncertainty by using the multiple sets of sampled group cross sections in a single Monte Carlo simulation. The proposed method was verified by estimating GODIVA benchmark problem and the results were compared with that of conventional sampling-based method. In this study, sampling-based method based on central limit theorem is proposed to improve calculation efficiency by reducing the number of repetitive Monte Carlo transport calculation required to obtain reliable uncertainty analysis results. Each set of sampled group cross sections is assigned to each active cycle group in a single Monte Carlo simulation. The criticality uncertainty for the GODIVA problem is evaluated by the proposed and previous method. The results show that the proposed sampling-based method can efficiently decrease the number of Monte Carlo simulation required for evaluate uncertainty of k eff . It is expected that the proposed method will improve computational efficiency of uncertainty analysis with sampling-based method
Directory of Open Access Journals (Sweden)
Valeriy Leonidovich Makarov
2014-12-01
Full Text Available On the basis of the authors’ methodology, the models of productive potential of the Russian Federation regions, including estimations of intellectual capital, were constructed. It is shown that characteristics of well-being level and quality of life make a significant impact on the regional production’s efficiency. The characteristics of regions’ readiness to innovate are identified, it is possible to name it as a factor of production’s efficiency. It is shown that the inclusion of different factors of efficiency in the production potential model can significantly increase the differentiation of technical efficiency estimates, besides these estimates and their grades depend on a set of efficiency’s factors. On the basis of a comparison of real GRP and boundary GRP ratings, it is identified locally effective regions with a relatively high estimation of efficiency among regions with similar amounts of GRP and locally ineffective regions. It is calculated marginal effects of influence of the efficiency’s factors on the result of industrial activity in the region. It seems constructively to use these estimates while analyzing the prospects for regions’ development, which is based on the possibility of targeting impact on controllable efficiency’s factors. The article is also offered the option of methodology of the public policy efficiency estimation on the knowledge economy formation — an agent-based model for Russia, which is learning the “knowledge economy” sector and considering their relationship with the rest of the macroeconomic system.
Estimation of efficiency project management
Directory of Open Access Journals (Sweden)
Novotorov Vladimir Yurevich
2011-03-01
Full Text Available In modern conditions, the effectiveness of the enterprises all in a greater degree depends on methods of management and business dealing forms. The organizations should choose the most effective for themselves strategy of management taking into account the existing legislation, concrete conditions of activity, financial and economic, investment potential and development strategy. Introduction of common system of planning and realization of strategy of the organization, it will allow to provide even development and long-term social and economic growth of the companies.
Saurer, Matthias; Renato, Spahni; Fortunat, Joos; David, Frank; Kerstin, Treydte; Rolf, Siegwolf
2015-04-01
Tree-ring d13C-based estimates of intrinsic water-use efficiency (iWUE, reflecting the ratio of assimilation A to stomatal conductance gs) generally show a strong increase during the industrial period, likely associated with the increase in atmospheric CO2. However, it is not clear, first, if tree-ring d13C-derived iWUE-values indeed reflect actual plant and ecosystem-scale variability in fluxes and, second, what physiological changes were the drivers of the observed iWUE increase, changes in A or gs or both. To address these questions, we used a complex dynamic vegetation model (LPX) that combines process-based vegetation dynamics with land-atmosphere carbon and water exchange. The analysis was conducted for three functional types, representing conifers, oaks, larch, and various sites in Europe, where tree-ring isotope data are available. The increase in iWUE over the 20th century was comparable in LPX-simulations as in tree-ring-estimates, strengthening confidence in these results. Furthermore, the results from the LPX model suggest that the cause of the iWUE increase was reduced stomatal conductance during recent decades rather than increased assimilation. High-frequency variation reflects the influence of climate, like for example the 1976 summer drought, resulting in strongly reduced A and g in the model, particularly for oak.
International Nuclear Information System (INIS)
Kazemi, Hossein; Bourkheili, Saeid Hassanpour; Kamkar, Behnam; Soltani, Afshin; Gharanjic, Kambiz; Nazari, Noor Mohammad
2016-01-01
Increasing the use of energy inputs in agricultural section has been led to numerous environmental concerns such as greenhouse gas (GHG) emissions, high consumption of non-renewable resources, loss of biodiversity and environment pollutions. The study was aimed to analyze the energy use efficiency (EUE) and estimation of GHG emissions from rainfed–based canola production systems (RCPSs) in Iran. In this study, data were collected from 35 farms in Golestan province (northeast of Iran) by a face to face questionnaire performed and statistical yearbooks of 2014. The amount of GHG emissions (per hectare) from inputs used in RCPSs was calculated using CO 2 emissions coefficient of agricultural inputs. Results showed that the EUE and net energy (NE) were as 3.44 and 35,537.81 MJ ha −1 , respectively. The value of these indices for the study area indicated that surveyed fields are approximately efficient in the use of energy for canola production. The highest share of energy consumption belonged to nitrogen fertilizer (42.09%) followed by diesel fuel (39.81%). In production of rainfed canola, GHG emission was estimated as 1009.91 kg CO 2 equivalent per hectare. Based on the results, nitrogen fertilizer (44.15%), diesel fuel (30.16%) and machinery (14.49%) for field operations had the highest share of GHG emission. The total consumed energy by inputs could be classified as direct energy (40.09%), and indirect energy (59.91%) or renewable energy (2.02%) and nonrenewable energy (97.98%). These results demonstrate that the share of renewable energies in canola production is very low in the studied region and agriculture in Iran is very much dependent on non-renewable energies. In this study, the energy use status in RCPSs has analyzed and the main involved causes have been interpreted. - Highlights: • Fertilizers had the highest share in GHG emission. • The share of renewable energy was low in canola production. • Canola production is efficient in Iran.
Sannigrahi, Srikanta; Sen, Somnath; Paul, Saikat
2016-04-01
Net Primary Production (NPP) of mangrove ecosystem and its capacity to sequester carbon from the atmosphere may be used to quantify the regulatory ecosystem services. Three major group of parameters has been set up as BioClimatic Parameters (BCP): (Photosynthetically Active Radiation (PAR), Absorbed PAR (APAR), Fraction of PAR (FPAR), Photochemical Reflectance Index (PRI), Light Use Efficiency (LUE)), BioPhysical Parameters (BPP) :(Normalize Difference Vegetation Index (NDVI), scaled NDVI, Enhanced Vegetation Index (EVI), scaled EVI, Optimised and Modified Soil Adjusted Vegetation Index (OSAVI, MSAVI), Leaf Area Index (LAI)), and Environmental Limiting Parameters (ELP) (Temperature Stress (TS), Land Surface Water Index (LSWI), Normalize Soil Water Index (NSWI), Water Stress Scalar (WS), Inversed WS (iWS) Land Surface Temperature (LST), scaled LST, Vapor Pressure Deficit (VPD), scaled VPD, and Soil Water Deficit Index (SWDI)). Several LUE models namely Carnegie Ames Stanford Approach (CASA), Eddy Covariance - LUE (EC-LUE), Global Production Efficiency Model (GloPEM), Vegetation Photosynthesis Model (VPM), MOD NPP model, Temperature and Greenness Model (TG), Greenness and Radiation model (GR) and MOD17 was adopted in this study to assess the spatiotemporal nature of carbon fluxes. Above and Below Ground Biomass (AGB & BGB) was calculated using field based estimation of OSAVI and NDVI. Microclimatic zonation has been set up to assess the impact of coastal climate on environmental limiting factors. MODerate Resolution Imaging Spectroradiometer (MODIS) based yearly Gross Primary Production (GPP) and NPP product MOD17 was also tested with LUE based results with standard model validation statistics: Root Mean Square of Error (RMSE), Mean Absolute Error (MEA), Bias, Coefficient of Variation (CV) and Coefficient of Determination (R2). The performance of CASA NPP was tested with the ground based NPP with R2 = 0.89 RMSE = 3.28 P = 0.01. Among the all adopted models, EC
Dubrovskaya, Ekaterina; Turkovskaya, Olga
2010-05-01
Estimation of the efficiency of hydrocarbon mineralization in soil by measuring CO2-emission and variations in the isotope composition of carbon dioxide E. Dubrovskaya1, O. Turkovskaya1, A. Tiunov2, N. Pozdnyakova1, A. Muratova1 1 - Institute of Biochemistry and Physiology of Plants and Microorganisms, RAS, Saratov, 2 - A.N. Severtsov Institute of Ecology and Evolution, RAS, Moscow, Russian Federation Hydrocarbon mineralization in soil undergoing phytoremediation was investigated in a laboratory experiment by estimating the variation in the 13С/12С ratio in the respired СО2. Hexadecane (HD) was used as a model hydrocarbon pollutant. The polluted soil was planted with winter rye (Secale cereale) inoculated with Azospirillum brasilense strain SR80, which combines the abilities to promote plant growth and to degrade oil hydrocarbon. Each vegetated treatment was accompanied with a corresponding nonvegetated one, and uncontaminated treatments were used as controls. Emission of carbon dioxide, its isotopic composition, and the residual concentration of HD in the soil were examined after two and four weeks. At the beginning of the experiment, the CO2-emission level was higher in the uncontaminated than in the contaminated soil. After two weeks, the quantity of emitted carbon dioxide decreased by about three times and did not change significantly in all uncontaminated treatments. The presence of HD in the soil initially increased CO2 emission, but later the respiration was reduced. During the first two weeks, nonvegetated soil had the highest CO2-emission level. Subsequently, the maximum increase in respiration was recorded in the vegetated contaminated treatments. The isotope composition of plant material determines the isotope composition of soil. The soil used in our experiment had an isotopic signature typical of soils formed by C3 plants (δ13C,-22.4‰). Generally, there was no significant fractionation of the carbon isotopes of the substrates metabolized by the
Zheng, T.; Chen, J. M.
2016-12-01
The maximum carboxylation rate (Vcmax), despite its importance in terrestrial carbon cycle modelling, remains challenging to obtain for large scales. In this study, an attempt has been made to invert the Vcmax using the gross primary productivity from sunlit leaves (GPPsun) with the physiological basis that the photosynthesis rate for leaves exposed to high solar radiation is mainly determined by the Vcmax. Since the GPPsun can be calculated through the sunlit light use efficiency (ɛsun), the main focus becomes the acquisition of ɛsun. Previous studies using site level reflectance observations have shown the ability of the photochemical reflectance ratio (PRR, defined as the ratio between the reflectance from an effective band centered around 531nm and a reference band) in tracking the variation of ɛsun for an evergreen coniferous stand and a deciduous broadleaf stand separately and the potential of a NDVI corrected PRR (NPRR, defined as the product of NDVI and PRR) in producing a general expression to describe the NPRR-ɛsun relationship across different plant function types. In this study, a significant correlation (R2 = 0.67, p<0.001) between the MODIS derived NPRR and the site level ɛsun calculated using flux data for four Canadian flux sites has been found for the year 2010. For validation purpose, the ɛsun in 2009 for the same sites are calculated using the MODIS NPRR and the expression from 2010. The MODIS derived ɛsun matches well with the flux calculated ɛsun (R2 = 0.57, p<0.001). Same expression has then been applied over a 217 × 193 km area in Saskatchewan, Canada to obtain the ɛsun and thus GPPsun for the region during the growing season in 2008 (day 150 to day 260). The Vcmax for the region is inverted using the GPPsun and the result is validated at three flux sites inside the area. The results show that the approach is able to obtain good estimations of Vcmax values with R2 = 0.68 and RMSE = 8.8 μmol m-2 s-1.
Echavarría-Heras, Héctor; Leal-Ramírez, Cecilia; Villa-Diharce, Enrique; Castillo, Oscar
2014-01-01
Eelgrass is a cosmopolitan seagrass species that provides important ecological services in coastal and near-shore environments. Despite its relevance, loss of eelgrass habitats is noted worldwide. Restoration by replanting plays an important role, and accurate measurements of the standing crop and productivity of transplants are important for evaluating restoration of the ecological functions of natural populations. Traditional assessments are destructive, and although they do not harm natural populations, in transplants the destruction of shoots might cause undesirable alterations. Non-destructive assessments of the aforementioned variables are obtained through allometric proxies expressed in terms of measurements of the lengths or areas of leaves. Digital imagery could produce measurements of leaf attributes without the removal of shoots, but sediment attachments, damage infringed by drag forces or humidity contents induce noise-effects, reducing precision. Available techniques for dealing with noise caused by humidity contents on leaves use the concepts of adjacency, vicinity, connectivity and tolerance of similarity between pixels. Selection of an interval of tolerance of similarity for efficient measurements requires extended computational routines with tied statistical inferences making concomitant tasks complicated and time consuming. The present approach proposes a simplified and cost-effective alternative, and also a general tool aimed to deal with any sort of noise modifying eelgrass leaves images. Moreover, this selection criterion relies only on a single statistics; the calculation of the maximum value of the Concordance Correlation Coefficient for reproducibility of observed areas of leaves through proxies obtained from digital images. Available data reveals that the present method delivers simplified, consistent estimations of areas of eelgrass leaves taken from noisy digital images. Moreover, the proposed procedure is robust because both the optimal
Mukhopadhyay, Nitai D; Sampson, Andrew J; Deniz, Daniel; Alm Carlsson, Gudrun; Williamson, Jeffrey; Malusek, Alexandr
2012-01-01
Correlated sampling Monte Carlo methods can shorten computing times in brachytherapy treatment planning. Monte Carlo efficiency is typically estimated via efficiency gain, defined as the reduction in computing time by correlated sampling relative to conventional Monte Carlo methods when equal statistical uncertainties have been achieved. The determination of the efficiency gain uncertainty arising from random effects, however, is not a straightforward task specially when the error distribution is non-normal. The purpose of this study is to evaluate the applicability of the F distribution and standardized uncertainty propagation methods (widely used in metrology to estimate uncertainty of physical measurements) for predicting confidence intervals about efficiency gain estimates derived from single Monte Carlo runs using fixed-collision correlated sampling in a simplified brachytherapy geometry. A bootstrap based algorithm was used to simulate the probability distribution of the efficiency gain estimates and the shortest 95% confidence interval was estimated from this distribution. It was found that the corresponding relative uncertainty was as large as 37% for this particular problem. The uncertainty propagation framework predicted confidence intervals reasonably well; however its main disadvantage was that uncertainties of input quantities had to be calculated in a separate run via a Monte Carlo method. The F distribution noticeably underestimated the confidence interval. These discrepancies were influenced by several photons with large statistical weights which made extremely large contributions to the scored absorbed dose difference. The mechanism of acquiring high statistical weights in the fixed-collision correlated sampling method was explained and a mitigation strategy was proposed. Copyright © 2011 Elsevier Ltd. All rights reserved.
Antti T. Kaartinen; Jeremy S. Fried; Paul A. Dunham
2002-01-01
Three Landsat TM-based GIS layers were evaluated as alternatives to conventional, photointerpretation-based stratification of FIA field plots. Estimates for timberland area, timber volume, and volume of down wood were calculated for California's North Coast Survey Unit of 2.5 million hectares. The estimates were compared on the basis of standard errors,...
Energy Technology Data Exchange (ETDEWEB)
Gonzales, John
2015-04-02
Presentation by Senior Engineer John Gonzales on Evaluating Investments in Natural Gas Vehicles and Infrastructure for Your Fleet using the Vehicle Infrastructure Cash-flow Estimation (VICE) 2.0 model.
National Oceanic and Atmospheric Administration, Department of Commerce — A method for estimation of Doppler spectrum, its moments, and polarimetric variables on pulsed weather radars which uses over sampled echo components at a rate...
Energy Technology Data Exchange (ETDEWEB)
Karali, Nihan [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Park, Won Young [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); McNeil, Michael A. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)
2015-06-18
Increasing concerns on non-sustainable energy use and climate change spur a growing research interest in energy efficiency potentials in various critical areas such as industrial production. This paper focuses on learning curve aspects of energy efficiency measures in the U.S iron and steel sector. A number of early-stage efficient technologies (i.e., emerging or demonstration technologies) are technically feasible and have the potential to make a significant contribution to energy saving and CO_{2} emissions reduction, but fall short economically to be included. However, they may also have the cost effective potential for significant cost reduction and/or performance improvement in the future under learning effects such as ‘learning-by-doing’. The investigation is carried out using ISEEM, a technology oriented, linear optimization model. We investigated how steel demand is balanced with/without the availability learning curve, compared to a Reference scenario. The retrofit (or investment in some cases) costs of energy efficient technologies decline in the scenario where learning curve is applied. The analysis also addresses market penetration of energy efficient technologies, energy saving, and CO_{2} emissions in the U.S. iron and steel sector with/without learning impact. Accordingly, the study helps those who use energy models better manage the price barriers preventing unrealistic diffusion of energy-efficiency technologies, better understand the market and learning system involved, predict future achievable learning rates more accurately, and project future savings via energy-efficiency technologies with presence of learning. We conclude from our analysis that, most of the existing energy efficiency technologies that are currently used in the U.S. iron and steel sector are cost effective. Penetration levels increases through the years, even though there is no price reduction. However, demonstration technologies are not economically
Histogram Estimators of Bivariate Densities
National Research Council Canada - National Science Library
Husemann, Joyce A
1986-01-01
One-dimensional fixed-interval histogram estimators of univariate probability density functions are less efficient than the analogous variable-interval estimators which are constructed from intervals...
Browne, William J; Steele, Fiona; Golalizadeh, Mousa; Green, Martin J
2009-06-01
We consider the application of Markov chain Monte Carlo (MCMC) estimation methods to random-effects models and in particular the family of discrete time survival models. Survival models can be used in many situations in the medical and social sciences and we illustrate their use through two examples that differ in terms of both substantive area and data structure. A multilevel discrete time survival analysis involves expanding the data set so that the model can be cast as a standard multilevel binary response model. For such models it has been shown that MCMC methods have advantages in terms of reducing estimate bias. However, the data expansion results in very large data sets for which MCMC estimation is often slow and can produce chains that exhibit poor mixing. Any way of improving the mixing will result in both speeding up the methods and more confidence in the estimates that are produced. The MCMC methodological literature is full of alternative algorithms designed to improve mixing of chains and we describe three reparameterization techniques that are easy to implement in available software. We consider two examples of multilevel survival analysis: incidence of mastitis in dairy cattle and contraceptive use dynamics in Indonesia. For each application we show where the reparameterization techniques can be used and assess their performance.
Nguyen, Van-Dung; Wu, Ling; Noels, Ludovic
2017-03-01
This work provides a unified treatment of arbitrary kinds of microscopic boundary conditions usually considered in the multi-scale computational homogenization method for nonlinear multi-physics problems. An efficient procedure is developed to enforce the multi-point linear constraints arising from the microscopic boundary condition either by the direct constraint elimination or by the Lagrange multiplier elimination methods. The macroscopic tangent operators are computed in an efficient way from a multiple right hand sides linear system whose left hand side matrix is the stiffness matrix of the microscopic linearized system at the converged solution. The number of vectors at the right hand side is equal to the number of the macroscopic kinematic variables used to formulate the microscopic boundary condition. As the resolution of the microscopic linearized system often follows a direct factorization procedure, the computation of the macroscopic tangent operators is then performed using this factorized matrix at a reduced computational time.
Latypov, A. F.
2009-03-01
The fuel economy was estimated at boost trajectory of aerospace plane during energy supply to the free stream. Initial and final velocities of the flight were given. A model of planning flight above cold air in infinite isobaric thermal wake was used. The comparison of fuel consumption was done at optimal trajectories. The calculations were done using a combined power plant consisting of ramjet and liquid-propellant engine. An exergy model was constructed in the first part of the paper for estimating the ramjet thrust and specific impulse. To estimate the aerodynamic drag of aircraft a quadratic dependence on aerodynamic lift is used. The energy for flow heating is obtained at the sacrifice of an equivalent decrease of exergy of combustion products. The dependencies are obtained for increasing the range coefficient of cruise flight at different Mach numbers. In the second part of the paper, a mathematical model is presented for the boost part of the flight trajectory of the flying vehicle and computational results for reducing the fuel expenses at the boost trajectory at a given value of the energy supplied in front of the aircraft.
International Nuclear Information System (INIS)
Niedzwiedz, W.; Cebulska-Wasilewska, A.; Piperakis, S.M.
2000-01-01
The purpose of this study was to examine the individual susceptibility to UV-C induced DNA damage in lymphocytes of Greece people occupationally exposed to pesticides and from reference group with reported no occupational exposure. We also analyzed if there are any differences in the cellular repair capacity between both groups. Lymphocytes were isolated from fresh blood samples collected in Greece from 50 persons recognized as non-exposed to pesticides and from 50 farmers at the end of the spraying season. The average age in exposed to pesticide and reference group was 42.08 and 42.19, respectively. Frozen lymphocytes were transported in a dry ice into DREB laboratory for DNA damage analysis. The DNA damage was measured with the application of single cell gel electrophoresis method (SCGE technique). Our results show that there was not any statistically significant difference concerning the level of the DNA damage detected in defrosted lymphocytes between exposed and non-exposed group. The photoproducts excision efficiency after exposure to UV-C (6 Jm 2 ) and difference in repair capacity by incubation in present and absent of PHA were also studied. There were no statistically significant differences detected directly after UV irradiation between both investigated groups (p >0.1). However, for group exposed to pesticide the ratio of DNA damage measured right after exposition and two hours later was higher (32.19) comparing to reference group (28.60). It may suggest that in exposed group photoproducts excision efficiency was higher or the rejoining rates of the breaks was lower. The differences between repair efficiency observed in lymphocytes from group exposed and non-exposed to pesticides (with or without stimulation to division) were also statistically insignificant (for Tail Length, Tail DNA and Tail moment parameters - p >0.1). Statistically significant differences in DNA damage repair capacities were observed (for all analyzed parameters) between lymphocytes
El Gharamti, Mohamad
2016-11-15
This study considers the assimilation problem of subsurface contaminants at the port of Rotterdam in the Netherlands. It involves the estimation of solute concentrations and biodegradation rates of four different chlorinated solvents. We focus on assessing the efficiency of an adaptive hybrid ensemble Kalman filter and optimal interpolation (EnKF-OI) and the exact second-order sampling formulation (EnKFESOS) for mitigating the undersampling of the estimation and observation errors covariances, respectively. A multi-dimensional and multi-species reactive transport model is coupled to simulate the migration of contaminants within a Pleistocene aquifer layer located around 25 m below mean sea level. The biodegradation chain of chlorinated hydrocarbons starting from tetrachloroethene and ending with vinyl chloride is modeled under anaerobic environmental conditions for 5 decades. Yearly pseudo-concentration data are used to condition the forecast concentration and degradation rates in the presence of model and observational errors. Assimilation results demonstrate the robustness of the hybrid EnKF-OI, for accurately calibrating the uncertain biodegradation rates. When implemented serially, the adaptive hybrid EnKF-OI scheme efficiently adjusts the weights of the involved covariances for each individual measurement. The EnKFESOS is shown to maintain the parameter ensemble spread much better leading to more robust estimates of the states and parameters. On average, a well tuned hybrid EnKF-OI and the EnKFESOS respectively suggest around 48 and 21 % improved concentration estimates, as well as around 70 and 23 % improved anaerobic degradation rates, over the standard EnKF. Incorporating large uncertainties in the flow model degrades the accuracy of the estimates of all schemes. Given that the performance of the hybrid EnKF-OI depends on the quality of the background statistics, satisfactory results were obtained only when the uncertainty imposed on the background
International Nuclear Information System (INIS)
Stuehrenberg, L.; Smith, D.; Johnson, O.W.
1995-10-01
Populations of salmonid smolts migrating through the hydropower system on the Columbia River incur some rate of mortality at each dam. To set priorities on options to minimize losses and provide safe passage of the smolts at dams, estimates of smolt survival at each dam are necessary. Two methods have been developed to obtain these survival estimates: the direct and the indirect method. With the indirect method, a test group of fish is released upstream and a-control group is released downstream from the area of interest. With the direct method, a single release of fish above the area of interest is used, with subsequent recovery below the area of interest. In 1988, the National Marine Fisheries Service (NMFS) began a 2-year study at McNary Dam to address possible sources of variation associated with the direct method of obtaining survival estimates. Five study objectives were established to determine whether (1) fish from the Columbia and Snake Rivers mixed as they migrated to McNary Dam (release-location tests); (2) collection rates for Columbia and Snake River stocks were the same (river-of-origin tests); (3) test-group release timing influenced recovery rates (time-of-release tests); (4) a collection-rate bias existed from use of test fish previously guided and collected at the recovery site (tests of previously guided fish); and (5) recovery rates obtained with PIT-tagged fish were comparable to those previously obtained with freeze-branded fish (PIT-tag vs. freeze-brand technology)
Energy Technology Data Exchange (ETDEWEB)
Gonzalez Arzola, K. [Department of Microbiology and Cell Biology, Faculty of Pharmacy, University of La Laguna, 38206 La Laguna, Tenerife (Spain); Arevalo, M.C. [Department of Physical Chemistry, Faculty of Chemistry, University of La Laguna, 38206 La Laguna, Tenerife (Spain)], E-mail: carevalo@ull.es; Falcon, M.A. [Department of Microbiology and Cell Biology, Faculty of Pharmacy, University of La Laguna, 38206 La Laguna, Tenerife (Spain)], E-mail: mafalcon@ull.es
2009-03-30
The electrochemical properties of eighteen natural and synthetic compounds commonly used to expand the oxidative capacity of laccases were evaluated in an aqueous buffered medium using cyclic voltammetry. This clarifies which compounds fulfil the requisites to be considered as redox mediators or enhancers. Cyclic voltammetry was also applied as a rapid way to assess the catalytic efficiency (CE) of those compounds which oxidise a non-phenolic lignin model (veratryl alcohol, VA) and a kraft lignin (KL). With the exception of gallic acid and catechol, all assayed compounds were capable of oxidising VA with varying CE. However, only some of them were able to oxidise KL. Although the oxidised forms of HBT and acetovanillone were not electrochemically stable, their reduced forms were quickly regenerated in the presence of VA. They thus act as chemical catalysts. Importantly, HBT and HPI did not attack the KL via the same mechanism as in VA oxidation. Electrochemical evidence suggests that violuric acid oxidises both substrates by an electron transfer mechanism, unlike the other N-OH compounds HBT and HPI. Acetovanillone was found to be efficient in oxidising VA and KL, even better than the synthetic mediators TEMPO, violuric acid or ABTS. Most of the compounds produced a generalised increase in the oxidative charge of KL, probably attributed to chain reactions arising between the phenolic and non-phenolic components of this complex molecule.
Rumosa Gwaze, Francisca; Chimonyo, Michael; Dzama, Kennedy
2010-08-01
It is often inappropriately assumed that the contributions of goats to household economies are similar across flock sizes. A monthly questionnaire was administered to 19, 21, and 13 farmers in Matatiele and Qawukeni and Mankone. Entries and exits to and from each flock were assessed. Most (P < 0.05) kids were born in August (4.92 +/- 0.362 and 2.03 +/- 0.329 for large and small flocks, respectively). Total entries were higher (P < 0.05) in August, September, and October than in the other months. Kid mortality peaked in May and September at 21%. Most (P < 0.05) goats were sold in December (1.06 +/- 0.127 for large flocks and 0.23 +/- 0.135 for small flocks). Goat production efficiency (GPE) was affected (P < 0.05) by month and flock size. The GPE was higher (P < 0.05) in small than in large flocks. The low GPE values obtained in this study indicates the need to devise other indices that factor in the neglected functions of goats when measuring goat production efficiency in communities where the main function of goats is not for sale or consumption.
International Nuclear Information System (INIS)
Gonzalez Arzola, K.; Arevalo, M.C.; Falcon, M.A.
2009-01-01
The electrochemical properties of eighteen natural and synthetic compounds commonly used to expand the oxidative capacity of laccases were evaluated in an aqueous buffered medium using cyclic voltammetry. This clarifies which compounds fulfil the requisites to be considered as redox mediators or enhancers. Cyclic voltammetry was also applied as a rapid way to assess the catalytic efficiency (CE) of those compounds which oxidise a non-phenolic lignin model (veratryl alcohol, VA) and a kraft lignin (KL). With the exception of gallic acid and catechol, all assayed compounds were capable of oxidising VA with varying CE. However, only some of them were able to oxidise KL. Although the oxidised forms of HBT and acetovanillone were not electrochemically stable, their reduced forms were quickly regenerated in the presence of VA. They thus act as chemical catalysts. Importantly, HBT and HPI did not attack the KL via the same mechanism as in VA oxidation. Electrochemical evidence suggests that violuric acid oxidises both substrates by an electron transfer mechanism, unlike the other N-OH compounds HBT and HPI. Acetovanillone was found to be efficient in oxidising VA and KL, even better than the synthetic mediators TEMPO, violuric acid or ABTS. Most of the compounds produced a generalised increase in the oxidative charge of KL, probably attributed to chain reactions arising between the phenolic and non-phenolic components of this complex molecule
Liao, Hstau Y; Hashem, Yaser; Frank, Joachim
2015-06-02
Single-particle cryogenic electron microscopy (cryo-EM) is a powerful tool for the study of macromolecular structures at high resolution. Classification allows multiple structural states to be extracted and reconstructed from the same sample. One classification approach is via the covariance matrix, which captures the correlation between every pair of voxels. Earlier approaches employ computing-intensive resampling and estimate only the eigenvectors of the matrix, which are then used in a separate fast classification step. We propose an iterative scheme to explicitly estimate the covariance matrix in its entirety. In our approach, the flexibility in choosing the solution domain allows us to examine a part of the molecule in greater detail. Three-dimensional covariance maps obtained in this way from experimental data (cryo-EM images of the eukaryotic pre-initiation complex) prove to be in excellent agreement with conclusions derived by using traditional approaches, revealing in addition the interdependencies of ligand bindings and structural changes. Copyright © 2015 Elsevier Ltd. All rights reserved.
Borghi, Giacomo; Tabacchini, Valerio; Seifert, Stefan; Schaart, Dennis R.
2015-02-01
Monolithic scintillator detectors can achieve excellent spatial resolution and coincidence resolving time. However, their practical use for positron emission tomography (PET) and other applications in the medical imaging field is still limited due to drawbacks of the different methods used to estimate the position of interaction. Common statistical methods for example require the collection of an extensive dataset of reference events with a narrow pencil beam aimed at a fine grid of reference positions. Such procedures are time consuming and not straightforwardly implemented in systems composed of many detectors. Here, we experimentally demonstrate for the first time a new calibration procedure for k-nearest neighbor ( k-NN) position estimation that utilizes reference data acquired with a fan beam. The procedure is tested on two detectors consisting of 16 mm ×16 mm ×10 mm and 16 mm ×16 mm ×20 mm monolithic, Ca-codoped LSO:Ce crystals and digital photon counter (DPC) arrays. For both detectors, the spatial resolution and the bias obtained with the new method are found to be practically the same as those obtained with the previously used method based on pencil-beam irradiation, while the calibration time is reduced by a factor of 20. Specifically, a FWHM of 1.1 mm and a FWTM of 2.7 mm were obtained using the fan-beam method with the 10 mm crystal, whereas a FWHM of 1.5 mm and a FWTM of 6 mm were achieved with the 20 mm crystal. Using a fan beam made with a 4.5 MBq 22Na point-source and a tungsten slit collimator with 0.5 mm aperture, the total measurement time needed to acquire the reference dataset was 3 hours for the thinner crystal and 2 hours for the thicker one.
Zhu, Feng; Hu, Xiaofeng; He, Xiaoyuan; Guo, Rui; Li, Kaiming; Yang, Lu
2017-11-01
In the military field, the performance evaluation of early-warning aircraft deployment or construction is always an important problem needing to be explored. As an effective approach of enterprise management and performance evaluation, Balanced Score Card (BSC) attracts more and more attentions and is studied more and more widely all over the world. It can also bring feasible ideas and technical approaches for studying the issue of the performance evaluation of the deployment or construction of early-warning aircraft which is the important component in early-warning detection system of systems (SoS). Therefore, the deep explored researches are carried out based on the previously research works. On the basis of the characteristics of space exploration and aerial detection effectiveness of early-warning detection SoS and the cardinal principle of BSC are analyzed simply, and the performance evaluation framework of the deployment or construction of early-warning aircraft is given, under this framework, aimed at the evaluation issue of aerial detection effectiveness of early-warning detection SoS with the cooperation efficiency factors of the early-warning aircraft and other land based radars, the evaluation indexes are further designed and the relative evaluation model is further established, especially the evaluation radar chart being also drawn to obtain the evaluation results from a direct sight angle. Finally, some practical computer simulations are launched to prove the validity and feasibility of the research thinking and technologic approaches which are proposed in the paper.
Directory of Open Access Journals (Sweden)
Rumana Aslam
2017-07-01
Full Text Available In the present investigation healthy and certified seeds of Capsicum annuum were treated with five concentrations of caffeine i.e. 0.10%, 0.25%, 0.50%, 0.75% and 1.0%. Germination percentage, plants survival and pollen fertility were decreased with the increase of caffeine concentrations. Similarly root length and shoot length were decreased as the concentrations increased in M1 generation. Different mutants were isolated in M1 generation. In M2 generation, various flower mutants with changes in number of sepals, petals, anther size colour i.e. Trimerous, tetramerous, pentamerous with fused petals, hexamerous etc were segregated. Heptamerous and anther change was not observed in lower concentration viz. 0.1%. All these mutants showed significant changes in morphological characters and good breeding values at lower and intermediate concentrations. Mutagenic effectiveness and efficiency was observed on the basis of M2 flower mutant frequency. It was generally decreased with the increase of mutagen concentrations. Cytological aberrations in mutants showed the decreasing trend at meiotic final stages. These mutants were further analysed through RAPD method and on the basis of appearance of polymorphic DNA bands, they distinguished these flower mutants genotypically. Among 93 bands 44 bands were polymorphic which showed great genetic variation produced by caffeine. As an outcome of that the above caffeine concentrations are good for the induction of genetic variability in Capsicum genotype.
International Nuclear Information System (INIS)
Takasu, Miyuki; Tani, Chihiro; Sakoda, Yasuko; Ishikawa, Miho; Tanitame, Keizo; Date, Shuji; Akiyama, Yuji; Awai, Kazuo; Sakai, Akira; Asaoku, Hideki; Kajima, Toshio
2012-01-01
To evaluate the effectiveness of the iterative decomposition of water and fat with echo asymmetric and least-squares estimation (IDEAL) MRI to quantify tumour infiltration into the lumbar vertebrae in myeloma patients without visible focal lesions. The lumbar spine was examined with 3 T MRI in 24 patients with multiple myeloma and in 26 controls. The fat-signal fraction was calculated as the mean value from three vertebral bodies. A post hoc test was used to compare the fat-signal fraction in controls and patients with monoclonal gammopathy of undetermined significance (MGUS), asymptomatic myeloma or symptomatic myeloma. Differences were considered significant at P 2 -microglobulin-to-albumin ratio were entered into the discriminant analysis. Fat-signal fractions were significantly lower in patients with symptomatic myelomas (43.9 ±19.7%, P 2 -microglobulin-to-albumin ratio facilitated discrimination of symptomatic myeloma from non-symptomatic myeloma in patients without focal bone lesions. circle A new magnetic resonance technique (IDEAL) offers new insights in multiple myeloma. (orig.)
Dahm, T.; Heimann, S.; Isken, M.; Vasyura-Bathke, H.; Kühn, D.; Sudhaus, H.; Kriegerowski, M.; Daout, S.; Steinberg, A.; Cesca, S.
2017-12-01
Seismic source and moment tensor waveform inversion is often ill-posed or non-unique if station coverage is poor or signals are weak. Therefore, the interpretation of moment tensors can become difficult, if not the full model space is explored, including all its trade-offs and uncertainties. This is especially true for non-double couple components of weak or shallow earthquakes, as for instance found in volcanic, geothermal or mining environments.We developed a bootstrap-based probabilistic optimization scheme (Grond), which is based on pre-calculated Greens function full waveform databases (e.g. fomosto tool, doi.org/10.5880/GFZ.2.1.2017.001). Grond is able to efficiently explore the full model space, the trade-offs and the uncertainties of source parameters. The program is highly flexible with respect to the adaption to specific problems, the design of objective functions, and the diversity of empirical datasets.It uses an integrated, robust waveform data processing based on a newly developed Python toolbox for seismology (Pyrocko, see Heimann et al., 2017, http://doi.org/10.5880/GFZ.2.1.2017.001), and allows for visual inspection of many aspects of the optimization problem. Grond has been applied to the CMT moment tensor inversion using W-phases, to nuclear explosions in Korea, to meteorite atmospheric explosions, to volcano-tectonic events during caldera collapse and to intra-plate volcanic and tectonic crustal events.Grond can be used to optimize simultaneously seismological waveforms, amplitude spectra and static displacements of geodetic data as InSAR and GPS (e.g. KITE, Isken et al., 2017, http://doi.org/10.5880/GFZ.2.1.2017.002). We present examples of Grond optimizations to demonstrate the advantage of a full exploration of source parameter uncertainties for interpretation.
Dimitriadis, Stavros I; Salis, Christos; Linden, David
2018-04-01
Limitations of the manual scoring of polysomnograms, which include data from electroencephalogram (EEG), electro-oculogram (EOG), electrocardiogram (ECG) and electromyogram (EMG) channels have long been recognized. Manual staging is resource intensive and time consuming, and thus considerable effort must be spent to ensure inter-rater reliability. As a result, there is a great interest in techniques based on signal processing and machine learning for a completely Automatic Sleep Stage Classification (ASSC). In this paper, we present a single-EEG-sensor ASSC technique based on the dynamic reconfiguration of different aspects of cross-frequency coupling (CFC) estimated between predefined frequency pairs over 5 s epoch lengths. The proposed analytic scheme is demonstrated using the PhysioNet Sleep European Data Format (EDF) Database with repeat recordings from 20 healthy young adults. We validate our methodology in a second sleep dataset. We achieved very high classification sensitivity, specificity and accuracy of 96.2 ± 2.2%, 94.2 ± 2.3%, and 94.4 ± 2.2% across 20 folds, respectively, and also a high mean F1 score (92%, range 90-94%) when a multi-class Naive Bayes classifier was applied. High classification performance has been achieved also in the second sleep dataset. Our method outperformed the accuracy of previous studies not only on different datasets but also on the same database. Single-sensor ASSC makes the entire methodology appropriate for longitudinal monitoring using wearable EEG in real-world and laboratory-oriented environments. Crown Copyright © 2018. Published by Elsevier B.V. All rights reserved.
Zhang, Qingyuan; Middleton, Elizabeth M.; Margolis, Hank A.; Drolet, Guillaume G.; Barr, Alan A.; Black, T. Andrew
2009-01-01
Gross primary production (GPP) is a key terrestrial ecophysiological process that links atmospheric composition and vegetation processes. Study of GPP is important to global carbon cycles and global warming. One of the most important of these processes, plant photosynthesis, requires solar radiation in the 0.4-0.7 micron range (also known as photosynthetically active radiation or PAR), water, carbon dioxide (CO2), and nutrients. A vegetation canopy is composed primarily of photosynthetically active vegetation (PAV) and non-photosynthetic vegetation (NPV; e.g., senescent foliage, branches and stems). A green leaf is composed of chlorophyll and various proportions of nonphotosynthetic components (e.g., other pigments in the leaf, primary/secondary/tertiary veins, and cell walls). The fraction of PAR absorbed by whole vegetation canopy (FAPAR(sub canopy)) has been widely used in satellite-based Production Efficiency Models to estimate GPP (as a product of FAPAR(sub canopy)x PAR x LUE(sub canopy), where LUE(sub canopy) is light use efficiency at canopy level). However, only the PAR absorbed by chlorophyll (a product of FAPAR(sub chl) x PAR) is used for photosynthesis. Therefore, remote sensing driven biogeochemical models that use FAPAR(sub chl) in estimating GPP (as a product of FAPAR(sub chl x PAR x LUE(sub chl) are more likely to be consistent with plant photosynthesis processes.
Directory of Open Access Journals (Sweden)
O. N. Korsun
2014-01-01
Full Text Available High information load of crew is one of the main problems of modern piloted aircraft therefore researches on approving data representation form, especially in critical situations are a challenge. The article considers one of opportunities to improve the interface of a modern pilot's cabin i.e. to use a spatial sound (3D - audio technology. The 3D - audio is a technology, which recreates a spatially directed sound in earphones or via loudspeakers. Spatial audio-helps, which together with information on danger will specify also the direction from which it proceeds, can reduce time of response to an event and, therefore, increase situational safety of flight. It is supposed that helps will be provided through pilot's headset therefore technology realization via earphones is discussed.Now the main hypothesis explaining the human ability to recognize the position of a sound source in space, asserts that the human estimates distortion of a sound signal spectrum at interaction with the head and an auricle depending on an arrangement of the sound source. For exact describing the signal spectrum variations there are such concepts as Head Related Impulse Response (HRIR and Head Related Transfer Function (HRTF. HRIR is measured in humans or dummies. At present the most full-scale public HRIR library is CIPIC HRTF Database of CIPIC Interface Laboratory at UC Davis.To have 3D audio effect, it is necessary to simulate a mono-signal conversion through the linear digital filters with anthropodependent pulse characteristics (HRIR for the left and right ear, which correspond to the chosen direction. Results should be united in a stereo file and applied for reproduction to the earphones.This scheme was realized in Matlab, and the received software was used for experiments to estimate the quantitative characteristics of technology. For processing and subsequent experiments the following sound signals were chosen: a fragment of the classical music piece "Polovetsky
International Nuclear Information System (INIS)
Katz, M.
1995-01-01
Whichever way the local distribution company (LDC) tries to convert residential customers to gas or expand their use of it, the process itself has become essential for the natural gas industry. The amount of gas used by each residential customer has been decreasing for 25 years -- since the energy crisis of the early 1970s. It's a direct result of better-insulated homes and more-efficient gas appliances, and that trend is continuing. So, LDCs have a choice of either finding new users and uses for gas, or recognizing that their throughput per customer is going to continue declining. The paper discusses strategies that several gas utilities are using to increase the number of gas appliances in the customer's homes. These and other strategies keep the gas industry optimistic about the future of the residential market: A.G.A. has projected that by 2010 demand will expand, from 1994's 5.1 quadrillion Btu (quads) to 5.7 quads, even with continued improvements in appliance efficiency. That estimate, however, will depend on the industry-s utilities and whether they keep converting, proselytizing, persuading and influencing customers to use more natural gas
Virtual Sensors: Efficiently Estimating Missing Spectra
National Aeronautics and Space Administration — Various instruments are used to create images of the Earth and other objects in the universe in a diverse set of wavelength bands with the aim of understanding...
How efficient is estimation with missing data?
DEFF Research Database (Denmark)
Karadogan, Seliz; Marchegiani, Letizia; Hansen, Lars Kai
2011-01-01
percentages (MDP) using a missing completely at random (MCAR) scheme. We compare three MDTs: pairwise deletion (PW), mean imputation (MI) and a maximum likelihood method that we call complete expectation maximization (CEM). We use a synthetic dataset, the Iris dataset and the Pima Indians Diabetes dataset. We...
Efficient nonparametric estimation of causal mediation effects
Chan, K. C. G.; Imai, K.; Yam, S. C. P.; Zhang, Z.
2016-01-01
An essential goal of program evaluation and scientific research is the investigation of causal mechanisms. Over the past several decades, causal mediation analysis has been used in medical and social sciences to decompose the treatment effect into the natural direct and indirect effects. However, all of the existing mediation analysis methods rely on parametric modeling assumptions in one way or another, typically requiring researchers to specify multiple regression models involving the treat...
DEFF Research Database (Denmark)
Jørgensen, Ivan Harald Holger; Bogason, Gudmundur; Bruun, Erik
1995-01-01
This paper proposes a new way to estimate the flow in a micromechanical flow channel. A neural network is used to estimate the delay of random temperature fluctuations induced in a fluid. The design and implementation of a hardware efficient neural flow estimator is described. The system...... is implemented using switched-current technique and is capable of estimating flow in the μl/s range. The neural estimator is built around a multiplierless neural network, containing 96 synaptic weights which are updated using the LMS1-algorithm. An experimental chip has been designed that operates at 5 V...
Efficiency model of Russian banks
Pavlyuk, Dmitry
2006-01-01
The article deals with problems related to the stochastic frontier model of bank efficiency measurement. The model is used to study the efficiency of the banking sector of The Russian Federation. It is based on the stochastic approach both to the efficiency frontier location and to individual bank efficiency values. The model allows estimating bank efficiency values, finding relations with different macro- and microeconomic factors and testing some economic hypotheses.
Efficiency in Microfinance Cooperatives
Directory of Open Access Journals (Sweden)
HARTARSKA, Valentina
2012-12-01
Full Text Available In recognition of cooperatives’ contribution to the socio-economic well-being of their participants, the United Nations has declared 2012 as the International Year of Cooperatives. Microfinance cooperatives make a large part of the microfinance industry. We study efficiency of microfinance cooperatives and provide estimates of the optimal size of such organizations. We employ the classical efficiency analysis consisting of estimating a system of equations and identify the optimal size of microfinance cooperatives in terms of their number of clients (outreach efficiency, as well as dollar value of lending and deposits (sustainability. We find that microfinance cooperatives have increasing returns to scale which means that the vast majority can lower cost if they become larger. We calculate that the optimal size is around $100 million in lending and half of that in deposits. We find less robust estimates in terms of reaching many clients with a range from 40,000 to 180,000 borrowers.
International Nuclear Information System (INIS)
2010-01-01
After a speech of the CEA's (Commissariat a l'Energie Atomique) general administrator about energy efficiency as a first rank challenge for the planet and for France, this publications proposes several contributions: a discussion of the efficiency of nuclear energy, an economic analysis of R and D's value in the field of fourth generation fast reactors, discussions about biofuels and the relationship between energy efficiency and economic competitiveness, and a discussion about solar photovoltaic efficiency
Energy Technology Data Exchange (ETDEWEB)
Asociacion de Tecnicos y Profesionistas en Aplicacion Energetica, A.C. [Mexico (Mexico)
2002-06-01
In the last years much attention has been given to the polluting gas discharges, in special of those that favor the green house effect (GHE), due to the negative sequels that its concentration causes to the atmosphere, particularly as the cause of the increase in the overall temperature of the planet, which has been denominated world-wide climatic change. There are many activities that allow to lessen or to elude the GHE gas emissions, and with the main ones the so-called projects of Energy Efficiency and Renewable Energy (EE/RE) have been structured. In order to carry out a project within the frame of the MDL, it is necessary to evaluate with quality, precision and transparency, the amount of emissions of GHE gases that are reduced or suppressed thanks to their application. For that reason, in our country we tried different methodologies directed to estimate the CO{sub 2} emissions that are attenuated or eliminated by means of the application of EE/RE projects. [Spanish] En los ultimos anos se ha puesto mucha atencion a las emisiones de gases contaminantes, en especial de los que favorecen el efecto invernadero (GEI), debido a las secuelas negativas que su concentracion ocasiona a la atmosfera, particularmente como causante del aumento en la temperatura general del planeta, en lo que se ha denominado cambio climatico mundial. Existen muchas actividades que permiten aminorar o eludir las emisiones de GEI, y con las principales se han estructurado los llamados proyectos de eficiencia energetica y energia renovables (EE/ER). Para llevar a cabo un proyecto dentro del marco del MDL, es necesario evaluar con calidad, precision y transparencia, la cantidad de emisiones de GEI que se reducen o suprimen gracias a su aplicacion. Por ello, en nuestro pais ensayamos diferentes metodologias encaminadas a estimar las emisiones de CO{sub 2} que se atenuan o eliminan mediante la aplicacion de proyectos de EE/ER.
Karacan, C Özgen
2013-07-30
Coal seam degasification and its efficiency are directly related to the safety of coal mining. Degasification activities in the Black Warrior basin started in the early 1980s by using vertical boreholes. Although the Blue Creek seam, which is part of the Mary Lee coal group, has been the main seam of interest for coal mining, vertical wellbores have also been completed in the Pratt, Mary Lee, and Black Creek coal groups of the Upper Pottsville formation to degasify multiple seams. Currently, the Blue Creek seam is further degasified 2-3 years in advance of mining using in-seam horizontal boreholes to ensure safe mining. The studied location in this work is located between Tuscaloosa and Jefferson counties in Alabama and was degasified using 81 vertical boreholes, some of which are still active. When the current long mine expanded its operation into this area in 2009, horizontal boreholes were also drilled in advance of mining for further degasification of only the Blue Creek seam to ensure a safe and a productive operation. This paper presents an integrated study and a methodology to combine history matching results from vertical boreholes with production modeling of horizontal boreholes using geostatistical simulation to evaluate spatial effectiveness of in-seam boreholes in reducing gas-in-place (GIP). Results in this study showed that in-seam wells' boreholes had an estimated effective drainage area of 2050 acres with cumulative production of 604 MMscf methane during ~2 years of operation. With horizontal borehole production, GIP in the Blue Creek seam decreased from an average of 1.52 MMscf to 1.23 MMscf per acre. It was also shown that effective gas flow capacity, which was independently modeled using vertical borehole data, affected horizontal borehole production. GIP and effective gas flow capacity of coal seam gas were also used to predict remaining gas potential for the Blue Creek seam.
DEFF Research Database (Denmark)
Sales-Cruz, Mauricio; Heitzig, Martina; Cameron, Ian
2011-01-01
of optimisation techniques coupled with dynamic solution of the underlying model. Linear and nonlinear approaches to parameter estimation are investigated. There is also the application of maximum likelihood principles in the estimation of parameters, as well as the use of orthogonal collocation to generate a set......In this chapter the importance of parameter estimation in model development is illustrated through various applications related to reaction systems. In particular, rate constants in a reaction system are obtained through parameter estimation methods. These approaches often require the application...... of algebraic equations as the basis for parameter estimation.These approaches are illustrated using estimations of kinetic constants from reaction system models....
Multi-directional program efficiency
DEFF Research Database (Denmark)
Asmild, Mette; Balezentis, Tomas; Hougaard, Jens Leth
2016-01-01
The present paper analyses both managerial and program efficiencies of Lithuanian family farms, in the tradition of Charnes et al. (Manag Sci 27(6):668–697, 1981) but with the important difference that multi-directional efficiency analysis rather than the traditional data envelopment analysis...... approach is used to estimate efficiency. This enables a consideration of input-specific efficiencies. The study shows clear differences between the efficiency scores on the different inputs as well as between the farm types of crop, livestock and mixed farms respectively. We furthermore find that crop...... farms have the highest program efficiency, but the lowest managerial efficiency and that the mixed farms have the lowest program efficiency (yet not the highest managerial efficiency)....
Improving efficiency in stereology
DEFF Research Database (Denmark)
Keller, Kresten Krarup; Andersen, Ina Trolle; Andersen, Johnnie Bremholm
2013-01-01
of the study was to investigate the time efficiency of the proportionator and the autodisector on virtual slides compared with traditional methods in a practical application, namely the estimation of osteoclast numbers in paws from mice with experimental arthritis and control mice. Tissue slides were scanned......, a proportionator sampling and a systematic, uniform random sampling were simulated. We found that the proportionator was 50% to 90% more time efficient than systematic, uniform random sampling. The time efficiency of the autodisector on virtual slides was 60% to 100% better than the disector on tissue slides. We...... conclude that both the proportionator and the autodisector on virtual slides may improve efficiency of cell counting in stereology....
DEFF Research Database (Denmark)
Andersen, Rikke Sand; Vedsted, Peter
2015-01-01
on institutional logics, we illustrate how a logic of efficiency organise and give shape to healthcare seeking practices as they manifest in local clinical settings. Overall, patient concerns are reconfigured to fit the local clinical setting and healthcare professionals and patients are required to juggle...... efficiency in order to deal with uncertainties and meet more complex or unpredictable needs. Lastly, building on the empirical case of cancer diagnostics, we discuss the implications of the pervasiveness of the logic of efficiency in the clinical setting and argue that provision of medical care in today......'s primary care settings requires careful balancing of increasing demands of efficiency, greater complexity of biomedical knowledge and consideration for individual patient needs....
Energy efficiency in Swedish industry
International Nuclear Information System (INIS)
Zhang, Shanshan; Lundgren, Tommy; Zhou, Wenchao
2016-01-01
This paper assesses energy efficiency in Swedish industry. Using unique firm-level panel data covering the years 2001–2008, the efficiency estimates are obtained for firms in 14 industrial sectors by using data envelopment analysis (DEA). The analysis accounts for multi-output technologies where undesirable outputs are produced alongside with the desirable output. The results show that there was potential to improve energy efficiency in all the sectors and relatively large energy inefficiencies existed in small energy-use industries in the sample period. Also, we assess how the EU ETS, the carbon dioxide (CO_2) tax and the energy tax affect energy efficiency by conducting a second-stage regression analysis. To obtain consistent estimates for the regression model, we apply a modified, input-oriented version of the double bootstrap procedure of Simar and Wilson (2007). The results of the regression analysis reveal that the EU ETS and the CO_2 tax did not have significant influences on energy efficiency in the sample period. However, the energy tax had a positive relation with the energy efficiency. - Highlights: • We use DEA to estimate firm-level energy efficiency in Swedish industry. • We examine impacts of climate and energy policies on energy efficiency. • The analyzed policies are Swedish carbon and energy taxes and the EU ETS. • Carbon tax and EU ETS did not have significant influences on energy efficiency. • The energy tax had a positive relation with energy efficiency.
Energy Technology Data Exchange (ETDEWEB)
Schwickerath, Ulrich; Silva, Ricardo; Uria, Christian, E-mail: Ulrich.Schwickerath@cern.c, E-mail: Ricardo.Silva@cern.c [CERN IT, 1211 Geneve 23 (Switzerland)
2010-04-01
A frequent source of concern for resource providers is the efficient use of computing resources in their centers. This has a direct impact on requests for new resources. There are two different but strongly correlated aspects to be considered: while users are mostly interested in a good turn-around time for their jobs, resource providers are mostly interested in a high and efficient usage of their available resources. Both things, the box usage and the efficiency of individual user jobs, need to be closely monitored so that the sources of the inefficiencies can be identified. At CERN, the Lemon monitoring system is used for both purposes. Examples of such sources are poorly written user code, inefficient access to mass storage systems, and dedication of resources to specific user groups. As a first step for improvements CERN has launched a project to develop a scheduler add-on that allows careful overloading of worker nodes that run idle jobs.
Coherence in quantum estimation
Giorda, Paolo; Allegra, Michele
2018-01-01
The geometry of quantum states provides a unifying framework for estimation processes based on quantum probes, and it establishes the ultimate bounds of the achievable precision. We show a relation between the statistical distance between infinitesimally close quantum states and the second order variation of the coherence of the optimal measurement basis with respect to the state of the probe. In quantum phase estimation protocols, this leads to propose coherence as the relevant resource that one has to engineer and control to optimize the estimation precision. Furthermore, the main object of the theory i.e. the symmetric logarithmic derivative, in many cases allows one to identify a proper factorization of the whole Hilbert space in two subsystems. The factorization allows one to discuss the role of coherence versus correlations in estimation protocols; to show how certain estimation processes can be completely or effectively described within a single-qubit subsystem; and to derive lower bounds for the scaling of the estimation precision with the number of probes used. We illustrate how the framework works for both noiseless and noisy estimation procedures, in particular those based on multi-qubit GHZ-states. Finally we succinctly analyze estimation protocols based on zero-temperature critical behavior. We identify the coherence that is at the heart of their efficiency, and we show how it exhibits the non-analyticities and scaling behavior proper of a large class of quantum phase transitions.
El Gharamti, Mohamad; Valstar, Johan; Janssen, Gijs; Marsman, Annemieke; Hoteit, Ibrahim
2016-01-01
This study considers the assimilation problem of subsurface contaminants at the port of Rotterdam in the Netherlands. It involves the estimation of solute concentrations and biodegradation rates of four different chlorinated solvents. We focus
Economics of appliance efficiency
International Nuclear Information System (INIS)
Tiedemann, K.H.
2009-01-01
Several significant developments occurred in 2001 that affect the impact of market transformation programs. This paper presented and applied an econometric approach to the identification and estimation of market models for refrigerators, clothes washers, dishwashers and room air conditioners. The purpose of the paper was to understand the impact of energy conservation policy developments on sales of energy efficient appliances. The paper discussed the approach with particular reference to building a database of sales and drivers of sales using publicly available information; estimation of the determinants of sales using econometric models; and estimation of the individual impacts of prices, gross domestic product (GDP) and energy conservation policies on sales using regression results. Market and policy developments were also presented, such as change a light, save the world promotion; the California energy crisis; and the Pacific Northwest drought induced hydro power shortage. It was concluded that an increase in GDP increased the sales of both more efficient and less efficient refrigerators, clothes washers, dishwashers, and room air conditioners. An increase in electricity price increased sales of Energy Star refrigerators, clothes washers, dishwashers, and room air conditioners. 4 refs., 8 tabs.
Czech Academy of Sciences Publication Activity Database
Mukhopadhyay, N. D.; Sampson, A. J.; Deniz, D.; Carlsson, G. A.; Williamson, J.; Malušek, Alexandr
2012-01-01
Roč. 70, č. 1 (2012), s. 315-323 ISSN 0969-8043 Institutional research plan: CEZ:AV0Z10480505 Keywords : Monte Carlo * correlated sampling * efficiency * uncertainty * bootstrap Subject RIV: BG - Nuclear, Atomic and Molecular Physics, Colliders Impact factor: 1.179, year: 2012 http://www.sciencedirect.com/science/article/pii/S0969804311004775
International Nuclear Information System (INIS)
Marvillet, Ch.; Tochon, P.; Mercier, P.
2004-01-01
World energy demand is constantly rising. This is a legitimate trend, insofar as access to energy enables enhanced quality of life and sanitation levels for populations. On the other hand, such increased consumption generates effects that may be catastrophic for the future of the planet (climate change, environmental imbalance), should this growth conform to the patterns followed, up to recent times, by most industrialized countries. Reduction of greenhouse gas emissions, development of new energy sources and energy efficiency are seen as the major challenges to be taken up for the world of tomorrow. In France, the National Energy Debate indeed emphasized, in 2003, the requirement to control both demand for, and offer of, energy, through a strategic orientation law for energy. The French position corresponds to a slightly singular situation - and a privileged one, compared to other countries - owing to massive use of nuclear power for electricity generation. This option allows France to be responsible for a mere 2% of worldwide greenhouse gas emissions. Real advances can nonetheless still be achieved as regards improved energy efficiency, particularly in the transportation and residential-tertiary sectors, following the lead, in this respect, shown by industry. These two sectors indeed account for over half of the country CO 2 emissions (26% and 25% respectively). With respect to transportation, the work carried out by CEA on the hydrogen pathway, energy converters, and electricity storage has been covered by the preceding chapters. As regards housing, a topic addressed by one of the papers in this chapter, investigations at CEA concern integration of the various devices enabling value-added use of renewable energies. At the same time, the organization is carrying through its activity in the extensive area of heat exchangers, allowing industry to benefit from improved understanding in the modeling of flows. An activity evidenced by advances in energy efficiency for
Efficiency in the Worst Production Situation Using Data Envelopment Analysis
Directory of Open Access Journals (Sweden)
Md. Kamrul Hossain
2013-01-01
Full Text Available Data envelopment analysis (DEA measures relative efficiency among the decision making units (DMU without considering noise in data. The least efficient DMU indicates that it is in the worst situation. In this paper, we measure efficiency of individual DMU whenever it losses the maximum output, and the efficiency of other DMUs is measured in the observed situation. This efficiency is the minimum efficiency of a DMU. The concept of stochastic data envelopment analysis (SDEA is a DEA method which considers the noise in data which is proposed in this study. Using bounded Pareto distribution, we estimate the DEA efficiency from efficiency interval. Small value of shape parameter can estimate the efficiency more accurately using the Pareto distribution. Rank correlations were estimated between observed efficiencies and minimum efficiency as well as between observed and estimated efficiency. The correlations are indicating the effectiveness of this SDEA model.
Robust estimation and hypothesis testing
Tiku, Moti L
2004-01-01
In statistical theory and practice, a certain distribution is usually assumed and then optimal solutions sought. Since deviations from an assumed distribution are very common, one cannot feel comfortable with assuming a particular distribution and believing it to be exactly correct. That brings the robustness issue in focus. In this book, we have given statistical procedures which are robust to plausible deviations from an assumed mode. The method of modified maximum likelihood estimation is used in formulating these procedures. The modified maximum likelihood estimators are explicit functions of sample observations and are easy to compute. They are asymptotically fully efficient and are as efficient as the maximum likelihood estimators for small sample sizes. The maximum likelihood estimators have computational problems and are, therefore, elusive. A broad range of topics are covered in this book. Solutions are given which are easy to implement and are efficient. The solutions are also robust to data anomali...
Energy Technology Data Exchange (ETDEWEB)
Imaki, Tadashi; Ishizuka, Hitoshi; Hayakawa, Junji
1987-12-21
According to the results of measuring crop photosynthesis and respiration of rice population, a comparative experiment on the variations of the ratio of respiration to photosynthesis (R/P ratio) due to the difference of the rice planting time or the heading time was carried out using extremely early-maturing and early or medium-maturing varieties. On the measured data of photosynthesis and respiration which were observed with change of the growth, the R/P ratio were obtained twenty days before to thirty days after the heading time. As the results, it was found that the R/P ratio of the group with the heading time at the end of July was about 10 to 20 % higher than that of another group with the heading time after middle of August. This means that the apparent energy conversion efficiency may be affected by the differences of the planting time, the heading time and the temperature condition. Hitherto, in determination of the rice-planting time, quantity of solar radiation in the growth and ripening stages was apt to be regarded as important. However, in consideration of true suitable planting time, the study from the aspect of the energy conversion efficiency is also required. (7 figs, 9 tabs, 8 refs)
International Nuclear Information System (INIS)
Aamir, K.M.; Maud, M.A.
2004-01-01
Small perturbations in signals (or any time series), at some particular instant, affect the whole frequency spectrum due to the global function e/sup j omega t/ in Fourier Transform formulation. However, the Fourier spectrum does not convey the time instant at which the perturbation occurred. Consequently the information on the particular time instance of occurrence of that perturbation is lost when spectrum is observed. Apparently Fourier analysis seems to be inadequate in such situations. This inadequacy is overcome by the use of Short Time Fourier Transform (STFT), which keeps track of time as well as frequency information. In STFT analysis, a fixed length window, say of length N, is moved sample by sample as the data arrives. The Discrete Fourier Transform (DFT) of this fixed window of length N is calculated using Fast Fourier Transform (FFT) algorithm. If the total number of points is M > N, the computational complexity of this scheme works out to be at least ((M-N) N log/sub 2/N). On the other hand, STFT is shown to be of computational complexity 6NM and 8NM in the literature. In this paper, two algorithms are presented which compute the same STFT more efficiently. The computational complexity of the proposed algorithms works out to be MN of one algorithm and even lesser in the other algorithm. This reduction in complexity becomes significant for large data sets. This algorithm also remains valid if a stationary part of signal is skipped. (author)
International Nuclear Information System (INIS)
Kalsi, P.C.
2010-01-01
The fission track registration efficiency of diethylene glycol bis allyl carbonate (dioctyl phthalate doped) (CR-39 (DOP)) solid state nuclear track detector (SSNTD) in solution medium (K wet ) has been experimentally determined and is found to be (9.7 ± 0.5).10 -4 cm. This is in good agreement with the values of other SSNTDs. The gamma irradiation effects in the dose range of 50.0-220.0 kGy on the bulk etch rate, V b and the activation energy for bulk etching, E of this solid state nuclear track detector (SSNTD) have also been studied. It is observed that the bulk etch rates increase and the activation energies for bulk etching decrease with the increase in gamma dose. These results have been explained on the basis of scission of the detector due to gamma irradiation
Motor-operated gearbox efficiency
Energy Technology Data Exchange (ETDEWEB)
DeWall, K.G.; Watkins, J.C.; Bramwell, D. [Idaho National Engineering Lab., Idaho Falls, ID (United States); Weidenhamer, G.H.
1996-12-01
Researchers at the Idaho National Engineering Laboratory recently conducted tests investigating the operating efficiency of the power train (gearbox) in motor-operators typically used in nuclear power plants to power motor-operated valves. Actual efficiency ratios were determined from in-line measurements of electric motor torque (input to the operator gearbox) and valve stem torque (output from the gearbox) while the operators were subjected to gradually increasing loads until the electric motor stalled. The testing included parametric studies under reduced voltage and elevated temperature conditions. As part of the analysis of the results, the authors compared efficiency values determined from testing to the values published by the operator manufacturer and typically used by the industry in calculations for estimating motor-operator capabilities. The operators they tested under load ran at efficiencies lower than the running efficiency (typically 50%) published by the operator manufacturer.
Motor-operated gearbox efficiency
International Nuclear Information System (INIS)
DeWall, K.G.; Watkins, J.C.; Bramwell, D.; Weidenhamer, G.H.
1996-01-01
Researchers at the Idaho National Engineering Laboratory recently conducted tests investigating the operating efficiency of the power train (gearbox) in motor-operators typically used in nuclear power plants to power motor-operated valves. Actual efficiency ratios were determined from in-line measurements of electric motor torque (input to the operator gearbox) and valve stem torque (output from the gearbox) while the operators were subjected to gradually increasing loads until the electric motor stalled. The testing included parametric studies under reduced voltage and elevated temperature conditions. As part of the analysis of the results, the authors compared efficiency values determined from testing to the values published by the operator manufacturer and typically used by the industry in calculations for estimating motor-operator capabilities. The operators they tested under load ran at efficiencies lower than the running efficiency (typically 50%) published by the operator manufacturer
Motor-operator gearbox efficiency
International Nuclear Information System (INIS)
DeWall, K.G.; Watkins, J.C.; Bramwell, D.
1996-01-01
Researchers at the Idaho National Engineering Laboratory recently conducted tests investigating the operating efficiency of the power train (gearbox) in motor-operators typically used in nuclear power plants to power motor-operated valves. Actual efficiency ratios were determined from in-line measurements of electric motor torque (input to the operator gearbox) and valve stem torque (output from the gearbox) while the operators were subjected to gradually increasing loads until the electric motor stalled. The testing included parametric studies under reduced voltage and elevated temperature conditions. As part of the analysis of the results, we compared efficiency values determined from testing to the values published by the operator manufacturer and typically used by the industry in calculations for estimating motor-operator capabilities. The operators we tested under load ran at efficiencies lower than the running efficiency (typically 50%) published by the operator manufacturer
Testing for Stochastic Dominance Efficiency
G.T. Post (Thierry); O. Linton; Y-J. Whang
2005-01-01
textabstractWe propose a new test of the stochastic dominance efficiency of a given portfolio over a class of portfolios. We establish its null and alternative asymptotic properties, and define a method for consistently estimating critical values. We present some numerical evidence that our
DEFF Research Database (Denmark)
Arndt, Channing; Simler, Kenneth R.
2010-01-01
A fundamental premise of absolute poverty lines is that they represent the same level of utility through time and space. Disturbingly, a series of recent studies in middle- and low-income economies show that even carefully derived poverty lines rarely satisfy this premise. This article proposes a......, with the current approach tending to systematically overestimate (underestimate) poverty in urban (rural) zones.......A fundamental premise of absolute poverty lines is that they represent the same level of utility through time and space. Disturbingly, a series of recent studies in middle- and low-income economies show that even carefully derived poverty lines rarely satisfy this premise. This article proposes...... an information-theoretic approach to estimating cost-of-basic-needs (CBN) poverty lines that are utility consistent. Applications to date illustrate that utility-consistent poverty measurements derived from the proposed approach and those derived from current CBN best practices often differ substantially...
Risk estimation using probability machines
2014-01-01
Background Logistic regression has been the de facto, and often the only, model used in the description and analysis of relationships between a binary outcome and observed features. It is widely used to obtain the conditional probabilities of the outcome given predictors, as well as predictor effect size estimates using conditional odds ratios. Results We show how statistical learning machines for binary outcomes, provably consistent for the nonparametric regression problem, can be used to provide both consistent conditional probability estimation and conditional effect size estimates. Effect size estimates from learning machines leverage our understanding of counterfactual arguments central to the interpretation of such estimates. We show that, if the data generating model is logistic, we can recover accurate probability predictions and effect size estimates with nearly the same efficiency as a correct logistic model, both for main effects and interactions. We also propose a method using learning machines to scan for possible interaction effects quickly and efficiently. Simulations using random forest probability machines are presented. Conclusions The models we propose make no assumptions about the data structure, and capture the patterns in the data by just specifying the predictors involved and not any particular model structure. So they do not run the same risks of model mis-specification and the resultant estimation biases as a logistic model. This methodology, which we call a “risk machine”, will share properties from the statistical machine that it is derived from. PMID:24581306
Generalized Centroid Estimators in Bioinformatics
Hamada, Michiaki; Kiryu, Hisanori; Iwasaki, Wataru; Asai, Kiyoshi
2011-01-01
In a number of estimation problems in bioinformatics, accuracy measures of the target problem are usually given, and it is important to design estimators that are suitable to those accuracy measures. However, there is often a discrepancy between an employed estimator and a given accuracy measure of the problem. In this study, we introduce a general class of efficient estimators for estimation problems on high-dimensional binary spaces, which represent many fundamental problems in bioinformatics. Theoretical analysis reveals that the proposed estimators generally fit with commonly-used accuracy measures (e.g. sensitivity, PPV, MCC and F-score) as well as it can be computed efficiently in many cases, and cover a wide range of problems in bioinformatics from the viewpoint of the principle of maximum expected accuracy (MEA). It is also shown that some important algorithms in bioinformatics can be interpreted in a unified manner. Not only the concept presented in this paper gives a useful framework to design MEA-based estimators but also it is highly extendable and sheds new light on many problems in bioinformatics. PMID:21365017
Isobars and the Efficient Market Hypothesis
Kristýna Ivanková
2010-01-01
Isobar surfaces, a method for describing the overall shape of multidimensional data, are estimated by nonparametric regression and used to evaluate the efficiency of selected markets based on returns of their stock market indices.
Estimation of population mean under systematic sampling
Noor-ul-amin, Muhammad; Javaid, Amjad
2017-11-01
In this study we propose a generalized ratio estimator under non-response for systematic random sampling. We also generate a class of estimators through special cases of generalized estimator using different combinations of coefficients of correlation, kurtosis and variation. The mean square errors and mathematical conditions are also derived to prove the efficiency of proposed estimators. Numerical illustration is included using three populations to support the results.
International Nuclear Information System (INIS)
Nipkow, J.; Brunner, C. U.
2005-01-01
This comprehensive article discusses the perspectives for reducing electricity consumption in Switzerland. The increase in consumption is discussed that has occurred in spite of the efforts of the Swiss national energy programmes 'Energy 2000' and 'SwissEnergy'. The fact that energy consumption is still on the increase although efficient and economically-viable technology is available is commented on. The authors are of the opinion that the market alone cannot provide a complete solution and that national and international efforts are needed to remedy things. In particular, the external costs that are often not included when estimating costs are stressed. Several technical options available, such as the use of fluorescent lighting, LCD monitors and efficient electric motors, are looked at as are other technologies quoted as being a means of reducing power consumption. Ways of reducing stand-by losses and system optimisation are looked at as are various scenarios for further development and measures that can be implemented in order to reduce power consumption
Radiation risk estimation based on measurement error models
Masiuk, Sergii; Shklyar, Sergiy; Chepurny, Mykola; Likhtarov, Illya
2017-01-01
This monograph discusses statistics and risk estimates applied to radiation damage under the presence of measurement errors. The first part covers nonlinear measurement error models, with a particular emphasis on efficiency of regression parameter estimators. In the second part, risk estimation in models with measurement errors is considered. Efficiency of the methods presented is verified using data from radio-epidemiological studies.
Effects of heterogeneity on bank efficiency scores
Bos, J. W. B.; Koetter, M.; Kolari, J. W.; Kool, C. J. M.
2009-01-01
Bank efficiency estimates often serve as a proxy of managerial skill since they quantify sub-optimal production choices. But such deviations can also be due to omitted systematic differences among banks. In this study, we examine the effects of heterogeneity on bank efficiency scores. We compare
Efficiency gains, bounds, and risk in finance
Sarisoy, Cisil
2015-01-01
This thesis consists of three chapters. The first chapter analyzes efficiency gains in the estimation of expected returns based on asset pricing models and examines the economic implications of such gains in portfolio allocation exercises. The second chapter provides nonparametric efficiency bounds
EPA’s Travel Efficiency Method (TEAM) AMPO Presentation
Presentation describes EPA’s Travel Efficiency Assessment Method (TEAM) assessing potential travel efficiency strategies for reducing travel activity and emissions, includes reduction estimates in Vehicle Miles Traveled in four different geographic areas.
2010-01-01
... consumption, estimated annual operating cost, and energy efficiency rating, and of water use rate. 305.5... RULE CONCERNING DISCLOSURES REGARDING ENERGY CONSUMPTION AND WATER USE OF CERTAIN HOME APPLIANCES AND... § 305.5 Determinations of estimated annual energy consumption, estimated annual operating cost, and...
Optimal Smoothing in Adaptive Location Estimation
Mammen, Enno; Park, Byeong U.
1997-01-01
In this paper higher order performance of kernel basedadaptive location estimators are considered. Optimalchoice of smoothing parameters is discussed and it isshown how much is lossed in efficiency by not knowingthe underlying translation density.
Directory of Open Access Journals (Sweden)
Douglas Sampaio Henrique
2005-06-01
Full Text Available Data of 320 animals were obtained from eight comparative slaughter studies performed under tropical conditions and used to estimate the total efficiency of utilization of the metabolizable energy intake (MEI, which varied from 77 to 419 kcal kg-0.75d-1. The provided data also contained direct measures of the recovered energy (RE, which allowed calculating the heat production (HE by difference. The RE was regressed on MEI and deviations from linearity were evaluated by using the F-test. The respective estimates of the fasting heat production and the intercept and the slope that composes the relationship between RE and MEI were 73 kcal kg-0.75d-1, 42 kcal kg-0.75d-1 and 0.37. Hence, the total efficiency was estimated by dividing the net energy for maintenance and growth by the metabolizable energy intake. The estimated total efficiency of the ME utilization and analogous estimates based on the beef cattle NRC model were employed in an additional study to evaluate their predictive powers in terms of the mean square deviations for both temperate and tropical conditions. The two approaches presented similar predictive powers but the proposed one had a 22% lower mean squared deviation even with its more simplified structure.Foram utilizadas 320 informações obtidas a partir de 8 estudos de abate comparativo conduzidos em condições tropicais para se estimar a eficiência total de utilização da energia metabolizável consumida, a qual variou de 77 a 419kcal kg-0.75d-1. Os dados também continham informações sobre a energia retida (RE, o que permitiu o cálculo da produção de calor por diferença. As estimativas da produção de calor em jejum e dos coeficientes linear e angular da regressão entre RE e MEI foram respectivamente, 73 kcal kg-0.75d-1, 42 kcal kg-0.75d-1 e 0,37. Em seguida, a eficiência total foi estimada dividindo-se a energia líquida para mantença e produção pelo consumo de energia metabolizável. A eficiência total de
Fast Katz and commuters : efficient estimation of social relatedness.
Energy Technology Data Exchange (ETDEWEB)
On, Byung-Won; Lakshmanan, Laks V. S.; Esfandiar, Pooya; Bonchi, Francesco; Grief, Chen; Gleich, David F.
2010-12-01
Motivated by social network data mining problems such as link prediction and collaborative filtering, significant research effort has been devoted to computing topological measures including the Katz score and the commute time. Existing approaches typically approximate all pairwise relationships simultaneously. In this paper, we are interested in computing: the score for a single pair of nodes, and the top-k nodes with the best scores from a given source node. For the pairwise problem, we apply an iterative algorithm that computes upper and lower bounds for the measures we seek. This algorithm exploits a relationship between the Lanczos process and a quadrature rule. For the top-k problem, we propose an algorithm that only accesses a small portion of the graph and is related to techniques used in personalized PageRank computing. To test the scalability and accuracy of our algorithms we experiment with three real-world networks and find that these algorithms run in milliseconds to seconds without any preprocessing.
Efficient estimation of burst-mode LDA power spectra
DEFF Research Database (Denmark)
Velte, Clara Marika; George, William K
2010-01-01
the velocity bias effects. Residence time weighting should also be used to compute velocity spectra. The residence time-weighted direct Fourier transform can, however, be computationally heavy, especially for the large data sets needed to eliminate finite time window effects and given the increased...
Comparison of Vehicle Efficiency Technology Attributes and Synergy Estimates
Energy Technology Data Exchange (ETDEWEB)
Duleep, G. [ICF Incorporated, LLC., Fairfax, VA (United States)
2011-02-01
Analyzing the future fuel economy of light-duty vehicles (LDVs) requires detailed knowledge of the vehicle technologies available to improve LDV fuel economy. The National Highway Transportation Safety Administration (NHTSA) has been relying on technology data from a 2001 National Academy of Sciences (NAS) study (NAS 2001) on corporate average fuel economy (CAFE) standards, but the technology parameters were updated in the new proposed rulemaking (EPA and NHTSA 2009) to set CAFE and greenhouse gas standards for the 2011 to 2016 period. The update is based largely on an Environmental Protection Agency (EPA) analysis of technology attributes augmented by NHTSA data and contractor staff assessments. These technology cost and performance data were documented in the Draft Joint Technical Support Document (TSD) issued by EPA and NHTSA in September 2009 (EPA/NHTSA 2009). For these tasks, the Energy and Environmental Analysis (EEA) division of ICF International (ICF) examined each technology and technology package in the Draft TSD and assessed their costs and performance potential based on U.S. Department of Energy (DOE) program assessments. ICF also assessed the technologies, other relevant attributes based on data from actual production vehicles, and recently published technical articles in engineering journals. ICF examined technology synergy issues through an ICF in-house model that uses a discrete parameter approach.
Comparison of Vehicle Efficiency Technology Attributes and Synergy Estimates
Energy Technology Data Exchange (ETDEWEB)
Duleep, G.
2011-02-01
Analyzing the future fuel economy of light-duty vehicles (LDVs) requires detailed knowledge of the vehicle technologies available to improve LDV fuel economy. The National Highway Transportation Safety Administration (NHTSA) has been relying on technology data from a 2001 National Academy of Sciences (NAS) study (NAS 2001) on corporate average fuel economy (CAFE) standards, but the technology parameters were updated in the new proposed rulemaking (EPA and NHTSA 2009) to set CAFE and greenhouse gas standards for the 2011 to 2016 period. The update is based largely on an Environmental Protection Agency (EPA) analysis of technology attributes augmented by NHTSA data and contractor staff assessments. These technology cost and performance data were documented in the Draft Joint Technical Support Document (TSD) issued by EPA and NHTSA in September 2009 (EPA/NHTSA 2009). For these tasks, the Energy and Environmental Analysis (EEA) division of ICF International (ICF) examined each technology and technology package in the Draft TSD and assessed their costs and performance potential based on U.S. Department of Energy (DOE) program assessments. ICF also assessed the technologies? other relevant attributes based on data from actual production vehicles and from recently published technical articles in engineering journals. ICF examined technology synergy issues through an ICF in-house model that uses a discrete parameter approach.
Transverse correlation: An efficient transverse flow estimator - initial results
DEFF Research Database (Denmark)
Holfort, Iben Kraglund; Henze, Lasse; Kortbek, Jacob
2008-01-01
, a standard deviation of 1.64% and a bias of 1.13% are obtained for a beam to flow angle of 90 degrees. Using the same setup a standard deviation of 2.21% and a bias of 1.07% are obtained for a beam to flow angle of 75 degrees. Using 20 emissions a standard deviation of 3.4% and a bias of 2.06% are obtained...... at 45 degrees. The method performs stable down to a signal-to-noise ratio of 0 dB, where a standard deviation of 5.5% and a bias of 1.2% is achieved....
ESTIMATION OF EFFICIENCY OF THE COMPETITIVE COOPERATION MODEL
Directory of Open Access Journals (Sweden)
Natalia N. Liparteliani
2014-01-01
Full Text Available Competitive cooperation model of regional travel agencies and travel market participants is considered. Evaluation of the model using mathematical and statistical methods was carried out. Relationship marketing provides a travel company certain economic advantages.
Estimating decontamination efficiency of territory, adjoining to the Chernobyl' NPP
International Nuclear Information System (INIS)
Konstantinov, E.A.; Sanitarov, V.A.; Stepanov, I.K.; Senin, E.V.; Filippov, E.M.; Shiraj, A.P.; Umanets, M.P.
1989-01-01
The NPP territory decontamination was fulfilled according to the next scheme: 1) removal of the soil upper layer and its moving avay for burial; 2) concrete plate laying; 3) concreting seams between plates and piling up polymer coatings. The conclusion is made that substantial improvement of radiation situation at the operating site is possible only after removal of radiation sources at the machine room roofing. 1 ref.; 5 figs.; 2 tabs
Stochastic Frontier Estimation of Efficient Learning in Video Games
Hamlen, Karla R.
2012-01-01
Stochastic Frontier Regression Analysis was used to investigate strategies and skills that are associated with the minimization of time required to achieve proficiency in video games among students in grades four and five. Students self-reported their video game play habits, including strategies and skills used to become good at the video games…
Social Responsibility and Economic Efficiency: aspects of estimation
Filippova, Irina
2014-01-01
The article discusses the concept of social responsibility of state and business in terms of effectiveness of social production system. In this perspective the traditional approaches to determining the effectiveness of the economic system were critically reviewed. Not only new approach to assessing the effectiveness being proposed in this paper, but also the link between effectiveness of public production system and social responsibility of core subjects is substantiated. Conventional approac...
Estimation of the reaction efficiency in polymerase chain reaction
Lalam, N.
2006-01-01
Polymerase chain reaction (PCR) is largely used in molecular biology for increasing the copy number of a specific DNA fragment. The succession of 20 replication cycles makes it possible to multiply the quantity of the fragment of interest by a factor of 1 million. The PCR technique has
Efficient lifetime estimation techniques for general multiaxial loading
Papuga, Jan; Halama, Radim; Fusek, Martin; Rojíček, Jaroslav; Fojtík, František; Horák, David; Pecha, Marek; Tomčala, Jiří; Čermák, Martin; Hapla, Václav; Sojka, Radim; Kružík, Jakub
2017-07-01
In this paper, we discuss and present our progress toward a project, which is focused on fatigue life prediction under multiaxial loading in the domain of low-cycle fatigue, i.e. cases, where the plasticity cannot be neglected. First, the elastic-plastic solution in the finite element analysis is enhanced and verified on own experiments. Second, the method by Jiang describing the instantaneous damage increase by analyses of load time by time, is in implementation phase. In addition, simplified routines for conversion of elastic stresses-strains to elastic-plastic ones as proposed by Firat and Ye et.al. are evaluated on the basis of data gathered from external sources. In order to produce high quality complex analyses, which could be feasible in an acceptable time, and allow the period for next analyses of results to be expanded; the core of PragTic fatigue solver used for all fatigue computations are being re-implemented to get the fully parallelized scalable solution.
Energy Technology Data Exchange (ETDEWEB)
Tschudi, William; Xu, Tengfang; Sartor, Dale; Koomey, Jon; Nordman, Bruce; Sezgen, Osman
2004-03-30
Data Center facilities, prevalent in many industries and institutions are essential to California's economy. Energy intensive data centers are crucial to California's industries, and many other institutions (such as universities) in the state, and they play an important role in the constantly evolving communications industry. To better understand the impact of the energy requirements and energy efficiency improvement potential in these facilities, the California Energy Commission's PIER Industrial Program initiated this project with two primary focus areas: First, to characterize current data center electricity use; and secondly, to develop a research ''roadmap'' defining and prioritizing possible future public interest research and deployment efforts that would improve energy efficiency. Although there are many opinions concerning the energy intensity of data centers and the aggregate effect on California's electrical power systems, there is very little publicly available information. Through this project, actual energy consumption at its end use was measured in a number of data centers. This benchmark data was documented in case study reports, along with site-specific energy efficiency recommendations. Additionally, other data center energy benchmarks were obtained through synergistic projects, prior PG&E studies, and industry contacts. In total, energy benchmarks for sixteen data centers were obtained. For this project, a broad definition of ''data center'' was adopted which included internet hosting, corporate, institutional, governmental, educational and other miscellaneous data centers. Typically these facilities require specialized infrastructure to provide high quality power and cooling for IT equipment. All of these data center types were considered in the development of an estimate of the total power consumption in California. Finally, a research ''roadmap'' was developed
Variance estimation for generalized Cavalieri estimators
Johanna Ziegel; Eva B. Vedel Jensen; Karl-Anton Dorph-Petersen
2011-01-01
The precision of stereological estimators based on systematic sampling is of great practical importance. This paper presents methods of data-based variance estimation for generalized Cavalieri estimators where errors in sampling positions may occur. Variance estimators are derived under perturbed systematic sampling, systematic sampling with cumulative errors and systematic sampling with random dropouts. Copyright 2011, Oxford University Press.
Energy-efficient cooking methods
Energy Technology Data Exchange (ETDEWEB)
De, Dilip K. [Department of Physics, University of Jos, P.M.B. 2084, Jos, Plateau State (Nigeria); Muwa Shawhatsu, N. [Department of Physics, Federal University of Technology, Yola, P.M.B. 2076, Yola, Adamawa State (Nigeria); De, N.N. [Department of Mechanical and Aerospace Engineering, The University of Texas at Arlington, Arlington, TX 76019 (United States); Ikechukwu Ajaeroh, M. [Department of Physics, University of Abuja, Abuja (Nigeria)
2013-02-15
Energy-efficient new cooking techniques have been developed in this research. Using a stove with 649{+-}20 W of power, the minimum heat, specific heat of transformation, and on-stove time required to completely cook 1 kg of dry beans (with water and other ingredients) and 1 kg of raw potato are found to be: 710 {+-}kJ, 613 {+-}kJ, and 1,144{+-}10 s, respectively, for beans and 287{+-}12 kJ, 200{+-}9 kJ, and 466{+-}10 s for Irish potato. Extensive researches show that these figures are, to date, the lowest amount of heat ever used to cook beans and potato and less than half the energy used in conventional cooking with a pressure cooker. The efficiency of the stove was estimated to be 52.5{+-}2 %. Discussion is made to further improve the efficiency in cooking with normal stove and solar cooker and to save food nutrients further. Our method of cooking when applied globally is expected to contribute to the clean development management (CDM) potential. The approximate values of the minimum and maximum CDM potentials are estimated to be 7.5 x 10{sup 11} and 2.2 x 10{sup 13} kg of carbon credit annually. The precise estimation CDM potential of our cooking method will be reported later.
The Efficiency of Educational Production
DEFF Research Database (Denmark)
Bogetoft, Peter; Heinesen, Eskil; Tranæs, Torben
is the most efficient Nordic country (often fully efficient), whereas Sweden and especially Norway and Denmark are clearly inefficient. However, using PISA test scores as indicators of student input quality in upper secondary education reduces the inefficiencies of these three countries. Also, when expected......Focusing in particular on upper secondary education, this paper examines whether the relatively high level of expenditure on education in the Nordic countries is matched by high output from the educational sector, both in terms of student enrolment and indicators of output quality in the form...... of graduation/completion rates and expected earnings after completed education. We use Data Envelopment Analysis (DEA) to compare (benchmark) the Nordic countries with a relevant group of rich OECD countries and calculate input efficiency scores for each country. We estimate a wide range of specifications...
The Efficiency of Educational Production
DEFF Research Database (Denmark)
Bogetoft, Peter; Heinesen, Eskil; Tranæs, Torben
2015-01-01
is the most efficient Nordic country (often fully efficient), whereas Sweden and especially Norway and Denmark are clearly inefficient. However, using PISA test scores as indicators of student input quality in upper secondary education reduces the inefficiencies of these three countries. Also, when expected......Focusing in particular on upper secondary education, this paper examines whether the relatively high level of expenditure on education in the Nordic countries is matched by high output from the educational sector, both in terms of student enrolment and indicators of output quality in the form...... of graduation/completion rates and expected earnings after completed education. We use data envelopment analysis (DEA) to compare (benchmark) the Nordic countries with a relevant group of rich OECD countries and calculate input efficiency scores for each country. We estimate a wide range of specifications...
Optimizing lengths of confidence intervals: fourth-order efficiency in location models
Klaassen, C.; Venetiaan, S.
2010-01-01
Under regularity conditions the maximum likelihood estimator of the location parameter in a location model is asymptotically efficient among translation equivariant estimators. Additional regularity conditions warrant third- and even fourth-order efficiency, in the sense that no translation
Estimation of biochemical variables using quantumbehaved particle ...
African Journals Online (AJOL)
To generate a more efficient neural network estimator, we employed the previously proposed quantum-behaved particle swarm optimization (QPSO) algorithm for neural network training. The experiment results of L-glutamic acid fermentation process showed that our established estimator could predict variables such as the ...
Efficiency of Finish power transmission network companies
International Nuclear Information System (INIS)
Anon.
2001-01-01
The Finnish Energy Market Authority has investigated the efficiency of power transmissions network companies. The results show that the intensification potential of the branch is 402 million FIM, corresponding to about 15% of the total costs of the branch and 7.3 % of the turnout. Energy Market Authority supervises the reasonableness of the power transmission prices, and it will use the results of the research in supervision. The research was carried out by the Quantitative Methods Research Group of Helsinki School of Economics. The main objective of the research was to create an efficiency estimation method for electric power distribution network business used for Finnish conditions. Data of the year 1998 was used as basic material in the research. Twenty-one of the 102 power distribution network operators was estimated to be totally efficient. Highest possible efficiency rate was 100, and the average of the efficiency rates of all the operators was 76.9, the minimum being 42.6
Relative Pose Estimation Algorithm with Gyroscope Sensor
Directory of Open Access Journals (Sweden)
Shanshan Wei
2016-01-01
Full Text Available This paper proposes a novel vision and inertial fusion algorithm S2fM (Simplified Structure from Motion for camera relative pose estimation. Different from current existing algorithms, our algorithm estimates rotation parameter and translation parameter separately. S2fM employs gyroscopes to estimate camera rotation parameter, which is later fused with the image data to estimate camera translation parameter. Our contributions are in two aspects. (1 Under the circumstance that no inertial sensor can estimate accurately enough translation parameter, we propose a translation estimation algorithm by fusing gyroscope sensor and image data. (2 Our S2fM algorithm is efficient and suitable for smart devices. Experimental results validate efficiency of the proposed S2fM algorithm.
Condition Number Regularized Covariance Estimation.
Won, Joong-Ho; Lim, Johan; Kim, Seung-Jean; Rajaratnam, Bala
2013-06-01
Estimation of high-dimensional covariance matrices is known to be a difficult problem, has many applications, and is of current interest to the larger statistics community. In many applications including so-called the "large p small n " setting, the estimate of the covariance matrix is required to be not only invertible, but also well-conditioned. Although many regularization schemes attempt to do this, none of them address the ill-conditioning problem directly. In this paper, we propose a maximum likelihood approach, with the direct goal of obtaining a well-conditioned estimator. No sparsity assumption on either the covariance matrix or its inverse are are imposed, thus making our procedure more widely applicable. We demonstrate that the proposed regularization scheme is computationally efficient, yields a type of Steinian shrinkage estimator, and has a natural Bayesian interpretation. We investigate the theoretical properties of the regularized covariance estimator comprehensively, including its regularization path, and proceed to develop an approach that adaptively determines the level of regularization that is required. Finally, we demonstrate the performance of the regularized estimator in decision-theoretic comparisons and in the financial portfolio optimization setting. The proposed approach has desirable properties, and can serve as a competitive procedure, especially when the sample size is small and when a well-conditioned estimator is required.
Condition Number Regularized Covariance Estimation*
Won, Joong-Ho; Lim, Johan; Kim, Seung-Jean; Rajaratnam, Bala
2012-01-01
Estimation of high-dimensional covariance matrices is known to be a difficult problem, has many applications, and is of current interest to the larger statistics community. In many applications including so-called the “large p small n” setting, the estimate of the covariance matrix is required to be not only invertible, but also well-conditioned. Although many regularization schemes attempt to do this, none of them address the ill-conditioning problem directly. In this paper, we propose a maximum likelihood approach, with the direct goal of obtaining a well-conditioned estimator. No sparsity assumption on either the covariance matrix or its inverse are are imposed, thus making our procedure more widely applicable. We demonstrate that the proposed regularization scheme is computationally efficient, yields a type of Steinian shrinkage estimator, and has a natural Bayesian interpretation. We investigate the theoretical properties of the regularized covariance estimator comprehensively, including its regularization path, and proceed to develop an approach that adaptively determines the level of regularization that is required. Finally, we demonstrate the performance of the regularized estimator in decision-theoretic comparisons and in the financial portfolio optimization setting. The proposed approach has desirable properties, and can serve as a competitive procedure, especially when the sample size is small and when a well-conditioned estimator is required. PMID:23730197
Boundary methods for mode estimation
Pierson, William E., Jr.; Ulug, Batuhan; Ahalt, Stanley C.
1999-08-01
This paper investigates the use of Boundary Methods (BMs), a collection of tools used for distribution analysis, as a method for estimating the number of modes associated with a given data set. Model order information of this type is required by several pattern recognition applications. The BM technique provides a novel approach to this parameter estimation problem and is comparable in terms of both accuracy and computations to other popular mode estimation techniques currently found in the literature and automatic target recognition applications. This paper explains the methodology used in the BM approach to mode estimation. Also, this paper quickly reviews other common mode estimation techniques and describes the empirical investigation used to explore the relationship of the BM technique to other mode estimation techniques. Specifically, the accuracy and computational efficiency of the BM technique are compared quantitatively to the a mixture of Gaussian (MOG) approach and a k-means approach to model order estimation. The stopping criteria of the MOG and k-means techniques is the Akaike Information Criteria (AIC).
Poisson sampling - The adjusted and unadjusted estimator revisited
Michael S. Williams; Hans T. Schreuder; Gerardo H. Terrazas
1998-01-01
The prevailing assumption, that for Poisson sampling the adjusted estimator "Y-hat a" is always substantially more efficient than the unadjusted estimator "Y-hat u" , is shown to be incorrect. Some well known theoretical results are applicable since "Y-hat a" is a ratio-of-means estimator and "Y-hat u" a simple unbiased estimator...
On Estimating Quantiles Using Auxiliary Information
Directory of Open Access Journals (Sweden)
Berger Yves G.
2015-03-01
Full Text Available We propose a transformation-based approach for estimating quantiles using auxiliary information. The proposed estimators can be easily implemented using a regression estimator. We show that the proposed estimators are consistent and asymptotically unbiased. The main advantage of the proposed estimators is their simplicity. Despite the fact the proposed estimators are not necessarily more efficient than their competitors, they offer a good compromise between accuracy and simplicity. They can be used under single and multistage sampling designs with unequal selection probabilities. A simulation study supports our finding and shows that the proposed estimators are robust and of an acceptable accuracy compared to alternative estimators, which can be more computationally intensive.
Online wave estimation using vessel motion measurements
DEFF Research Database (Denmark)
H. Brodtkorb, Astrid; Nielsen, Ulrik D.; J. Sørensen, Asgeir
2018-01-01
parameters and motion transfer functions are required as input. Apart from this the method is signal-based, with no assumptions on the wave spectrum shape, and as a result it is computationally efficient. The algorithm is implemented in a dynamic positioning (DP)control system, and tested through simulations......In this paper, a computationally efficient online sea state estimation algorithm isproposed for estimation of the on site sea state. The algorithm finds the wave spectrum estimate from motion measurements in heave, roll and pitch by iteratively solving a set of linear equations. The main vessel...
Dose-response curve estimation: a semiparametric mixture approach.
Yuan, Ying; Yin, Guosheng
2011-12-01
In the estimation of a dose-response curve, parametric models are straightforward and efficient but subject to model misspecifications; nonparametric methods are robust but less efficient. As a compromise, we propose a semiparametric approach that combines the advantages of parametric and nonparametric curve estimates. In a mixture form, our estimator takes a weighted average of the parametric and nonparametric curve estimates, in which a higher weight is assigned to the estimate with a better model fit. When the parametric model assumption holds, the semiparametric curve estimate converges to the parametric estimate and thus achieves high efficiency; when the parametric model is misspecified, the semiparametric estimate converges to the nonparametric estimate and remains consistent. We also consider an adaptive weighting scheme to allow the weight to vary according to the local fit of the models. We conduct extensive simulation studies to investigate the performance of the proposed methods and illustrate them with two real examples. © 2011, The International Biometric Society.
Efficiency in the Community College Sector: Stochastic Frontier Analysis
Agasisti, Tommaso; Belfield, Clive
2017-01-01
This paper estimates technical efficiency scores across the community college sector in the United States. Using stochastic frontier analysis and data from the Integrated Postsecondary Education Data System for 2003-2010, we estimate efficiency scores for 950 community colleges and perform a series of sensitivity tests to check for robustness. We…
Efficiency profile method to study the hit efficiency of drift chambers
International Nuclear Information System (INIS)
Abyzov, A.; Bel'kov, A.; Lanev, A.; Spiridonov, A.; Walter, M.; Hulsbergen, W.
2002-01-01
A method based on the usage of efficiency profile is proposed to estimate the hit efficiency of drift chambers with a large number of channels. The performance of the method under real conditions of the detector operation has been tested analysing the experimental data from the HERA-B drift chambers
Moving Horizon Estimation and Control
DEFF Research Database (Denmark)
Jørgensen, John Bagterp
successful and applied methodology beyond PID-control for control of industrial processes. The main contribution of this thesis is introduction and definition of the extended linear quadratic optimal control problem for solution of numerical problems arising in moving horizon estimation and control...... problems. Chapter 1 motivates moving horizon estimation and control as a paradigm for control of industrial processes. It introduces the extended linear quadratic control problem and discusses its central role in moving horizon estimation and control. Introduction, application and efficient solution....... It provides an algorithm for computation of the maximal output admissible set for linear model predictive control. Appendix D provides results concerning linear regression. Appendix E discuss prediction error methods for identification of linear models tailored for model predictive control....
Sparse DOA estimation with polynomial rooting
DEFF Research Database (Denmark)
Xenaki, Angeliki; Gerstoft, Peter; Fernandez Grande, Efren
2015-01-01
Direction-of-arrival (DOA) estimation involves the localization of a few sources from a limited number of observations on an array of sensors. Thus, DOA estimation can be formulated as a sparse signal reconstruction problem and solved efficiently with compressive sensing (CS) to achieve highresol......Direction-of-arrival (DOA) estimation involves the localization of a few sources from a limited number of observations on an array of sensors. Thus, DOA estimation can be formulated as a sparse signal reconstruction problem and solved efficiently with compressive sensing (CS) to achieve...... highresolution imaging. Utilizing the dual optimal variables of the CS optimization problem, it is shown with Monte Carlo simulations that the DOAs are accurately reconstructed through polynomial rooting (Root-CS). Polynomial rooting is known to improve the resolution in several other DOA estimation methods...
International Nuclear Information System (INIS)
Choi, Yongrok; Zhang, Ning; Zhou, P.
2012-01-01
Highlights: ► We employ a slacks-based DEA model to estimate the energy efficiency and shadow prices of CO 2 emissions in China. ► The empirical study shows that China was not performing CO 2 -efficiently. ► The average of estimated shadow prices of CO 2 emissions is about $7.2. -- Abstract: This paper uses nonparametric efficiency analysis technique to estimate the energy efficiency, potential emission reductions and marginal abatement costs of energy-related CO 2 emissions in China. We employ a non-radial slacks-based data envelopment analysis (DEA) model for estimating the potential reductions and efficiency of CO 2 emissions for China. The dual model of the slacks-based DEA model is then used to estimate the marginal abatement costs of CO 2 emissions. An empirical study based on China’s panel data (2001–2010) is carried out and some policy implications are also discussed.
Regional and global exergy and energy efficiencies
Energy Technology Data Exchange (ETDEWEB)
Nakicenovic, N; Kurz, R [International Inst. for Applied Systems Analysis, Laxenburg (Austria). Environmentally Compatible Energy Strategies (Ecuador) Project; Gilli, P V [Graz Univ. of Technology (Austria)
1996-03-01
We present estimates of global energy efficiency by applying second-law (exergy) analysis to regional and global energy balances. We use a uniform analysis of national and regional energy balances and aggregate these balances first for three main economic regions and subsequently into world totals. The procedure involves assessment of energy and exergy efficiencies at each step of energy conversion, from primary exergy to final and useful exergy. Ideally, the analysis should be extended to include actual delivered energy services; unfortunately, data are scarce and only rough estimates can be given for this last stage of energy conversion. The overall result is that the current global primary to useful exergy efficiency is about one-tenth of the theoretical maximum and the service efficiency is even lower. (Author)
90 measurement of technical efficiency and its determinants in crop ...
African Journals Online (AJOL)
OLUWOLE AKINNAGBE
They found that education, number of working animals, credit per acre and number of extension visits significantly increased cost efficiency while large land holding size significantly decreased cost efficiency. In single estimation approach of the technical efficiency model for Indian farmers,. Colli et al (1998), found that years ...
On the Optimality of Multivariate S-Estimators
Croux, C.; Dehon, C.; Yadine, A.
2010-01-01
In this paper we maximize the efficiency of a multivariate S-estimator under a constraint on the breakdown point. In the linear regression model, it is known that the highest possible efficiency of a maximum breakdown S-estimator is bounded above by 33% for Gaussian errors. We prove the surprising
Reconsidering energy efficiency
International Nuclear Information System (INIS)
Goldoni, Giovanni
2007-01-01
Energy and environmental policies are reconsidering energy efficiency. In a perfect market, rational and well informed consumers reach economic efficiency which, at the given prices of energy and capital, corresponds to physical efficiency. In the real world, market failures and cognitive frictions distort the consumers from perfectly rational and informed choices. Green incentive schemes aim at balancing market failures and directing consumers toward more efficient goods and services. The problem is to fine tune the incentive schemes [it
Energy efficiency; Efficacite energetique
Energy Technology Data Exchange (ETDEWEB)
NONE
2006-06-15
This road-map proposes by the Group Total aims to inform the public on the energy efficiency. It presents the energy efficiency and intensity around the world with a particular focus on Europe, the energy efficiency in industry and Total commitment. (A.L.B.)
Seismic efficiency of meteor airbursts
Svetsov, V. V.; Artemieva, N. A.; Shuvalov, V. V.
2017-08-01
We present the results of numerical simulation for impacts of relatively small asteroids and ice bodies of 30-100 m in size, decelerated in the atmosphere and exploding before they reach the surface, but still producing seismic effects due to the impact wave reaching the surface. The calculated magnitudes fall within the range of 4 to 6, and average seismic efficiency of these events is 2.5 × 10-5. The results obtained allow the seismic hazard from impacts of cosmic bodies to be estimated.
Corporate Accounting Policy Efficiency Improvement
Directory of Open Access Journals (Sweden)
Elena K. Vorobei
2013-01-01
Full Text Available The article is focused on the issues of efficient use of different methods of tax accounting for the optimization of income tax expenses and their consolidation in corporate accounting policy. The article makes reasoned conclusions, concerning optimal selection of depreciation methods for tax and bookkeeping accounting and their consolidation in corporate accounting policy and consolidation of optimal methods of cost recovery in production, considering business environment. The impact of the selected methods on corporate income tax rates and corporate property tax rates was traced and tax recovery was estimated.
Waldo, Staffan
2007-01-01
While individual data form the base for much empirical analysis in education, this is not the case for analysis of technical efficiency. In this paper, efficiency is estimated using individual data which is then aggregated to larger groups of students. Using an individual approach to technical efficiency makes it possible to carry out studies on a…
Coordination of Energy Efficiency and Demand Response
Energy Technology Data Exchange (ETDEWEB)
Goldman, Charles; Reid, Michael; Levy, Roger; Silverstein, Alison
2010-01-29
This paper reviews the relationship between energy efficiency and demand response and discusses approaches and barriers to coordinating energy efficiency and demand response. The paper is intended to support the 10 implementation goals of the National Action Plan for Energy Efficiency's Vision to achieve all cost-effective energy efficiency by 2025. Improving energy efficiency in our homes, businesses, schools, governments, and industries - which consume more than 70 percent of the nation's natural gas and electricity - is one of the most constructive, cost-effective ways to address the challenges of high energy prices, energy security and independence, air pollution, and global climate change. While energy efficiency is an increasingly prominent component of efforts to supply affordable, reliable, secure, and clean electric power, demand response is becoming a valuable tool in utility and regional resource plans. The Federal Energy Regulatory Commission (FERC) estimated the contribution from existing U.S. demand response resources at about 41,000 megawatts (MW), about 5.8 percent of 2008 summer peak demand (FERC, 2008). Moreover, FERC recently estimated nationwide achievable demand response potential at 138,000 MW (14 percent of peak demand) by 2019 (FERC, 2009).2 A recent Electric Power Research Institute study estimates that 'the combination of demand response and energy efficiency programs has the potential to reduce non-coincident summer peak demand by 157 GW' by 2030, or 14-20 percent below projected levels (EPRI, 2009a). This paper supports the Action Plan's effort to coordinate energy efficiency and demand response programs to maximize value to customers. For information on the full suite of policy and programmatic options for removing barriers to energy efficiency, see the Vision for 2025 and the various other Action Plan papers and guides available at www.epa.gov/eeactionplan.
Efficient flapping flight of pterosaurs
Strang, Karl Axel
In the late eighteenth century, humans discovered the first pterosaur fossil remains and have been fascinated by their existence ever since. Pterosaurs exploited their membrane wings in a sophisticated manner for flight control and propulsion, and were likely the most efficient and effective flyers ever to inhabit our planet. The flapping gait is a complex combination of motions that sustains and propels an animal in the air. Because pterosaurs were so large with wingspans up to eleven meters, if they could have sustained flapping flight, they would have had to achieve high propulsive efficiencies. Identifying the wing motions that contribute the most to propulsive efficiency is key to understanding pterosaur flight, and therefore to shedding light on flapping flight in general and the design of efficient ornithopters. This study is based on published results for a very well-preserved specimen of Coloborhynchus robustus, for which the joints are well-known and thoroughly described in the literature. Simplifying assumptions are made to estimate the characteristics that can not be inferred directly from the fossil remains. For a given animal, maximizing efficiency is equivalent to minimizing power at a given thrust and speed. We therefore aim at finding the flapping gait, that is the joint motions, that minimize the required flapping power. The power is computed from the aerodynamic forces created during a given wing motion. We develop an unsteady three-dimensional code based on the vortex-lattice method, which correlates well with published results for unsteady motions of rectangular wings. In the aerodynamic model, the rigid pterosaur wing is defined by the position of the bones. In the aeroelastic model, we add the flexibility of the bones and of the wing membrane. The nonlinear structural behavior of the membrane is reduced to a linear modal decomposition, assuming small deflections about the reference wing geometry. The reference wing geometry is computed for
Estimates of variance components for postweaning feed intake and ...
African Journals Online (AJOL)
Feed efficiency is of major economic importance in beef production. The objective of this work was to evaluate alternative measures of feed efficiency for use in genetic evaluation. To meet this objective, genetic parameters were estimated for the components of efficiency. These parameters were then used in multiple-trait ...
Distributed fusion estimation for sensor networks with communication constraints
Zhang, Wen-An; Song, Haiyu; Yu, Li
2016-01-01
This book systematically presents energy-efficient robust fusion estimation methods to achieve thorough and comprehensive results in the context of network-based fusion estimation. It summarizes recent findings on fusion estimation with communication constraints; several novel energy-efficient and robust design methods for dealing with energy constraints and network-induced uncertainties are presented, such as delays, packet losses, and asynchronous information... All the results are presented as algorithms, which are convenient for practical applications.
Sequential bayes estimation algorithm with cubic splines on uniform meshes
International Nuclear Information System (INIS)
Hossfeld, F.; Mika, K.; Plesser-Walk, E.
1975-11-01
After outlining the principles of some recent developments in parameter estimation, a sequential numerical algorithm for generalized curve-fitting applications is presented combining results from statistical estimation concepts and spline analysis. Due to its recursive nature, the algorithm can be used most efficiently in online experimentation. Using computer-sumulated and experimental data, the efficiency and the flexibility of this sequential estimation procedure is extensively demonstrated. (orig.) [de
Barriers to Industrial Energy Efficiency - Study (Appendix A), June 2015
Energy Technology Data Exchange (ETDEWEB)
None
2015-06-01
This study examines barriers that impede the adoption of energy efficient technologies and practices in the industrial sector, and identifies successful examples and opportunities to overcome these barriers. Three groups of energy efficiency technologies and measures were examined: industrial end-use energy efficiency, industrial demand response, and industrial combined heat and power. This study also includes the estimated economic benefits from hypothetical Federal energy efficiency matching grants, as directed by the Act.
Barriers to Industrial Energy Efficiency - Report to Congress, June 2015
Energy Technology Data Exchange (ETDEWEB)
None
2015-06-01
This report examines barriers that impede the adoption of energy efficient technologies and practices in the industrial sector, and identifies successful examples and opportunities to overcome these barriers. Three groups of energy efficiency technologies and measures were examined: industrial end-use energy efficiency, industrial demand response, and industrial combined heat and power. This report also includes the estimated economic benefits from hypothetical Federal energy efficiency matching grants, as directed by the Act.
Estimating the Doppler centroid of SAR data
DEFF Research Database (Denmark)
Madsen, Søren Nørvang
1989-01-01
attractive properties. An evaluation based on an existing SEASAT processor is reported. The time-domain algorithms are shown to be extremely efficient with respect to requirements on calculations and memory, and hence they are well suited to real-time systems where the Doppler estimation is based on raw SAR......After reviewing frequency-domain techniques for estimating the Doppler centroid of synthetic-aperture radar (SAR) data, the author describes a time-domain method and highlights its advantages. In particular, a nonlinear time-domain algorithm called the sign-Doppler estimator (SDE) is shown to have...... data. For offline processors where the Doppler estimation is performed on processed data, which removes the problem of partial coverage of bright targets, the ΔE estimator and the CDE (correlation Doppler estimator) algorithm give similar performance. However, for nonhomogeneous scenes it is found...
Market conditions affecting energy efficiency investments
International Nuclear Information System (INIS)
Seabright, J.
1996-01-01
The global energy efficiency market is growing, due in part to energy sector and macroeconomic reforms and increased awareness of the environmental benefits of energy efficiency. Many countries have promoted open, competitive markets, thereby stimulating economic growth. They have reduced or removed subsidies on energy prices, and governments have initiated energy conservation programs that have spurred the wider adoption of energy efficiency technologies. The market outlook for energy efficiency is quite positive. The global market for end-use energy efficiency in the industrial, residential and commercial sectors is now estimated to total more than $34 billion per year. There is still enormous technical potential to implement energy conservation measures and to upgrade to the best available technologies for new investments. For many technologies, energy-efficient designs now represent less than 10--20% of new product sales. Thus, creating favorable market conditions should be a priority. There are a number of actions that can be taken to create favorable market conditions for investing in energy efficiency. Fostering a market-oriented energy sector will lead to energy prices that reflect the true cost of supply. Policy initiatives should address known market failures and should support energy efficiency initiatives. And market transformation for energy efficiency products and services can be facilitated by creating an institutional and legal structure that favors commercially-oriented entities
Interactive inverse kinematics for human motion estimation
DEFF Research Database (Denmark)
Engell-Nørregård, Morten Pol; Hauberg, Søren; Lapuyade, Jerome
2009-01-01
We present an application of a fast interactive inverse kinematics method as a dimensionality reduction for monocular human motion estimation. The inverse kinematics solver deals efficiently and robustly with box constraints and does not suffer from shaking artifacts. The presented motion...... to significantly speed up the particle filtering. It should be stressed that the observation part of the system has not been our focus, and as such is described only from a sense of completeness. With our approach it is possible to construct a robust and computationally efficient system for human motion estimation....
CEE Energy Efficiency Report - Slovakia
International Nuclear Information System (INIS)
Hecl, V.
2005-01-01
need to be increased significantly if the proposed targets are to be realised. This increase in budget allocation would enable the implementation of programmes to significantly reduce energy imports and therefore lead to an improvement in the balance of payments. The adoption of these instruments will be beneficial for the entire economy. The most obvious impact is related to the level of energy imports, and therefore the balance of payments. The reduction in energy imports is estimated between 8% (low targets) and 12% (high targets) for natural gas, and between 8% (low targets) and 14% (high targets) for petroleum products. Furthermore it is estimated that the implementation of the proposed energy efficiency could create approximately 10,000 new jobs. The annual reduction in CO2 emissions has been estimated between 9 million tonnes (low targets) and 16 million tonnes (high targets)
Efficient boiler operations sourcebook
Energy Technology Data Exchange (ETDEWEB)
Payne, F.W. (comp.)
1985-01-01
This book emphasizes the practical aspects of industrial and commercial boiler operations. It starts with a comprehensive review of general combustion and boiler fundamentals and then deals with specific efficiency improvement methods, and the cost savings which result. The book has the following chapter headings: boiler combustion fundamentals; boiler efficiency goals; major factors controlling boiler efficiency; boiler efficiency calculations; heat loss; graphical solutions; preparation for boiler testing; boiler test procedures; efficiency-related boiler maintenance procedures; boiler tune-up; boiler operational modifications; effect of water side and gas side scale deposits; load management; auxillary equipment to increase boiler efficiency; air preheaters and economizers; other types of auxillary equipment; combustion control systems and instrumentation; boiler O/sub 2/ trim controls; should you purchase a new boiler.; financial evaluation procedures; case studies. The last chapter includes a case study of a boiler burning pulverized coal and a case study of stoker-fired coal.
Estimation of stochastic environment force for master–slave robotic ...
Indian Academy of Sciences (India)
Neelu Nagpal
Subsequently, convergence analysis of error in the estimates is performed. Also, an expression of ... nonlinear and composite adaptive controller [7, 9] and disturbance ... block processing method and acts as an efficient estimator since this estimation ...... 0949-2. [32] Smith L 2006 Sequential Monte Carlo particle filtering for.
A Fast Iterative Bayesian Inference Algorithm for Sparse Channel Estimation
DEFF Research Database (Denmark)
Pedersen, Niels Lovmand; Manchón, Carles Navarro; Fleury, Bernard Henri
2013-01-01
representation of the Bessel K probability density function; a highly efficient, fast iterative Bayesian inference method is then applied to the proposed model. The resulting estimator outperforms other state-of-the-art Bayesian and non-Bayesian estimators, either by yielding lower mean squared estimation error...
Joint Sparsity and Frequency Estimation for Spectral Compressive Sensing
DEFF Research Database (Denmark)
Nielsen, Jesper Kjær; Christensen, Mads Græsbøll; Jensen, Søren Holdt
2014-01-01
various interpolation techniques to estimate the continuous frequency parameters. In this paper, we show that solving the problem in a probabilistic framework instead produces an asymptotically efficient estimator which outperforms existing methods in terms of estimation accuracy while still having a low...
Efficiency of European Dairy Processing Firms
Soboh, R.A.M.E.; Oude Lansink, A.G.J.M.; Dijk, van G.
2014-01-01
This paper compares the technical efficiency and production frontier of dairy processing cooperativesand investor owned firms in six major dairy producing European countries. Two parametric produc-tion frontiers are estimated, i.e. for cooperatives and investor owned firms separately, which are
The energy efficiency of lead selfsputtering
DEFF Research Database (Denmark)
Andersen, Hans Henrik
1968-01-01
The sputtering efficiency (i.e. ratio between sputtered energy and impinging ion energy) has been measured for 30–75‐keV lead ions impinging on polycrystalline lead. The results are in good agreement with recent theoretical estimates. © 1968 The American Institute of Physics...
Cost Efficiency in Public Higher Education.
Robst, John
This study used the frontier cost function framework to examine cost efficiency in public higher education. The frontier cost function estimates the minimum predicted cost for producing a given amount of output. Data from the annual Almanac issues of the "Chronicle of Higher Education" were used to calculate state level enrollments at two-year and…
Sustaining with efficiency the renewable energy sources
International Nuclear Information System (INIS)
Bano, L.; Lorenzoni, A.
2008-01-01
European energy policy requires actions, in favour of a more widespread diffusion of renewable energy sources. Is essential to have an efficient financial support to reduce costs. Are presented an estimated of electric power from renewable energy sources and some criticism. Is proposed a modification of green certificates market based on bilateral tradable agreements [it
DEFF Research Database (Denmark)
Amore, Mario Daniele; Bennedsen, Morten; Larsen, Birthe
We study the effect of CEO education on a firm’s energy efficiency. Using a unique dataset of Danish firms, we document that firms led by more educated CEOs exhibit greater energy efficiency. We establish causality by employing exogenous CEO hospitalization episodes: the hospitalization of highly......-educated CEOs induces a drop in a firm’s energy efficiency, whereas the hospitalization of low-education CEOs does not have any significant effect. Disentangling the effect of educational length from that of the field of study, we find that the greater energy efficiency is mostly driven by the cumulated years...
Stochastic estimation of electricity consumption
International Nuclear Information System (INIS)
Kapetanovic, I.; Konjic, T.; Zahirovic, Z.
1999-01-01
Electricity consumption forecasting represents a part of the stable functioning of the power system. It is very important because of rationality and increase of control process efficiency and development planning of all aspects of society. On a scientific basis, forecasting is a possible way to solve problems. Among different models that have been used in the area of forecasting, the stochastic aspect of forecasting as a part of quantitative models takes a very important place in applications. ARIMA models and Kalman filter as stochastic estimators have been treated together for electricity consumption forecasting. Therefore, the main aim of this paper is to present the stochastic forecasting aspect using short time series. (author)
Typology of efficiency of functioning of enterprise
Directory of Open Access Journals (Sweden)
I.I. Svitlyshyn
2015-03-01
Full Text Available Measuring and estimation of the efficiency of functioning of enterprises of agrarian sector traditionally performed by applying only some of its types, which focuses mainly on operating activity. Investment and financial activity, as inalienable constituent of economic process of enterprise, remain regardless thus. In addition, in scientific literature and practical activity to research of efficiency focuses on the stages «production-exchange». The stages of «distribution» and «consumption» at the level of enterprise are not examined. This distorts the results of measuring and estimation of efficiency and makes uneffective proposals for its growth. Coming from what, approach is worked out to determination and systematization of basic types of efficiency of functioning of enterprises of agrarian sector. Approach is based on the offered model, that system represents all stages and types of economic activity of the enterprise. The basic lines of efficiency are interpreted on every stage and in the cut of types of economic activity of enterprise. It allows to provide a complexity and system during its measuring and estimation.
Enhancement of heliothermal systems efficiency
International Nuclear Information System (INIS)
Draganov, B.; Fara, L.
1994-01-01
An attempt to solve the problem of optimization of economic indicators of a solar system is presented. The three main subsystems are considered: solar collector, thermal accumulator and heat exchanger. The use of eurystic-algorithmic models is proposed to determine the degree of influence of the most important technological and energetic parameters on the extreme values of system efficiency indicators, as well as to take into account the interaction of various subsystems and the optimum of the entire system. As thermal accumulator the underground system is considered, and a mathematical model is proposed for solving the process of heat exchange and mass transfer in rocks, e.g. filtering of the thermal agent and its thermal shift. The increase in energy efficiency of solar collectors and heat exchangers is achieved by using rubbed tubes. A strong effect on the heat exchange intensification is caused through flow turbulence which could be increased by transverse prominences. The application of exergetic analysis is proposed for estimating the energetic efficiency of the solar installations and their subsystems. 7 refs. (orig.)
State Estimation for Tensegrity Robots
Caluwaerts, Ken; Bruce, Jonathan; Friesen, Jeffrey M.; Sunspiral, Vytas
2016-01-01
Tensegrity robots are a class of compliant robots that have many desirable traits when designing mass efficient systems that must interact with uncertain environments. Various promising control approaches have been proposed for tensegrity systems in simulation. Unfortunately, state estimation methods for tensegrity robots have not yet been thoroughly studied. In this paper, we present the design and evaluation of a state estimator for tensegrity robots. This state estimator will enable existing and future control algorithms to transfer from simulation to hardware. Our approach is based on the unscented Kalman filter (UKF) and combines inertial measurements, ultra wideband time-of-flight ranging measurements, and actuator state information. We evaluate the effectiveness of our method on the SUPERball, a tensegrity based planetary exploration robotic prototype. In particular, we conduct tests for evaluating both the robot's success in estimating global position in relation to fixed ranging base stations during rolling maneuvers as well as local behavior due to small-amplitude deformations induced by cable actuation.
Oil pipeline energy consumption and efficiency
Energy Technology Data Exchange (ETDEWEB)
Hooker, J.N.
1981-01-01
This report describes an investigation of energy consumption and efficiency of oil pipelines in the US in 1978. It is based on a simulation of the actual movement of oil on a very detailed representation of the pipeline network, and it uses engineering equations to calculate the energy that pipeline pumps must have exerted on the oil to move it in this manner. The efficiencies of pumps and drivers are estimated so as to arrive at the amount of energy consumed at pumping stations. The throughput in each pipeline segment is estimated by distributing each pipeline company's reported oil movements over its segments in proportions predicted by regression equations that show typical throughput and throughput capacity as functions of pipe diameter. The form of the equations is justified by a generalized cost-engineering study of pipelining, and their parameters are estimated using new techniques developed for the purpose. A simplified model of flow scheduling is chosen on the basis of actual energy use data obtained from a few companies. The study yields energy consumption and intensiveness estimates for crude oil trunk lines, crude oil gathering lines and oil products lines, for the nation as well as by state and by pipe diameter. It characterizes the efficiency of typical pipelines of various diameters operating at capacity. Ancillary results include estimates of oil movements by state and by diameter and approximate pipeline capacity utilization nationwide.
Efficient Breeding by Genomic Mating.
Akdemir, Deniz; Sánchez, Julio I
2016-01-01
Selection in breeding programs can be done by using phenotypes (phenotypic selection), pedigree relationship (breeding value selection) or molecular markers (marker assisted selection or genomic selection). All these methods are based on truncation selection, focusing on the best performance of parents before mating. In this article we proposed an approach to breeding, named genomic mating, which focuses on mating instead of truncation selection. Genomic mating uses information in a similar fashion to genomic selection but includes information on complementation of parents to be mated. Following the efficiency frontier surface, genomic mating uses concepts of estimated breeding values, risk (usefulness) and coefficient of ancestry to optimize mating between parents. We used a genetic algorithm to find solutions to this optimization problem and the results from our simulations comparing genomic selection, phenotypic selection and the mating approach indicate that current approach for breeding complex traits is more favorable than phenotypic and genomic selection. Genomic mating is similar to genomic selection in terms of estimating marker effects, but in genomic mating the genetic information and the estimated marker effects are used to decide which genotypes should be crossed to obtain the next breeding population.
Institutions, Equilibria and Efficiency
DEFF Research Database (Denmark)
Competition and efficiency is at the core of economic theory. This volume collects papers of leading scholars, which extend the conventional general equilibrium model in important ways. Efficiency and price regulation are studied when markets are incomplete and existence of equilibria in such set...... in OLG, learning in OLG and in games, optimal pricing of derivative securities, the impact of heterogeneity...
Energy Efficiency Collaboratives
Energy Technology Data Exchange (ETDEWEB)
Li, Michael [US Department of Energy, Washington, DC (United States); Bryson, Joe [US Environmental Protection Agency, Washington, DC (United States)
2015-09-01
Collaboratives for energy efficiency have a long and successful history and are currently used, in some form, in more than half of the states. Historically, many state utility commissions have used some form of collaborative group process to resolve complex issues that emerge during a rate proceeding. Rather than debate the issues through the formality of a commission proceeding, disagreeing parties are sent to discuss issues in a less-formal setting and bring back resolutions to the commission. Energy efficiency collaboratives take this concept and apply it specifically to energy efficiency programs—often in anticipation of future issues as opposed to reacting to a present disagreement. Energy efficiency collaboratives can operate long term and can address the full suite of issues associated with designing, implementing, and improving energy efficiency programs. Collaboratives can be useful to gather stakeholder input on changing program budgets and program changes in response to performance or market shifts, as well as to provide continuity while regulators come and go, identify additional energy efficiency opportunities and innovations, assess the role of energy efficiency in new regulatory contexts, and draw on lessons learned and best practices from a diverse group. Details about specific collaboratives in the United States are in the appendix to this guide. Collectively, they demonstrate the value of collaborative stakeholder processes in producing successful energy efficiency programs.
efficience technico-economique
African Journals Online (AJOL)
USER
ABSTRACT. TECHNICO-ECONOMIC EFFICIENCY : CASE OF THE PRODUCERS OF ONION AND POTATO IN KNOW IN MOROCCO. In the new context ... Key words : Technico-economic efficiency, stochastic frontier, potato production, onion production, Morocco ..... sont le labour, le cover-cropage et le traçage. 82 % des ...
Oenema, O.
2015-01-01
There is a need for communications about resource use efficiency and for measures to increase the use efficiency of nutrients in relation to food production. This holds especially for nitrogen. Nitrogen (N) is essential for life and a main nutrient element. It is needed in relatively large
Logistics, Management and Efficiency
Mircea UDRESCU; Sandu CUTURELA
2014-01-01
The problem related to the efficiency of the management for organization is general being the concern off all managers. In the present essay we consider that the efficacy of the organization begins from the structural systemization of the organizational management into general management, management of logistics and management of production which demands a new managerial process, more competitive based on economic efficiency.
Energy efficiency and behaviour
DEFF Research Database (Denmark)
Carstensen, Trine Agervig; Kunnasvirta, Annika; Kiviluoto, Katariina
separate key aspects hinders strategic energy efficiency planning. For this reason, the PLEEC project – “Planning for Energy Efficient Cities” – funded by the EU Seventh Framework Programme uses an integrative approach to achieve the sus‐ tainable, energy– efficient, smart city. By coordinating strategies...... to conduct behavioural interventions, to be presented in Deliverable 5.5., the final report. This report will also provide valuable information for the WP6 general model for an Energy-Smart City. Altogether 38 behavioural interventions are analysed in this report. Each collected and analysed case study...... of the European Union’s 20‐20‐20 plan is to improve energy efficiency by 20% in 2020. However, holistic knowledge about energy efficiency potentials in cities is far from complete. Currently, a WP4 location in PLEEC project page 3 variety of individual strategies and approaches by different stakeholders tackling...
Propulsive efficiency and non- expert swimmers performance
Directory of Open Access Journals (Sweden)
Tiago Barbosa
2009-12-01
Full Text Available Propulsive efficiency is one of the most interesting issues for competitive swimming researchers, has it presents significant relationships with the swimmer’s biophysical behavior and his/her performance. Although propulsive efficiency is a variable that has been quite studied in elite swimmers, there is no research on this issue in young and non-expert swimmers. Thus, the aim of this study was to: (i estimate the propulsive efficiency on non-expert swimmers; (ii identify biomechanical and anthropometrical parameters that are associated with propulsive efficiency; (iii identify the association between the propulsive efficiency and swim performance. Twenty-eight non-expert swimmers participated on this study. It was assessed the propulsive efficiency, biomechanical and anthropometrical parameters, as well as, the swim performance. The propulsive efficiency of non-expert swimmers is lower than data reported in the literature to higher competitive levels swimmers and there are no significant differences between boys and girls. It was also noted that several biomechanical and anthropometrical parameters, as well as, the swim performance are associated with the propulsive efficiency.
Variable Kernel Density Estimation
Terrell, George R.; Scott, David W.
1992-01-01
We investigate some of the possibilities for improvement of univariate and multivariate kernel density estimates by varying the window over the domain of estimation, pointwise and globally. Two general approaches are to vary the window width by the point of estimation and by point of the sample observation. The first possibility is shown to be of little efficacy in one variable. In particular, nearest-neighbor estimators in all versions perform poorly in one and two dimensions, but begin to b...
Adaptive measurement selection for progressive damage estimation
Zhou, Wenfan; Kovvali, Narayan; Papandreou-Suppappola, Antonia; Chattopadhyay, Aditi; Peralta, Pedro
2011-04-01
Noise and interference in sensor measurements degrade the quality of data and have a negative impact on the performance of structural damage diagnosis systems. In this paper, a novel adaptive measurement screening approach is presented to automatically select the most informative measurements and use them intelligently for structural damage estimation. The method is implemented efficiently in a sequential Monte Carlo (SMC) setting using particle filtering. The noise suppression and improved damage estimation capability of the proposed method is demonstrated by an application to the problem of estimating progressive fatigue damage in an aluminum compact-tension (CT) sample using noisy PZT sensor measurements.
Chatterji, Gano
2011-01-01
Conclusions: Validated the fuel estimation procedure using flight test data. A good fuel model can be created if weight and fuel data are available. Error in assumed takeoff weight results in similar amount of error in the fuel estimate. Fuel estimation error bounds can be determined.
Optimal fault signal estimation
Stoorvogel, Antonie Arij; Niemann, H.H.; Saberi, A.; Sannuti, P.
2002-01-01
We consider here both fault identification and fault signal estimation. Regarding fault identification, we seek either exact or almost fault identification. On the other hand, regarding fault signal estimation, we seek either $H_2$ optimal, $H_2$ suboptimal or Hinfinity suboptimal estimation. By
Nonparametric Collective Spectral Density Estimation and Clustering
Maadooliat, Mehdi
2017-04-12
In this paper, we develop a method for the simultaneous estimation of spectral density functions (SDFs) for a collection of stationary time series that share some common features. Due to the similarities among the SDFs, the log-SDF can be represented using a common set of basis functions. The basis shared by the collection of the log-SDFs is estimated as a low-dimensional manifold of a large space spanned by a pre-specified rich basis. A collective estimation approach pools information and borrows strength across the SDFs to achieve better estimation efficiency. Also, each estimated spectral density has a concise representation using the coefficients of the basis expansion, and these coefficients can be used for visualization, clustering, and classification purposes. The Whittle pseudo-maximum likelihood approach is used to fit the model and an alternating blockwise Newton-type algorithm is developed for the computation. A web-based shiny App found at
Nonparametric Collective Spectral Density Estimation and Clustering
Maadooliat, Mehdi; Sun, Ying; Chen, Tianbo
2017-01-01
In this paper, we develop a method for the simultaneous estimation of spectral density functions (SDFs) for a collection of stationary time series that share some common features. Due to the similarities among the SDFs, the log-SDF can be represented using a common set of basis functions. The basis shared by the collection of the log-SDFs is estimated as a low-dimensional manifold of a large space spanned by a pre-specified rich basis. A collective estimation approach pools information and borrows strength across the SDFs to achieve better estimation efficiency. Also, each estimated spectral density has a concise representation using the coefficients of the basis expansion, and these coefficients can be used for visualization, clustering, and classification purposes. The Whittle pseudo-maximum likelihood approach is used to fit the model and an alternating blockwise Newton-type algorithm is developed for the computation. A web-based shiny App found at
Efficiency of emergency exercises
International Nuclear Information System (INIS)
Zander, N.; Sogalla, M.
2011-01-01
In order to cope with accidents beyond the design basis within German nuclear power plants which possibly lead to relevant radiological consequences, the utilities as well as the competent authorities exist emergency organisations. The efficiency, capacity for teamwork and preparedness of such organisations should be tested by regular, efficient exercise activities. Such activities can suitably be based on scenarios which provide challenging tasks for all units of the respective emergency organisation. Thus, the demonstration and further development of the efficiency of the respective organisational structures, including their ability to collaborate, is promoted. (orig.)
Learning efficient correlated equilibria
Borowski, Holly P.; Marden, Jason R.; Shamma, Jeff S.
2014-01-01
The majority of distributed learning literature focuses on convergence to Nash equilibria. Correlated equilibria, on the other hand, can often characterize more efficient collective behavior than even the best Nash equilibrium. However, there are no existing distributed learning algorithms that converge to specific correlated equilibria. In this paper, we provide one such algorithm which guarantees that the agents' collective joint strategy will constitute an efficient correlated equilibrium with high probability. The key to attaining efficient correlated behavior through distributed learning involves incorporating a common random signal into the learning environment.
Learning efficient correlated equilibria
Borowski, Holly P.
2014-12-15
The majority of distributed learning literature focuses on convergence to Nash equilibria. Correlated equilibria, on the other hand, can often characterize more efficient collective behavior than even the best Nash equilibrium. However, there are no existing distributed learning algorithms that converge to specific correlated equilibria. In this paper, we provide one such algorithm which guarantees that the agents\\' collective joint strategy will constitute an efficient correlated equilibrium with high probability. The key to attaining efficient correlated behavior through distributed learning involves incorporating a common random signal into the learning environment.
DEFF Research Database (Denmark)
Gambalemoke, Mbalitini; Mukinzi, Itoka; Amundala, Drazo
2008-01-01
We investigated the efficiency of four trap types (pitfall, Sherman LFA, Victor snap and Museum Special snap traps) to capture shrews. This experiment was conducted in five inter-riverine forest blocks in the region of Kisangani. The total trapping effort was 6,300, 9,240, 5,280 and 5,460 trap......, our results indicate that pitfall traps are the most efficient for capturing shrews: not only do they have a higher efficiency (yield), but the taxonomic diversity of shrews is also higher when pitfall traps are used....
The Efficient Windows Collaborative
Energy Technology Data Exchange (ETDEWEB)
Petermann, Nils
2006-03-31
The Efficient Windows Collaborative (EWC) is a coalition of manufacturers, component suppliers, government agencies, research institutions, and others who partner to expand the market for energy efficient window products. Funded through a cooperative agreement with the U.S. Department of Energy, the EWC provides education, communication and outreach in order to transform the residential window market to 70% energy efficient products by 2005. Implementation of the EWC is managed by the Alliance to Save Energy, with support from the University of Minnesota and Lawrence Berkeley National Laboratory.
Unified definition of a class of Monte Carlo estimators
International Nuclear Information System (INIS)
Lux, I.
1978-01-01
A unified definition of a wide class of Monte Carlo reaction rate estimators is presented, since most commonly used estimators belong to that class. The definition is given through an integral transformation of an arbitrary estimator of the class. Since the transformation contains an arbitrary function, in principle an infinite number of new estimators can be defined on the basis of one known estimator. It is shown that the most common estimators belonging to the class, such as the track-length and expectation estimators, are special cases of transformation, corresponding to the simplest transformation kernels when transforming the usual collision estimator. A pair of new estimators is defined and their variances are compared to the variance of the expectation estimator. One of the new estimators, called the trexpectation estimator, seems to be appropriate for flux-integral estimation in moderator regions. The other one, which uses an intermediate estimation of the final result and is therefore called the self-improving estimator, always yields a lower variance than the expectation estimator. As is shown, this estimator approximates well to possibly the best estimator of the class. Numerical results are presented for the simplest geometries, and these results indicate that for absorbers that are not too strong, in practical cases the standard deviation of the self-improving estimator is less than that of the expectation estimator by more than 10%. The experiments also suggest that the self-improving estimator is always superior to the track-length estimator as well, i.e., that it is the best of all known estimators belonging to the class. In the Appendices, for simplified cases, approximate conditions are given for which the trexpectation and track-length estimators show a higher efficiency than the expectation estimator
Energy Technology Data Exchange (ETDEWEB)
NONE
2010-07-01
Transport is the sector with the highest final energy consumption and, without any significant policy changes, is forecast to remain so. In 2008, the IEA published 25 energy efficiency recommendations, among which four are for the transport sector. The recommendations focus on road transport and include policies on improving tyre energy efficiency, fuel economy standards for both light-duty vehicles and heavy-duty vehicles, and eco-driving. Implementation of the recommendations has been weaker in the transport sector than others. This paper updates the progress that has been made in implementing the transport energy efficiency recommendations in IEA countries since March 2009. Many countries have in the last year moved from 'planning to implement' to 'implementation underway', but none have fully implemented all transport energy efficiency recommendations. The IEA calls therefore for full and immediate implementation of the recommendations.
Efficient incremental relaying
Fareed, Muhammad Mehboob; Alouini, Mohamed-Slim
2013-01-01
We propose a novel relaying scheme which improves the spectral efficiency of cooperative diversity systems by utilizing limited feedback from destination. Our scheme capitalizes on the fact that relaying is only required when direct transmission
Meneghelli, Barry J.; Notardonato, William; Fesmire, James E.
2016-01-01
The Cryogenics Test Laboratory, NASA Kennedy Space Center, works to provide practical solutions to low-temperature problems while focusing on long-term technology targets for the energy-efficient use of cryogenics on Earth and in space.
Efficiency of clinical and combined diagnosis of breast cancer
International Nuclear Information System (INIS)
Solov'ev, I.E.
1986-01-01
Problems on clinical, instrumental, laboratory diagnosis of mammary glands cancer are described. Efficiency of clinical examination, mammography, cytological examination, ultrasonic, radioisotopic diagnosis, some biochemical tests are estimated. The conclusion concerning advisability of complex diagnosis of mammary glands cancer especially its early forms is made. Perspectivity of application of polyamine test in diagnosis of primary cancer of the mammary gland is mark to estimate efficiency of its treatment
Efficient incremental relaying
Fareed, Muhammad Mehboob
2013-07-01
We propose a novel relaying scheme which improves the spectral efficiency of cooperative diversity systems by utilizing limited feedback from destination. Our scheme capitalizes on the fact that relaying is only required when direct transmission suffers deep fading. We calculate the packet error rate for the proposed efficient incremental relaying scheme with both amplify and forward and decode and forward relaying. Numerical results are also presented to verify their analytical counterparts. © 2013 IEEE.
DEFF Research Database (Denmark)
Zhang, Tian
2015-01-01
The solar-to-biomass conversion efficiency of natural photosynthesis is between 2.9 and 4.3% for most crops (1, 2). Improving the efficiency of photosynthesis could help increase the appeal of biologically derived fuels and chemicals in comparison with traditional petrochemical processes. One app...... competition between biotechnology and the food industry and avoid the environmental perturbation caused by intensive agriculture (3)....
Thermodynamically efficient solar concentrators
Winston, Roland
2012-10-01
Non-imaging Optics is the theory of thermodynamically efficient optics and as such depends more on thermodynamics than on optics. Hence in this paper a condition for the "best" design is proposed based on purely thermodynamic arguments, which we believe has profound consequences for design of thermal and even photovoltaic systems. This new way of looking at the problem of efficient concentration depends on probabilities, the ingredients of entropy and information theory while "optics" in the conventional sense recedes into the background.
Efficient Windows Collaborative
Energy Technology Data Exchange (ETDEWEB)
Nils Petermann
2010-02-28
The project goals covered both the residential and commercial windows markets and involved a range of audiences such as window manufacturers, builders, homeowners, design professionals, utilities, and public agencies. Essential goals included: (1) Creation of 'Master Toolkits' of information that integrate diverse tools, rating systems, and incentive programs, customized for key audiences such as window manufacturers, design professionals, and utility programs. (2) Delivery of education and outreach programs to multiple audiences through conference presentations, publication of articles for builders and other industry professionals, and targeted dissemination of efficient window curricula to professionals and students. (3) Design and implementation of mechanisms to encourage and track sales of more efficient products through the existing Window Products Database as an incentive for manufacturers to improve products and participate in programs such as NFRC and ENERGY STAR. (4) Development of utility incentive programs to promote more efficient residential and commercial windows. Partnership with regional and local entities on the development of programs and customized information to move the market toward the highest performing products. An overarching project goal was to ensure that different audiences adopt and use the developed information, design and promotion tools and thus increase the market penetration of energy efficient fenestration products. In particular, a crucial success criterion was to move gas and electric utilities to increase the promotion of energy efficient windows through demand side management programs as an important step toward increasing the market share of energy efficient windows.
Feedback and efficient behavior.
Directory of Open Access Journals (Sweden)
Sandro Casal
Full Text Available Feedback is an effective tool for promoting efficient behavior: it enhances individuals' awareness of choice consequences in complex settings. Our study aims to isolate the mechanisms underlying the effects of feedback on achieving efficient behavior in a controlled environment. We design a laboratory experiment in which individuals are not aware of the consequences of different alternatives and, thus, cannot easily identify the efficient ones. We introduce feedback as a mechanism to enhance the awareness of consequences and to stimulate exploration and search for efficient alternatives. We assess the efficacy of three different types of intervention: provision of social information, manipulation of the frequency, and framing of feedback. We find that feedback is most effective when it is framed in terms of losses, that it reduces efficiency when it includes information about inefficient peers' behavior, and that a lower frequency of feedback does not disrupt efficiency. By quantifying the effect of different types of feedback, our study suggests useful insights for policymakers.
A logistic regression estimating function for spatial Gibbs point processes
DEFF Research Database (Denmark)
Baddeley, Adrian; Coeurjolly, Jean-François; Rubak, Ege
We propose a computationally efficient logistic regression estimating function for spatial Gibbs point processes. The sample points for the logistic regression consist of the observed point pattern together with a random pattern of dummy points. The estimating function is closely related to the p......We propose a computationally efficient logistic regression estimating function for spatial Gibbs point processes. The sample points for the logistic regression consist of the observed point pattern together with a random pattern of dummy points. The estimating function is closely related...
Fractal stock markets: International evidence of dynamical (in)efficiency
Bianchi, Sergio; Frezza, Massimiliano
2017-07-01
The last systemic financial crisis has reawakened the debate on the efficient nature of financial markets, traditionally described as semimartingales. The standard approaches to endow the general notion of efficiency of an empirical content turned out to be somewhat inconclusive and misleading. We propose a topological-based approach to quantify the informational efficiency of a financial time series. The idea is to measure the efficiency by means of the pointwise regularity of a (stochastic) function, given that the signature of a martingale is that its pointwise regularity equals 1/2 . We provide estimates for real financial time series and investigate their (in)efficient behavior by comparing three main stock indexes.
Theoretical and observational assessments of flare efficiencies
International Nuclear Information System (INIS)
Leahey, D.M.; Preston, K.; Strosher, M.
2000-01-01
During the processing of hydrocarbon materials, gaseous wastes are flared in an effort to completely burn the waste material and therefore leave behind very little by-products. Complete combustion, however is rarely successful because entrainment of air into the region of combusting gases restricts flame sizes to less than optimum values. The resulting flames are often too small to dissipate the amount of heat associated with complete (100 per cent) combustion efficiency. Flaring, therefore, often results in emissions of gases with more complex molecular structures than just carbon dioxide and water. Polycyclic aromatic hydrocarbons and volatile organic compounds which are indicative of incomplete combustion are often associated with flaring. This theoretical study of flame efficiencies was based on the knowledge of the full range of chemical reactions and associated kinetics. In this study, equations developed by Leahey and Schroeder were used to estimate flame lengths, areas and volumes as functions of flare stack exit velocity, stoichiometric mixing ratio and wind speed. This was followed by an estimate of heats released as part of the combustion process. This was derived from the knowledge of the flame dimensions together with an assumed flame temperature of 1200 K. Combustion efficiencies were then obtained by taking the ratio of estimated actual heat release values to those associated with complete combustion. It was concluded that combustion efficiency decreases significantly with wind speed increases from 1 to 6 m/s. After that initial increase, combustion efficiencies level off at values between 10 to 15 per cent. Propane and ethane were found to burn more efficiently than methane or hydrogen sulfide. 24 refs., 4 tabs., 1 fig., 1 append
Estimating Function Approaches for Spatial Point Processes
Deng, Chong
Spatial point pattern data consist of locations of events that are often of interest in biological and ecological studies. Such data are commonly viewed as a realization from a stochastic process called spatial point process. To fit a parametric spatial point process model to such data, likelihood-based methods have been widely studied. However, while maximum likelihood estimation is often too computationally intensive for Cox and cluster processes, pairwise likelihood methods such as composite likelihood, Palm likelihood usually suffer from the loss of information due to the ignorance of correlation among pairs. For many types of correlated data other than spatial point processes, when likelihood-based approaches are not desirable, estimating functions have been widely used for model fitting. In this dissertation, we explore the estimating function approaches for fitting spatial point process models. These approaches, which are based on the asymptotic optimal estimating function theories, can be used to incorporate the correlation among data and yield more efficient estimators. We conducted a series of studies to demonstrate that these estmating function approaches are good alternatives to balance the trade-off between computation complexity and estimating efficiency. First, we propose a new estimating procedure that improves the efficiency of pairwise composite likelihood method in estimating clustering parameters. Our approach combines estimating functions derived from pairwise composite likeli-hood estimation and estimating functions that account for correlations among the pairwise contributions. Our method can be used to fit a variety of parametric spatial point process models and can yield more efficient estimators for the clustering parameters than pairwise composite likelihood estimation. We demonstrate its efficacy through a simulation study and an application to the longleaf pine data. Second, we further explore the quasi-likelihood approach on fitting
Adjusting estimative prediction limits
Masao Ueki; Kaoru Fueda
2007-01-01
This note presents a direct adjustment of the estimative prediction limit to reduce the coverage error from a target value to third-order accuracy. The adjustment is asymptotically equivalent to those of Barndorff-Nielsen & Cox (1994, 1996) and Vidoni (1998). It has a simpler form with a plug-in estimator of the coverage probability of the estimative limit at the target value. Copyright 2007, Oxford University Press.
Supernovae Discovery Efficiency
John, Colin
2018-01-01
Abstract:We present supernovae (SN) search efficiency measurements for recent Hubble Space Telescope (HST) surveys. Efficiency is a key component to any search, and is important parameter as a correction factor for SN rates. To achieve an accurate value for efficiency, many supernovae need to be discoverable in surveys. This cannot be achieved from real SN only, due to their scarcity, so fake SN are planted. These fake supernovae—with a goal of realism in mind—yield an understanding of efficiency based on position related to other celestial objects, and brightness. To improve realism, we built a more accurate model of supernovae using a point-spread function. The next improvement to realism is planting these objects close to galaxies and of various parameters of brightness, magnitude, local galactic brightness and redshift. Once these are planted, a very accurate SN is visible and discoverable by the searcher. It is very important to find factors that affect this discovery efficiency. Exploring the factors that effect detection yields a more accurate correction factor. Further inquires into efficiency give us a better understanding of image processing, searching techniques and survey strategies, and result in an overall higher likelihood to find these events in future surveys with Hubble, James Webb, and WFIRST telescopes. After efficiency is discovered and refined with many unique surveys, it factors into measurements of SN rates versus redshift. By comparing SN rates vs redshift against the star formation rate we can test models to determine how long star systems take from the point of inception to explosion (delay time distribution). This delay time distribution is compared to SN progenitors models to get an accurate idea of what these stars were like before their deaths.
Hydraulic efficiency of a Rushton turbine impeller
Chara, Z.; Kysela, B.; Fort, I.
2017-07-01
Based on CFD simulations hydraulic efficiency of a standard Rushton turbine impeller in a baffled tank was determined at a Reynolds number of ReM=33330. Instantaneous values of pressure and velocity components were used to draw up the macroscopic balance of the mechanical energy. It was shown that the hydraulic efficiency of the Rushton turbine impeller (energy dissipated in a bulk volume) is about 57%. Using this result we estimated a length scale in a non-dimensional equation of kinetic energy dissipation rate in the bulk volume as L=D/2.62.