WorldWideScience

Sample records for timing estimates due

  1. Methodology for estimation of time-dependent surface heat flux due to cryogen spray cooling.

    Science.gov (United States)

    Tunnell, James W; Torres, Jorge H; Anvari, Bahman

    2002-01-01

    Cryogen spray cooling (CSC) is an effective technique to protect the epidermis during cutaneous laser therapies. Spraying a cryogen onto the skin surface creates a time-varying heat flux, effectively cooling the skin during and following the cryogen spurt. In previous studies mathematical models were developed to predict the human skin temperature profiles during the cryogen spraying time. However, no studies have accounted for the additional cooling due to residual cryogen left on the skin surface following the spurt termination. We formulate and solve an inverse heat conduction (IHC) problem to predict the time-varying surface heat flux both during and following a cryogen spurt. The IHC formulation uses measured temperature profiles from within a medium to estimate the surface heat flux. We implement a one-dimensional sequential function specification method (SFSM) to estimate the surface heat flux from internal temperatures measured within an in vitro model in response to a cryogen spurt. Solution accuracy and experimental errors are examined using simulated temperature data. Heat flux following spurt termination appears substantial; however, it is less than that during the spraying time. The estimated time-varying heat flux can subsequently be used in forward heat conduction models to estimate temperature profiles in skin during and following a cryogen spurt and predict appropriate timing for onset of the laser pulse.

  2. Estimates of over-diagnosis of breast cancer due to population-based mammography screening in South Australia after adjustment for lead time effects.

    Science.gov (United States)

    Beckmann, Kerri; Duffy, Stephen W; Lynch, John; Hiller, Janet; Farshid, Gelareh; Roder, David

    2015-09-01

    To estimate over-diagnosis due to population-based mammography screening using a lead time adjustment approach, with lead time measures based on symptomatic cancers only. Women aged 40-84 in 1989-2009 in South Australia eligible for mammography screening. Numbers of observed and expected breast cancer cases were compared, after adjustment for lead time. Lead time effects were modelled using age-specific estimates of lead time (derived from interval cancer rates and predicted background incidence, using maximum likelihood methods) and screening sensitivity, projected background breast cancer incidence rates (in the absence of screening), and proportions screened, by age and calendar year. Lead time estimates were 12, 26, 43 and 53 months, for women aged 40-49, 50-59, 60-69 and 70-79 respectively. Background incidence rates were estimated to have increased by 0.9% and 1.2% per year for invasive and all breast cancer. Over-diagnosis among women aged 40-84 was estimated at 7.9% (0.1-12.0%) for invasive cases and 12.0% (5.7-15.4%) when including ductal carcinoma in-situ (DCIS). We estimated 8% over-diagnosis for invasive breast cancer and 12% inclusive of DCIS cancers due to mammography screening among women aged 40-84. These estimates may overstate the extent of over-diagnosis if the increasing prevalence of breast cancer risk factors has led to higher background incidence than projected. © The Author(s) 2015.

  3. Estimating Global Burden of Disease due to congenital anomaly

    DEFF Research Database (Denmark)

    Boyle, Breidge; Addor, Marie-Claude; Arriola, Larraitz

    2018-01-01

    OBJECTIVE: To validate the estimates of Global Burden of Disease (GBD) due to congenital anomaly for Europe by comparing infant mortality data collected by EUROCAT registries with the WHO Mortality Database, and by assessing the significance of stillbirths and terminations of pregnancy for fetal...... the burden of disease due to congenital anomaly, and thus declining YLL over time may obscure lack of progress in primary, secondary and tertiary prevention....

  4. Maximum likelihood window for time delay estimation

    International Nuclear Information System (INIS)

    Lee, Young Sup; Yoon, Dong Jin; Kim, Chi Yup

    2004-01-01

    Time delay estimation for the detection of leak location in underground pipelines is critically important. Because the exact leak location depends upon the precision of the time delay between sensor signals due to leak noise and the speed of elastic waves, the research on the estimation of time delay has been one of the key issues in leak lovating with the time arrival difference method. In this study, an optimal Maximum Likelihood window is considered to obtain a better estimation of the time delay. This method has been proved in experiments, which can provide much clearer and more precise peaks in cross-correlation functions of leak signals. The leak location error has been less than 1 % of the distance between sensors, for example the error was not greater than 3 m for 300 m long underground pipelines. Apart from the experiment, an intensive theoretical analysis in terms of signal processing has been described. The improved leak locating with the suggested method is due to the windowing effect in frequency domain, which offers a weighting in significant frequencies.

  5. Accuracy of Travel Time Estimation using Bluetooth Technology

    DEFF Research Database (Denmark)

    Araghi, Bahar Namaki; Skoven Pedersen, Kristian; Tørholm Christensen, Lars

    2012-01-01

    Short-term travel time information plays a critical role in Advanced Traffic Information Systems (ATIS) and Advanced Traffic Management Systems (ATMS). In this context, the need for accurate and reliable travel time information sources is becoming increasingly important. Bluetooth Technology (BT......) has been used as a relatively new cost-effective source of travel time estimation. However, due to low sampling rate of BT compared to other sensor technologies, existence of outliers may significantly affect the accuracy and reliability of the travel time estimates obtained using BT. In this study......, the concept of outliers and corresponding impacts on travel time accuracy are discussed. Four different estimators named Min-BT, Max-BT, Med-BT and Avg-BT with different outlier detection logic are presented in this paper. These methods are used to estimate travel times using a BT derived dataset. In order...

  6. SU-E-T-95: Delivery Time Estimator

    International Nuclear Information System (INIS)

    Kantor, M; Balter, P; Ohrt, J

    2014-01-01

    Purpose: The development and testing of a tool for the inclusion of delivery time as a parameter in plan optimization. Methods: We developed an algorithm that estimates the time required for the machine and personnel movements required to deliver a treatment plan on a linear accelerator. We included dose rate, leaf motion, collimator motion, gantry motion, and couch motions (including time to enter the room to rotate the couch safely). Vault-specific parameters to account for time to enter and perform couch angle adjustments were also included. This algorithm works for static, step and shoot IMRT, and VMAT beams photon beams and for fixed electron beams. This was implemented as a script in our treatment planning system. We validated the estimator against actual recorded delivery time from our R and V system as well as recorded times from our IMRT QA delivery. Results: Data was collected (Figure 1) for 12 treatment plans by examining the R and V beam start times, and by manually timing the QA treatment for a reference, but the QA measurements were only significant to the nearest minute. The average difference between the estimated and R and V times was 15%, and 11% when excluding the major outliers. Outliers arose due to respiratory aides and gating techniques which could not be accounted for in the estimator. Conclusion: Non-mechanical factors such as the time a therapist needs to walk in and out of the room to adjust the couch needed to be fine-tuned and cycled back into the algorithm to improve the estimate. The algorithm has been demonstrated to provide reasonable and useful estimates for delivery time. This estimate has provided a useful additional input for clinical decision-making when comparing several potential radiation treatment options

  7. An Analysis of Variance Approach for the Estimation of Response Time Distributions in Tests

    Science.gov (United States)

    Attali, Yigal

    2010-01-01

    Generalizability theory and analysis of variance methods are employed, together with the concept of objective time pressure, to estimate response time distributions and the degree of time pressure in timed tests. By estimating response time variance components due to person, item, and their interaction, and fixed effects due to item types and…

  8. On the fast estimation of transit times application to BWR simulated data

    International Nuclear Information System (INIS)

    Antonopoulos-Domis, M.; Marseguerra, M.; Padovani, E.

    1996-01-01

    Real time estimators of transit times are proposed. BWR noise is simulated including a global component due to rod vibration. The time obtained form the simulation is used to investigate the robustness and noise immunity of the estimators. It is found that, in presence of a coincident (global) signal, the cross-correlation function is the worst estimator. (authors)

  9. Hygienic estimation of population doses due to stratospheric fallout

    International Nuclear Information System (INIS)

    Marej, A.N.; Knizhnikov, V.A.

    1980-01-01

    The hygienic estimation of external and internal irradiation of the USSR population due to stratospheric global fallouts of fission products after nuclear explosions and weapon tests, is carried out. Numerical values which characterize the dose-effect dependence in the case of radiation of marrow, bone tissue and whole body are presented. Values of mean individual and population doses of irradiation due to global fallouts within 1963-1975, types of injury and the number of mortal cases due to malignant neoplasms are presented. A conclusion is made that the contribution of radiation due to stratospheric fallouts in the mortality due to malignant neoplasms is insignificant. Annual radiation doses, conditioned by global fallouts within the period of 1963-1975 constitute but several percent from the dose of radiation of the natural radiation background. Results of estimation of genetic consequences of irradiation due to atmospheric fallouts are presented

  10. Exploratory Study for Continuous-time Parameter Estimation of Ankle Dynamics

    Science.gov (United States)

    Kukreja, Sunil L.; Boyle, Richard D.

    2014-01-01

    Recently, a parallel pathway model to describe ankle dynamics was proposed. This model provides a relationship between ankle angle and net ankle torque as the sum of a linear and nonlinear contribution. A technique to identify parameters of this model in discrete-time has been developed. However, these parameters are a nonlinear combination of the continuous-time physiology, making insight into the underlying physiology impossible. The stable and accurate estimation of continuous-time parameters is critical for accurate disease modeling, clinical diagnosis, robotic control strategies, development of optimal exercise protocols for longterm space exploration, sports medicine, etc. This paper explores the development of a system identification technique to estimate the continuous-time parameters of ankle dynamics. The effectiveness of this approach is assessed via simulation of a continuous-time model of ankle dynamics with typical parameters found in clinical studies. The results show that although this technique improves estimates, it does not provide robust estimates of continuous-time parameters of ankle dynamics. Due to this we conclude that alternative modeling strategies and more advanced estimation techniques be considered for future work.

  11. Estimate of the global-scale joule heating rates in the thermosphere due to time mean currents

    International Nuclear Information System (INIS)

    Roble, R.G.; Matsushita, S.

    1975-01-01

    An estimate of the global-scale joule heating rates in the thermosphere is made based on derived global equivalent overhead electric current systems in the dynamo region during geomagnetically quiet and disturbed periods. The equivalent total electric field distribution is calculated from Ohm's law. The global-scale joule heating rates are calculated for various monthly average periods in 1965. The calculated joule heating rates maximize at high latitudes in the early evening and postmidnight sectors. During geomagnetically quiet times the daytime joule heating rates are considerably lower than heating by solar EUV radiation. However, during geomagnetically disturbed periods the estimated joule heating rates increase by an order of magnitude and can locally exceed the solar EUV heating rates. The results show that joule heating is an important and at times the dominant energy source at high latitudes. However, the global mean joule heating rates calculated near solar minimum are generally small compared to the global mean solar EUV heating rates. (auth)

  12. Toward a Real-Time Measurement-Based System for Estimation of Helicopter Engine Degradation Due to Compressor Erosion

    Science.gov (United States)

    Litt, Jonathan S.; Simo, Donald L.

    2007-01-01

    This paper presents a preliminary demonstration of an automated health assessment tool, capable of real-time on-board operation using existing engine control hardware. The tool allows operators to discern how rapidly individual turboshaft engines are degrading. As the compressor erodes, performance is lost, and with it the ability to generate power. Thus, such a tool would provide an instant assessment of the engine s fitness to perform a mission, and would help to pinpoint any abnormal wear or performance anomalies before they became serious, thereby decreasing uncertainty and enabling improved maintenance scheduling. The research described in the paper utilized test stand data from a T700-GE-401 turboshaft engine that underwent sand-ingestion testing to scale a model-based compressor efficiency degradation estimation algorithm. This algorithm was then applied to real-time Health Usage and Monitoring System (HUMS) data from a T700-GE-701C to track compressor efficiency on-line. The approach uses an optimal estimator called a Kalman filter. The filter is designed to estimate the compressor efficiency using only data from the engine s sensors as input.

  13. Estimation of Surface Deformation due to Pasni Earthquake Using SAR Interferometry

    Science.gov (United States)

    Ali, M.; Shahzad, M. I.; Nazeer, M.; Kazmi, J. H.

    2018-04-01

    Earthquake cause ground deformation in sedimented surface areas like Pasni and that is a hazard. Such earthquake induced ground displacements can seriously damage building structures. On 7 February 2017, an earthquake with 6.3 magnitudes strike near to Pasni. We have successfully distinguished widely spread ground displacements for the Pasni earthquake by using InSAR-based analysis with Sentinel-1 satellite C-band data. The maps of surface displacement field resulting from the earthquake are generated. Sentinel-1 Wide Swath data acquired from 9 December 2016 to 28 February 2017 was used to generate displacement map. The interferogram revealed the area of deformation. The comparison map of interferometric vertical displacement in different time period was treated as an evidence of deformation caused by earthquake. Profile graphs of interferogram were created to estimate the vertical displacement range and trend. Pasni lies in strong earthquake magnitude effected area. The major surface deformation areas are divided into different zones based on significance of deformation. The average displacement in Pasni is estimated about 250 mm. Maximum pasni area is uplifted by earthquake and maximum uplifting occurs was about 1200 mm. Some of areas was subsidized like the areas near to shoreline and maximum subsidence was estimated about 1500 mm. Pasni is facing many problems due to increasing sea water intrusion under prevailing climatic change where land deformation due to a strong earthquake can augment its vulnerability.

  14. Estimation of train dwell time at short stops based on track occupation event data

    NARCIS (Netherlands)

    Li, D.; Daamen, W.; Goverde, R.M.P.

    2015-01-01

    Train dwell time is one of the most unpredictable components of railway operations mainly due to the varying volumes of alighting and boarding passengers. For reliable estimations of train running times and route conflicts on main lines it is however necessary to obtain accurate estimations of dwell

  15. Accuracy of prehospital transport time estimation.

    Science.gov (United States)

    Wallace, David J; Kahn, Jeremy M; Angus, Derek C; Martin-Gill, Christian; Callaway, Clifton W; Rea, Thomas D; Chhatwal, Jagpreet; Kurland, Kristen; Seymour, Christopher W

    2014-01-01

    Estimates of prehospital transport times are an important part of emergency care system research and planning; however, the accuracy of these estimates is unknown. The authors examined the accuracy of three estimation methods against observed transport times in a large cohort of prehospital patient transports. This was a validation study using prehospital records in King County, Washington, and southwestern Pennsylvania from 2002 to 2006 and 2005 to 2011, respectively. Transport time estimates were generated using three methods: linear arc distance, Google Maps, and ArcGIS Network Analyst. Estimation error, defined as the absolute difference between observed and estimated transport time, was assessed, as well as the proportion of estimated times that were within specified error thresholds. Based on the primary results, a regression estimate was used that incorporated population density, time of day, and season to assess improved accuracy. Finally, hospital catchment areas were compared using each method with a fixed drive time. The authors analyzed 29,935 prehospital transports to 44 hospitals. The mean (± standard deviation [±SD]) absolute error was 4.8 (±7.3) minutes using linear arc, 3.5 (±5.4) minutes using Google Maps, and 4.4 (±5.7) minutes using ArcGIS. All pairwise comparisons were statistically significant (p Google Maps, and 11.6 [±10.9] minutes for ArcGIS). Estimates were within 5 minutes of observed transport time for 79% of linear arc estimates, 86.6% of Google Maps estimates, and 81.3% of ArcGIS estimates. The regression-based approach did not substantially improve estimation. There were large differences in hospital catchment areas estimated by each method. Route-based transport time estimates demonstrate moderate accuracy. These methods can be valuable for informing a host of decisions related to the system organization and patient access to emergency medical care; however, they should be employed with sensitivity to their limitations.

  16. Two-Step Time of Arrival Estimation for Pulse-Based Ultra-Wideband Systems

    Directory of Open Access Journals (Sweden)

    H. Vincent Poor

    2008-05-01

    Full Text Available In cooperative localization systems, wireless nodes need to exchange accurate position-related information such as time-of-arrival (TOA and angle-of-arrival (AOA, in order to obtain accurate location information. One alternative for providing accurate position-related information is to use ultra-wideband (UWB signals. The high time resolution of UWB signals presents a potential for very accurate positioning based on TOA estimation. However, it is challenging to realize very accurate positioning systems in practical scenarios, due to both complexity/cost constraints and adverse channel conditions such as multipath propagation. In this paper, a two-step TOA estimation algorithm is proposed for UWB systems in order to provide accurate TOA estimation under practical constraints. In order to speed up the estimation process, the first step estimates a coarse TOA of the received signal based on received signal energy. Then, in the second step, the arrival time of the first signal path is estimated by considering a hypothesis testing approach. The proposed scheme uses low-rate correlation outputs and is able to perform accurate TOA estimation in reasonable time intervals. The simulation results are presented to analyze the performance of the estimator.

  17. An Efficient Code-Timing Estimator for DS-CDMA Systems over Resolvable Multipath Channels

    Directory of Open Access Journals (Sweden)

    Jian Li

    2005-04-01

    Full Text Available We consider the problem of training-based code-timing estimation for the asynchronous direct-sequence code-division multiple-access (DS-CDMA system. We propose a modified large-sample maximum-likelihood (MLSML estimator that can be used for the code-timing estimation for the DS-CDMA systems over the resolvable multipath channels in closed form. Simulation results show that MLSML can be used to provide a high correct acquisition probability and a high estimation accuracy. Simulation results also show that MLSML can have very good near-far resistant capability due to employing a data model similar to that for adaptive array processing where strong interferences can be suppressed.

  18. Finite Time Fault Tolerant Control for Robot Manipulators Using Time Delay Estimation and Continuous Nonsingular Fast Terminal Sliding Mode Control.

    Science.gov (United States)

    Van, Mien; Ge, Shuzhi Sam; Ren, Hongliang

    2016-04-28

    In this paper, a novel finite time fault tolerant control (FTC) is proposed for uncertain robot manipulators with actuator faults. First, a finite time passive FTC (PFTC) based on a robust nonsingular fast terminal sliding mode control (NFTSMC) is investigated. Be analyzed for addressing the disadvantages of the PFTC, an AFTC are then investigated by combining NFTSMC with a simple fault diagnosis scheme. In this scheme, an online fault estimation algorithm based on time delay estimation (TDE) is proposed to approximate actuator faults. The estimated fault information is used to detect, isolate, and accommodate the effect of the faults in the system. Then, a robust AFTC law is established by combining the obtained fault information and a robust NFTSMC. Finally, a high-order sliding mode (HOSM) control based on super-twisting algorithm is employed to eliminate the chattering. In comparison to the PFTC and other state-of-the-art approaches, the proposed AFTC scheme possess several advantages such as high precision, strong robustness, no singularity, less chattering, and fast finite-time convergence due to the combined NFTSMC and HOSM control, and requires no prior knowledge of the fault due to TDE-based fault estimation. Finally, simulation results are obtained to verify the effectiveness of the proposed strategy.

  19. Bayesian Nonparametric Mixture Estimation for Time-Indexed Functional Data in R

    Directory of Open Access Journals (Sweden)

    Terrance D. Savitsky

    2016-08-01

    Full Text Available We present growfunctions for R that offers Bayesian nonparametric estimation models for analysis of dependent, noisy time series data indexed by a collection of domains. This data structure arises from combining periodically published government survey statistics, such as are reported in the Current Population Study (CPS. The CPS publishes monthly, by-state estimates of employment levels, where each state expresses a noisy time series. Published state-level estimates from the CPS are composed from household survey responses in a model-free manner and express high levels of volatility due to insufficient sample sizes. Existing software solutions borrow information over a modeled time-based dependence to extract a de-noised time series for each domain. These solutions, however, ignore the dependence among the domains that may be additionally leveraged to improve estimation efficiency. The growfunctions package offers two fully nonparametric mixture models that simultaneously estimate both a time and domain-indexed dependence structure for a collection of time series: (1 A Gaussian process (GP construction, which is parameterized through the covariance matrix, estimates a latent function for each domain. The covariance parameters of the latent functions are indexed by domain under a Dirichlet process prior that permits estimation of the dependence among functions across the domains: (2 An intrinsic Gaussian Markov random field prior construction provides an alternative to the GP that expresses different computation and estimation properties. In addition to performing denoised estimation of latent functions from published domain estimates, growfunctions allows estimation of collections of functions for observation units (e.g., households, rather than aggregated domains, by accounting for an informative sampling design under which the probabilities for inclusion of observation units are related to the response variable. growfunctions includes plot

  20. Measurement error in mobile source air pollution exposure estimates due to residential mobility during pregnancy.

    Science.gov (United States)

    Pennington, Audrey Flak; Strickland, Matthew J; Klein, Mitchel; Zhai, Xinxin; Russell, Armistead G; Hansen, Craig; Darrow, Lyndsey A

    2017-09-01

    Prenatal air pollution exposure is frequently estimated using maternal residential location at the time of delivery as a proxy for residence during pregnancy. We describe residential mobility during pregnancy among 19,951 children from the Kaiser Air Pollution and Pediatric Asthma Study, quantify measurement error in spatially resolved estimates of prenatal exposure to mobile source fine particulate matter (PM 2.5 ) due to ignoring this mobility, and simulate the impact of this error on estimates of epidemiologic associations. Two exposure estimates were compared, one calculated using complete residential histories during pregnancy (weighted average based on time spent at each address) and the second calculated using only residence at birth. Estimates were computed using annual averages of primary PM 2.5 from traffic emissions modeled using a Research LINE-source dispersion model for near-surface releases (RLINE) at 250 m resolution. In this cohort, 18.6% of children were born to mothers who moved at least once during pregnancy. Mobile source PM 2.5 exposure estimates calculated using complete residential histories during pregnancy and only residence at birth were highly correlated (r S >0.9). Simulations indicated that ignoring residential mobility resulted in modest bias of epidemiologic associations toward the null, but varied by maternal characteristics and prenatal exposure windows of interest (ranging from -2% to -10% bias).

  1. Sensitivity of Process Design due to Uncertainties in Property Estimates

    DEFF Research Database (Denmark)

    Hukkerikar, Amol; Jones, Mark Nicholas; Sarup, Bent

    2012-01-01

    The objective of this paper is to present a systematic methodology for performing analysis of sensitivity of process design due to uncertainties in property estimates. The methodology provides the following results: a) list of properties with critical importance on design; b) acceptable levels of...... in chemical processes. Among others vapour pressure accuracy for azeotropic mixtures is critical and needs to be measured or estimated with a ±0.25% accuracy to satisfy acceptable safety levels in design....

  2. Error due to unresolved scales in estimation problems for atmospheric data assimilation

    Science.gov (United States)

    Janjic, Tijana

    The error arising due to unresolved scales in data assimilation procedures is examined. The problem of estimating the projection of the state of a passive scalar undergoing advection at a sequence of times is considered. The projection belongs to a finite- dimensional function space and is defined on the continuum. Using the continuum projection of the state of a passive scalar, a mathematical definition is obtained for the error arising due to the presence, in the continuum system, of scales unresolved by the discrete dynamical model. This error affects the estimation procedure through point observations that include the unresolved scales. In this work, two approximate methods for taking into account the error due to unresolved scales and the resulting correlations are developed and employed in the estimation procedure. The resulting formulas resemble the Schmidt-Kalman filter and the usual discrete Kalman filter, respectively. For this reason, the newly developed filters are called the Schmidt-Kalman filter and the traditional filter. In order to test the assimilation methods, a two- dimensional advection model with nonstationary spectrum was developed for passive scalar transport in the atmosphere. An analytical solution on the sphere was found depicting the model dynamics evolution. Using this analytical solution the model error is avoided, and the error due to unresolved scales is the only error left in the estimation problem. It is demonstrated that the traditional and the Schmidt- Kalman filter work well provided the exact covariance function of the unresolved scales is known. However, this requirement is not satisfied in practice, and the covariance function must be modeled. The Schmidt-Kalman filter cannot be computed in practice without further approximations. Therefore, the traditional filter is better suited for practical use. Also, the traditional filter does not require modeling of the full covariance function of the unresolved scales, but only

  3. HIV due to female sex work: regional and global estimates.

    Directory of Open Access Journals (Sweden)

    Annette Prüss-Ustün

    Full Text Available Female sex workers (FSWs are at high risk of HIV infection. Our objective was to determine the proportion of HIV prevalence in the general female adult population that is attributable to the occupational exposure of female sex work, due to unprotected sexual intercourse.Population attributable fractions of HIV prevalence due to female sex work were estimated for 2011. A systematic search was conducted to retrieve required input data from available sources. Data gaps of HIV prevalence in FSWs for 2011 were filled using multilevel modeling and multivariate linear regression. The fraction of HIV attributable to female sex work was estimated as the excess HIV burden in FSWs deducting the HIV burden in FSWs due to injecting drug use.An estimated fifteen percent of HIV in the general female adult population is attributable to (unsafe female sex work. The region with the highest attributable fraction is Sub Saharan Africa, but the burden is also substantial for the Caribbean, Latin America and South and Southeast Asia. We estimate 106,000 deaths from HIV are a result of female sex work globally, 98,000 of which occur in Sub-Saharan Africa. If HIV prevalence in other population groups originating from sexual contact with FSWs had been considered, the overall attributable burden would probably be much larger.Female sex work is an important contributor to HIV transmission and the global HIV burden. Effective HIV prevention measures exist and have been successfully targeted at key populations in many settings. These must be scaled up.FSWs suffer from high HIV burden and are a crucial core population for HIV transmission. Surveillance, prevention and treatment of HIV in FSWs should benefit both this often neglected vulnerable group and the general population.

  4. HIV Due to Female Sex Work: Regional and Global Estimates

    Science.gov (United States)

    Prüss-Ustün, Annette; Wolf, Jennyfer; Driscoll, Tim; Degenhardt, Louisa; Neira, Maria; Calleja, Jesus Maria Garcia

    2013-01-01

    Introduction Female sex workers (FSWs) are at high risk of HIV infection. Our objective was to determine the proportion of HIV prevalence in the general female adult population that is attributable to the occupational exposure of female sex work, due to unprotected sexual intercourse. Methods Population attributable fractions of HIV prevalence due to female sex work were estimated for 2011. A systematic search was conducted to retrieve required input data from available sources. Data gaps of HIV prevalence in FSWs for 2011 were filled using multilevel modeling and multivariate linear regression. The fraction of HIV attributable to female sex work was estimated as the excess HIV burden in FSWs deducting the HIV burden in FSWs due to injecting drug use. Results An estimated fifteen percent of HIV in the general female adult population is attributable to (unsafe) female sex work. The region with the highest attributable fraction is Sub Saharan Africa, but the burden is also substantial for the Caribbean, Latin America and South and Southeast Asia. We estimate 106,000 deaths from HIV are a result of female sex work globally, 98,000 of which occur in Sub-Saharan Africa. If HIV prevalence in other population groups originating from sexual contact with FSWs had been considered, the overall attributable burden would probably be much larger. Discussion Female sex work is an important contributor to HIV transmission and the global HIV burden. Effective HIV prevention measures exist and have been successfully targeted at key populations in many settings. These must be scaled up. Conclusion FSWs suffer from high HIV burden and are a crucial core population for HIV transmission. Surveillance, prevention and treatment of HIV in FSWs should benefit both this often neglected vulnerable group and the general population. PMID:23717432

  5. Estimating anesthesia and surgical procedure times from medicare anesthesia claims.

    Science.gov (United States)

    Silber, Jeffrey H; Rosenbaum, Paul R; Zhang, Xuemei; Even-Shoshan, Orit

    2007-02-01

    Procedure times are important variables that often are included in studies of quality and efficiency. However, due to the need for costly chart review, most studies are limited to single-institution analyses. In this article, the authors describe how well the anesthesia claim from Medicare can estimate chart times. The authors abstracted information on time of induction and entrance to the recovery room ("anesthesia chart time") from the charts of 1,931 patients who underwent general and orthopedic surgical procedures in Pennsylvania. The authors then merged the associated bills from claims data supplied from Medicare (Part B data) that included a variable denoting the time in minutes for the anesthesia service. The authors also investigated the time from incision to closure ("surgical chart time") on a subset of 1,888 patients. Anesthesia claim time from Medicare was highly predictive of anesthesia chart time (Kendall's rank correlation tau = 0.85, P < 0.0001, median absolute error = 5.1 min) but somewhat less predictive of surgical chart time (Kendall's tau = 0.73, P < 0.0001, median absolute error = 13.8 min). When predicting chart time from Medicare bills, variables reflecting procedure type, comorbidities, and hospital type did not significantly improve the prediction, suggesting that errors in predicting the chart time from the anesthesia bill time are not related to these factors; however, the individual hospital did have some influence on these estimates. Anesthesia chart time can be well estimated using Medicare claims, thereby facilitating studies with vastly larger sample sizes and much lower costs of data collection.

  6. Estimating Global Burden of Disease due to congenital anomaly: an analysis of European data

    Science.gov (United States)

    Boyle, Breidge; Addor, Marie-Claude; Arriola, Larraitz; Barisic, Ingeborg; Bianchi, Fabrizio; Csáky-Szunyogh, Melinda; de Walle, Hermien E K; Dias, Carlos Matias; Draper, Elizabeth; Gatt, Miriam; Garne, Ester; Haeusler, Martin; Källén, Karin; Latos-Bielenska, Anna; McDonnell, Bob; Mullaney, Carmel; Nelen, Vera; Neville, Amanda J; O’Mahony, Mary; Queisser-Wahrendorf, Annette; Randrianaivo, Hanitra; Rankin, Judith; Rissmann, Anke; Ritvanen, Annukka; Rounding, Catherine; Tucker, David; Verellen-Dumoulin, Christine; Wellesley, Diana; Wreyford, Ben; Zymak-Zakutnia, Natalia; Dolk, Helen

    2018-01-01

    Objective To validate the estimates of Global Burden of Disease (GBD) due to congenital anomaly for Europe by comparing infant mortality data collected by EUROCAT registries with the WHO Mortality Database, and by assessing the significance of stillbirths and terminations of pregnancy for fetal anomaly (TOPFA) in the interpretation of infant mortality statistics. Design, setting and outcome measures EUROCAT is a network of congenital anomaly registries collecting data on live births, fetal deaths from 20 weeks’ gestation and TOPFA. Data from 29 registries in 19 countries were analysed for 2005–2009, and infant mortality (deaths of live births at age congenital anomaly. In 11 EUROCAT countries, average infant mortality with congenital anomaly was 1.1 per 1000 births, with higher rates where TOPFA is illegal (Malta 3.0, Ireland 2.1). The rate of stillbirths with congenital anomaly was 0.6 per 1000. The average TOPFA prevalence was 4.6 per 1000, nearly three times more prevalent than stillbirths and infant deaths combined. TOPFA also impacted on the prevalence of postneonatal survivors with non-lethal congenital anomaly. Conclusions By excluding TOPFA and stillbirths from GBD years of life lost (YLL) estimates, GBD underestimates the burden of disease due to congenital anomaly, and thus declining YLL over time may obscure lack of progress in primary, secondary and tertiary prevention. PMID:28667189

  7. Accurate estimation of indoor travel times

    DEFF Research Database (Denmark)

    Prentow, Thor Siiger; Blunck, Henrik; Stisen, Allan

    2014-01-01

    The ability to accurately estimate indoor travel times is crucial for enabling improvements within application areas such as indoor navigation, logistics for mobile workers, and facility management. In this paper, we study the challenges inherent in indoor travel time estimation, and we propose...... the InTraTime method for accurately estimating indoor travel times via mining of historical and real-time indoor position traces. The method learns during operation both travel routes, travel times and their respective likelihood---both for routes traveled as well as for sub-routes thereof. InTraTime...... allows to specify temporal and other query parameters, such as time-of-day, day-of-week or the identity of the traveling individual. As input the method is designed to take generic position traces and is thus interoperable with a variety of indoor positioning systems. The method's advantages include...

  8. Parameter Estimation in Continuous Time Domain

    Directory of Open Access Journals (Sweden)

    Gabriela M. ATANASIU

    2016-12-01

    Full Text Available This paper will aim to presents the applications of a continuous-time parameter estimation method for estimating structural parameters of a real bridge structure. For the purpose of illustrating this method two case studies of a bridge pile located in a highly seismic risk area are considered, for which the structural parameters for the mass, damping and stiffness are estimated. The estimation process is followed by the validation of the analytical results and comparison with them to the measurement data. Further benefits and applications for the continuous-time parameter estimation method in civil engineering are presented in the final part of this paper.

  9. Algorithms for Brownian first-passage-time estimation

    Science.gov (United States)

    Adib, Artur B.

    2009-09-01

    A class of algorithms in discrete space and continuous time for Brownian first-passage-time estimation is considered. A simple algorithm is derived that yields exact mean first-passage times (MFPTs) for linear potentials in one dimension, regardless of the lattice spacing. When applied to nonlinear potentials and/or higher spatial dimensions, numerical evidence suggests that this algorithm yields MFPT estimates that either outperform or rival Langevin-based (discrete time and continuous space) estimates.

  10. Estimating Global Burden of Disease due to congenital anomaly: an analysis of European data.

    Science.gov (United States)

    Boyle, Breidge; Addor, Marie-Claude; Arriola, Larraitz; Barisic, Ingeborg; Bianchi, Fabrizio; Csáky-Szunyogh, Melinda; de Walle, Hermien E K; Dias, Carlos Matias; Draper, Elizabeth; Gatt, Miriam; Garne, Ester; Haeusler, Martin; Källén, Karin; Latos-Bielenska, Anna; McDonnell, Bob; Mullaney, Carmel; Nelen, Vera; Neville, Amanda J; O'Mahony, Mary; Queisser-Wahrendorf, Annette; Randrianaivo, Hanitra; Rankin, Judith; Rissmann, Anke; Ritvanen, Annukka; Rounding, Catherine; Tucker, David; Verellen-Dumoulin, Christine; Wellesley, Diana; Wreyford, Ben; Zymak-Zakutnia, Natalia; Dolk, Helen

    2018-01-01

    To validate the estimates of Global Burden of Disease (GBD) due to congenital anomaly for Europe by comparing infant mortality data collected by EUROCAT registries with the WHO Mortality Database, and by assessing the significance of stillbirths and terminations of pregnancy for fetal anomaly (TOPFA) in the interpretation of infant mortality statistics. EUROCAT is a network of congenital anomaly registries collecting data on live births, fetal deaths from 20 weeks' gestation and TOPFA. Data from 29 registries in 19 countries were analysed for 2005-2009, and infant mortality (deaths of live births at age congenital anomaly. In 11 EUROCAT countries, average infant mortality with congenital anomaly was 1.1 per 1000 births, with higher rates where TOPFA is illegal (Malta 3.0, Ireland 2.1). The rate of stillbirths with congenital anomaly was 0.6 per 1000. The average TOPFA prevalence was 4.6 per 1000, nearly three times more prevalent than stillbirths and infant deaths combined. TOPFA also impacted on the prevalence of postneonatal survivors with non-lethal congenital anomaly. By excluding TOPFA and stillbirths from GBD years of life lost (YLL) estimates, GBD underestimates the burden of disease due to congenital anomaly, and thus declining YLL over time may obscure lack of progress in primary, secondary and tertiary prevention. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.

  11. Life time estimation for irradiation assisted mechanical cracking of PWR RCCA rodlets

    Energy Technology Data Exchange (ETDEWEB)

    Matsuoka, Takanori; Yamaguchi, Youichirou [Nuclear Development Corp., Tokai, Ibaraki (Japan)

    1999-09-01

    Intergranular cracks of cladding tubes had been observed at the tips of the rodlets of PWR rod cluster control assemblies (RCCAs). Because RCCAs were important core components, an investigation was carried out to estimate their service lifetime. The reviews on their mechanism and the life time estimation are shown in this paper. The summaries are as follows. (1) The mechanism of the intergranular crack of the cladding tube was not IASCC but irradiation assisted mechanical cracking (IAMC) caused by an increase in hoop strain due to the swelling of the absorber and a decrease in elongation due to neutron irradiation. (2) The crack initiation limit of cylindrical shells made of low ductile material and subjected to internal pressure was determined in relation to the uniform strain of the material and was in accordance with that of the RCCA rodlets in an actual plant. (3) From the above investigation, the method of estimating the lifetime and countermeasures for its extension were obtained. (author)

  12. DUE GlobBiomass - Estimates of Biomass on a Global Scale

    Science.gov (United States)

    Eberle, J.; Schmullius, C.

    2017-12-01

    For the last three years, a new ESA Data User Element (DUE) project had focussed on creating improved knowledge about the Essential Climate Variable Biomass. The main purpose of the DUE GlobBiomass project is to better characterize and to reduce uncertainties of AGB estimates by developing an innovative synergistic mapping approach in five regional sites (Sweden, Poland, Mexico, Kalimantan, South Africa) for the epochs 2005, 2010 and 2015 and for one global map for the year 2010. The project team includes leading Earth Observation experts of Europe and is linked through Partnership Agreements with further national bodies from Brazil, Canada, China, Russia and South Africa. GlobBiomass has demonstrated how EO observation data can be integrated with in situ measurements and ecological understanding to provide improved biomass estimates that can be effectively exploited by users. The target users had mainly be drawn from the climate and carbon cycle modelling communities and included users concerned with carbon emissions and uptake due to biomass changes within initiatives such as REDD+. GlobBiomass provided a harmonised structure that can be exploited to address user needs for biomass information, but will be capable of being progressively refined as new data and methods become available. This presentation will give an overview of the technical prerequisites and final results of the GlobBiomass project.

  13. Radiochemical separation and effective dose estimation due to ingestion of 90Sr

    International Nuclear Information System (INIS)

    Ilic, Z.; Vidic, A.; Deljkic, D.; Sirko, D.; Zovko, E.; Samek, D.

    2009-01-01

    Since 2007. Institute for Public Health of Federation of Bosnia and Herzegovina-Radiation Protection Centre, within the framework of monitoring of radioactivity of environment carried out measurement of specific activity of 90 Sr content in selected food and water samples. The paper described the methods of measurement and radiochemical separation. Presented results, as average values of specific activity of 90 Sr, were used for estimation of effective dose due to ingestion of 90 Sr for 2007. and 2008. Estimated effective dose for 2007. due to ingestion of 90 Sr for adults was 1,36 μSv and 2,03 μSv for children (10 year old), and for 2008. 0,67 μSv (adults) and 1,01 μSv (children 10 year old). Estimated effective doses for 2007. and 2008. are varied because of different average specific activity radionuclide 90 Sr in selected samples of food, their number, species and origin. (author) [sr

  14. Real-time fault-tolerant moving horizon air data estimation for the RECONFIGURE benchmark

    NARCIS (Netherlands)

    Wan, Y.; Keviczky, T.

    2018-01-01

    This paper proposes a real-time fault-tolerant estimation approach for combined sensor fault diagnosis and air data reconstruction. Due to simultaneous influence of winds and latent faults on monitored sensors, it is challenging to address the tradeoff between robustness to wind disturbances and

  15. Hydrological model uncertainty due to spatial evapotranspiration estimation methods

    Science.gov (United States)

    Yu, Xuan; Lamačová, Anna; Duffy, Christopher; Krám, Pavel; Hruška, Jakub

    2016-05-01

    Evapotranspiration (ET) continues to be a difficult process to estimate in seasonal and long-term water balances in catchment models. Approaches to estimate ET typically use vegetation parameters (e.g., leaf area index [LAI], interception capacity) obtained from field observation, remote sensing data, national or global land cover products, and/or simulated by ecosystem models. In this study we attempt to quantify the uncertainty that spatial evapotranspiration estimation introduces into hydrological simulations when the age of the forest is not precisely known. The Penn State Integrated Hydrologic Model (PIHM) was implemented for the Lysina headwater catchment, located 50°03‧N, 12°40‧E in the western part of the Czech Republic. The spatial forest patterns were digitized from forest age maps made available by the Czech Forest Administration. Two ET methods were implemented in the catchment model: the Biome-BGC forest growth sub-model (1-way coupled to PIHM) and with the fixed-seasonal LAI method. From these two approaches simulation scenarios were developed. We combined the estimated spatial forest age maps and two ET estimation methods to drive PIHM. A set of spatial hydrologic regime and streamflow regime indices were calculated from the modeling results for each method. Intercomparison of the hydrological responses to the spatial vegetation patterns suggested considerable variation in soil moisture and recharge and a small uncertainty in the groundwater table elevation and streamflow. The hydrologic modeling with ET estimated by Biome-BGC generated less uncertainty due to the plant physiology-based method. The implication of this research is that overall hydrologic variability induced by uncertain management practices was reduced by implementing vegetation models in the catchment models.

  16. Towards Real-Time Maneuver Detection: Automatic State and Dynamics Estimation with the Adaptive Optimal Control Based Estimator

    Science.gov (United States)

    Lubey, D.; Scheeres, D.

    Tracking objects in Earth orbit is fraught with complications. This is due to the large population of orbiting spacecraft and debris that continues to grow, passive (i.e. no direct communication) and data-sparse observations, and the presence of maneuvers and dynamics mismodeling. Accurate orbit determination in this environment requires an algorithm to capture both a system's state and its state dynamics in order to account for mismodelings. Previous studies by the authors yielded an algorithm called the Optimal Control Based Estimator (OCBE) - an algorithm that simultaneously estimates a system's state and optimal control policies that represent dynamic mismodeling in the system for an arbitrary orbit-observer setup. The stochastic properties of these estimated controls are then used to determine the presence of mismodelings (maneuver detection), as well as characterize and reconstruct the mismodelings. The purpose of this paper is to develop the OCBE into an accurate real-time orbit tracking and maneuver detection algorithm by automating the algorithm and removing its linear assumptions. This results in a nonlinear adaptive estimator. In its original form the OCBE had a parameter called the assumed dynamic uncertainty, which is selected by the user with each new measurement to reflect the level of dynamic mismodeling in the system. This human-in-the-loop approach precludes real-time application to orbit tracking problems due to their complexity. This paper focuses on the Adaptive OCBE, a version of the estimator where the assumed dynamic uncertainty is chosen automatically with each new measurement using maneuver detection results to ensure that state uncertainties are properly adjusted to account for all dynamic mismodelings. The paper also focuses on a nonlinear implementation of the estimator. Originally, the OCBE was derived from a nonlinear cost function then linearized about a nominal trajectory, which is assumed to be ballistic (i.e. the nominal optimal

  17. A novel case-control design to estimate the extent of over-diagnosis of breast cancer due to organised population-based mammography screening.

    Science.gov (United States)

    Beckmann, Kerri R; Lynch, John W; Hiller, Janet E; Farshid, Gelareh; Houssami, Nehmat; Duffy, Stephen W; Roder, David M

    2015-03-15

    Debate about the extent of breast cancer over-diagnosis due to mammography screening has continued for over a decade, without consensus. Estimates range from 0 to 54%, but many studies have been criticized for having flawed methodology. In this study we used a novel study design to estimate over-diagnosis due to organised mammography screening in South Australia (SA). To estimate breast cancer incidence at and following screening we used a population-based, age-matched case-control design involving 4,931 breast cancer cases and 22,914 controls to obtain OR for yearly time intervals since women's last screening mammogram. The level of over-diagnosis was estimated by comparing the cumulative breast cancer incidence with and without screening. The former was derived by applying ORs for each time window to incidence rates in the absence of screening, and the latter, by projecting pre-screening incidence rates. Sensitivity analyses were undertaken to assess potential biases. Over-diagnosis was estimated to be 8% (95%CI 2-14%) and 14% (95%CI 8-19%) among SA women aged 45 to 85 years from 2006-2010, for invasive breast cancer and all breast cancer respectively. These estimates were robust when applying various sensitivity analyses, except for adjustment for potential confounding assuming higher risk among screened than non-screened women, which reduced levels of over-diagnosis to 1% (95%CI 5-7%) and 8% (95%CI 2-14%) respectively when incidence rates for screening participants were adjusted by 10%. Our results indicate that the level of over-diagnosis due to mammography screening is modest and considerably lower than many previous estimates, including others for Australia. © 2014 UICC.

  18. Estimation of time-delayed mutual information and bias for irregularly and sparsely sampled time-series

    International Nuclear Information System (INIS)

    Albers, D.J.; Hripcsak, George

    2012-01-01

    Highlights: ► Time-delayed mutual information for irregularly sampled time-series. ► Estimation bias for the time-delayed mutual information calculation. ► Fast, simple, PDF estimator independent, time-delayed mutual information bias estimate. ► Quantification of data-set-size limits of the time-delayed mutual calculation. - Abstract: A method to estimate the time-dependent correlation via an empirical bias estimate of the time-delayed mutual information for a time-series is proposed. In particular, the bias of the time-delayed mutual information is shown to often be equivalent to the mutual information between two distributions of points from the same system separated by infinite time. Thus intuitively, estimation of the bias is reduced to estimation of the mutual information between distributions of data points separated by large time intervals. The proposed bias estimation techniques are shown to work for Lorenz equations data and glucose time series data of three patients from the Columbia University Medical Center database.

  19. Time Delay Estimation Algoritms for Echo Cancellation

    Directory of Open Access Journals (Sweden)

    Kirill Sakhnov

    2011-01-01

    Full Text Available The following case study describes how to eliminate echo in a VoIP network using delay estimation algorithms. It is known that echo with long transmission delays becomes more noticeable to users. Thus, time delay estimation, as a part of echo cancellation, is an important topic during transmission of voice signals over packetswitching telecommunication systems. An echo delay problem associated with IP-based transport networks is discussed in the following text. The paper introduces the comparative study of time delay estimation algorithm, used for estimation of the true time delay between two speech signals. Experimental results of MATLab simulations that describe the performance of several methods based on cross-correlation, normalized crosscorrelation and generalized cross-correlation are also presented in the paper.

  20. System and method for traffic signal timing estimation

    KAUST Repository

    Dumazert, Julien; Claudel, Christian G.

    2015-01-01

    A method and system for estimating traffic signals. The method and system can include constructing trajectories of probe vehicles from GPS data emitted by the probe vehicles, estimating traffic signal cycles, combining the estimates, and computing the traffic signal timing by maximizing a scoring function based on the estimates. Estimating traffic signal cycles can be based on transition times of the probe vehicles starting after a traffic signal turns green.

  1. System and method for traffic signal timing estimation

    KAUST Repository

    Dumazert, Julien

    2015-12-30

    A method and system for estimating traffic signals. The method and system can include constructing trajectories of probe vehicles from GPS data emitted by the probe vehicles, estimating traffic signal cycles, combining the estimates, and computing the traffic signal timing by maximizing a scoring function based on the estimates. Estimating traffic signal cycles can be based on transition times of the probe vehicles starting after a traffic signal turns green.

  2. estimating formwork striking time for concrete mixes estimating

    African Journals Online (AJOL)

    eobe

    In this study, we estimated the time for strength development in concrete cured up to 56 days. Water. In this .... regression analysis using MS Excel 2016 Software performed on the ..... [1] Abolfazl, K. R, Peroti S. and Rahemi L 'The Effect of.

  3. System-theoretic analysis of due-time performance in production systems

    Directory of Open Access Journals (Sweden)

    David Jacobs

    1995-01-01

    Full Text Available Along with the average production rate, the due-time performance is an important characteristic of manufacturing systems. Unlike the production rate, the due-time performance has received relatively little attention in the literature, especially in the context of large volume production. This paper is devoted to this topic. Specifically, the notion of due-time performance is formalized as the probability that the number of parts produced during the shipping period reaches the required shipment size. This performance index is analyzed for both lean and mass manufacturing environments. In particular, it is shown that, to achieve a high due-time performance in a lean environment, the production system should be scheduled for a sufficiently small fraction of its average production rate. In mass production, due-time performance arbitrarily close to one can be achieved for any scheduling practice, up to the average production rate.

  4. Real-time moving horizon estimation for a vibrating active cantilever

    Science.gov (United States)

    Abdollahpouri, Mohammad; Takács, Gergely; Rohaľ-Ilkiv, Boris

    2017-03-01

    Vibrating structures may be subject to changes throughout their operating lifetime due to a range of environmental and technical factors. These variations can be considered as parameter changes in the dynamic model of the structure, while their online estimates can be utilized in adaptive control strategies, or in structural health monitoring. This paper implements the moving horizon estimation (MHE) algorithm on a low-cost embedded computing device that is jointly observing the dynamic states and parameter variations of an active cantilever beam in real time. The practical behavior of this algorithm has been investigated in various experimental scenarios. It has been found, that for the given field of application, moving horizon estimation converges faster than the extended Kalman filter; moreover, it handles atypical measurement noise, sensor errors or other extreme changes, reliably. Despite its improved performance, the experiments demonstrate that the disadvantage of solving the nonlinear optimization problem in MHE is that it naturally leads to an increase in computational effort.

  5. Years of Life Gained Due to Leisure-Time Physical Activity in the United States

    Science.gov (United States)

    Janssen, Ian; Carson, Valerie; Lee, I-Min; Katzmarzyk, Peter T.; Blair, Steven N.

    2013-01-01

    Background Physical inactivity is an important modifiable risk factor for non-communicable disease. The degree to which physical activity affects the life expectancy of Americans is unknown. This study estimated the potential years of life gained due to leisure-time physical activity across the adult lifespan in the United States. Methods Data from the National Health and Nutrition Examination Survey (2007–2010), National Health Interview Study mortality linkage (1990–2006), and US Life Tables (2006) were used to estimate and compare life expectancy at each age of adult life for inactive (no moderate-to-vigorous physical activity), somewhat active (some moderate-to-vigorous activity but active (≥500 metabolic equivalent min/week of moderate-to-vigorous activity) adults. Analyses were conducted in 2012. Results Somewhat active and active non-Hispanic white men had a life expectancy at age 20 that was around 2.4 years longer than the inactive men; this life expectancy advantage was 1.2 years at age 80. Similar observations were made in non-Hispanic white women, with a higher life expectancy within the active category of 3.0 years at age 20 and 1.6 years at age 80. In non-Hispanic black women, as many as 5.5 potential years of life were gained due to physical activity. Significant increases in longevity were also observed within somewhat active and active non-Hispanic black men; however, among Hispanics the years of life gained estimates were more variable and not significantly different from 0 years gained. Conclusions Leisure-time physical activity is associated with increases in longevity in the United States. PMID:23253646

  6. Years of life gained due to leisure-time physical activity in the U.S.

    Science.gov (United States)

    Janssen, Ian; Carson, Valerie; Lee, I-Min; Katzmarzyk, Peter T; Blair, Steven N

    2013-01-01

    Physical inactivity is an important modifiable risk factor for noncommunicable disease. The degree to which physical activity affects the life expectancy of Americans is unknown. This study estimated the potential years of life gained due to leisure-time physical activity in the U.S. Data from the National Health and Nutrition Examination Survey (2007-2010); National Health Interview Study mortality linkage (1990-2006); and U.S. Life Tables (2006) were used to estimate and compare life expectancy at each age of adult life for inactive (no moderate to vigorous physical activity); somewhat-active (some moderate to vigorous activity but active (≥ 500 MET minutes/week of moderate to vigorous activity) adults. Analyses were conducted in 2012. Somewhat-active and active non-Hispanic white men had a life expectancy at age 20 years that was ~2.4 years longer than that for the inactive men; this life expectancy advantage was 1.2 years at age 80 years. Similar observations were made in non-Hispanic white women, with a higher life expectancy within the active category of 3.0 years at age 20 years and 1.6 years at age 80 years. In non-Hispanic black women, as many as 5.5 potential years of life were gained due to physical activity. Significant increases in longevity were also observed within somewhat-active and active non-Hispanic black men; however, among Hispanics the years-of-life-gained estimates were not significantly different from 0 years gained. Leisure-time physical activity is associated with increases in longevity. Copyright © 2013 American Journal of Preventive Medicine. Published by Elsevier Inc. All rights reserved.

  7. Simple approximation for estimating centerline gamma absorbed dose rates due to a continuous Gaussian plume

    International Nuclear Information System (INIS)

    Overcamp, T.J.; Fjeld, R.A.

    1987-01-01

    A simple approximation for estimating the centerline gamma absorbed dose rates due to a continuous Gaussian plume was developed. To simplify the integration of the dose integral, this approach makes use of the Gaussian cloud concentration distribution. The solution is expressed in terms of the I1 and I2 integrals which were developed for estimating long-term dose due to a sector-averaged Gaussian plume. Estimates of tissue absorbed dose rates for the new approach and for the uniform cloud model were compared to numerical integration of the dose integral over a Gaussian plume distribution

  8. Comparative study for the estimation of To shift due to irradiation embrittlement

    International Nuclear Information System (INIS)

    Lee, Jin Ho; Park, Youn won; Choi, Young Hwan; Kim, Seok Hun; Revka, Volodymyr

    2002-01-01

    Recently, an approach called the 'Master Curve' method was proposed which has opened a new means to acquire a directly measured material-specific fracture toughness curve. For the entire application of the Master Curve method, several technical issues should be solved. One of them is to utilize existing Charpy impact test data in the evaluation of a fracture transition temperature shift due to irradiation damage. In the U.S. and most Western countries, the Charpy impact test data have been used to estimate the irradiation effects on fracture toughness changes of RPV materials. For the determination of the irradiation shift the indexing energy level of 41 joule is used irrespective of the material yield strength. The Russian Code also requires the Charpy impact test data to determine the extent of radiation embrittlement. Unlike the U.S. Code, however, the Russian approach uses the indexing energy level varying according to the material strength. The objective of this study is to determine a method by which the reference transition temperature shift (ΔT o ) due to irradiation can be estimated. By comparing the irradiation shift estimated according to the U.S. procedure (ΔT 41J ) with that estimated according to the Russian procedure (ΔT F ), it was found that one-to-one relation exists between ΔT o and ΔT F

  9. Real-Time Head Pose Estimation on Mobile Platforms

    Directory of Open Access Journals (Sweden)

    Jianfeng Ren

    2010-06-01

    Full Text Available Many computer vision applications such as augmented reality require head pose estimation. As far as the real-time implementation of head pose estimation on relatively resource limited mobile platforms is concerned, it is required to satisfy real-time constraints while maintaining reasonable head pose estimation accuracy. The introduced head pose estimation approach in this paper is an attempt to meet this objective. The approach consists of the following components: Viola-Jones face detection, color-based face tracking using an online calibration procedure, and head pose estimation using Hu moment features and Fisher linear discriminant. Experimental results running on an actual mobile device are reported exhibiting both the real- time and accuracy aspects of the developed approach.

  10. Estimate of overdiagnosis of breast cancer due to mammography after adjustment for lead time. A service screening study in Italy

    Science.gov (United States)

    Paci, Eugenio; Miccinesi, Guido; Puliti, Donella; Baldazzi, Paola; De Lisi, Vincenzo; Falcini, Fabio; Cirilli, Claudia; Ferretti, Stefano; Mangone, Lucia; Finarelli, Alba Carola; Rosso, Stefano; Segnan, Nereo; Stracci, Fabrizio; Traina, Adele; Tumino, Rosario; Zorzi, Manuel

    2006-01-01

    Introduction Excess of incidence rates is the expected consequence of service screening. The aim of this paper is to estimate the quota attributable to overdiagnosis in the breast cancer screening programmes in Northern and Central Italy. Methods All patients with breast cancer diagnosed between 50 and 74 years who were resident in screening areas in the six years before and five years after the start of the screening programme were included. We calculated a corrected-for-lead-time number of observed cases for each calendar year. The number of observed incident cases was reduced by the number of screen-detected cases in that year and incremented by the estimated number of screen-detected cases that would have arisen clinically in that year. Results In total we included 13,519 and 13,999 breast cancer cases diagnosed in the pre-screening and screening years, respectively. In total, the excess ratio of observed to predicted in situ and invasive cases was 36.2%. After correction for lead time the excess ratio was 4.6% (95% confidence interval 2 to 7%) and for invasive cases only it was 3.2% (95% confidence interval 1 to 6%). Conclusion The remaining excess of cancers after individual correction for lead time was lower than 5%. PMID:17147789

  11. System-theoretic analysis of due-time performance in production systems

    OpenAIRE

    Jacobs David; Meerkov Semyon M.

    1995-01-01

    Along with the average production rate, the due-time performance is an important characteristic of manufacturing systems. Unlike the production rate, the due-time performance has received relatively little attention in the literature, especially in the context of large volume production. This paper is devoted to this topic. Specifically, the notion of due-time performance is formalized as the probability that the number of parts produced during the shipping period reaches the required shipme...

  12. Consequences of varied soil hydraulic and meteorological complexity on unsaturated zone time lag estimates.

    Science.gov (United States)

    Vero, S E; Ibrahim, T G; Creamer, R E; Grant, J; Healy, M G; Henry, T; Kramers, G; Richards, K G; Fenton, O

    2014-12-01

    The true efficacy of a programme of agricultural mitigation measures within a catchment to improve water quality can be determined only after a certain hydrologic time lag period (subsequent to implementation) has elapsed. As the biophysical response to policy is not synchronous, accurate estimates of total time lag (unsaturated and saturated) become critical to manage the expectations of policy makers. The estimation of the vertical unsaturated zone component of time lag is vital as it indicates early trends (initial breakthrough), bulk (centre of mass) and total (Exit) travel times. Typically, estimation of time lag through the unsaturated zone is poor, due to the lack of site specific soil physical data, or by assuming saturated conditions. Numerical models (e.g. Hydrus 1D) enable estimates of time lag with varied levels of input data. The current study examines the consequences of varied soil hydraulic and meteorological complexity on unsaturated zone time lag estimates using simulated and actual soil profiles. Results indicated that: greater temporal resolution (from daily to hourly) of meteorological data was more critical as the saturated hydraulic conductivity of the soil decreased; high clay content soils failed to converge reflecting prevalence of lateral component as a contaminant pathway; elucidation of soil hydraulic properties was influenced by the complexity of soil physical data employed (textural menu, ROSETTA, full and partial soil water characteristic curves), which consequently affected time lag ranges; as the importance of the unsaturated zone increases with respect to total travel times the requirements for high complexity/resolution input data become greater. The methodology presented herein demonstrates that decisions made regarding input data and landscape position will have consequences for the estimated range of vertical travel times. Insufficiencies or inaccuracies regarding such input data can therefore mislead policy makers regarding

  13. Highway travel time estimation with data fusion

    CERN Document Server

    Soriguera Martí, Francesc

    2016-01-01

    This monograph presents a simple, innovative approach for the measurement and short-term prediction of highway travel times based on the fusion of inductive loop detector and toll ticket data. The methodology is generic and not technologically captive, allowing it to be easily generalized for other equivalent types of data. The book shows how Bayesian analysis can be used to obtain fused estimates that are more reliable than the original inputs, overcoming some of the drawbacks of travel-time estimations based on unique data sources. The developed methodology adds value and obtains the maximum (in terms of travel time estimation) from the available data, without recurrent and costly requirements for additional data. The application of the algorithms to empirical testing in the AP-7 toll highway in Barcelona proves that it is possible to develop an accurate real-time, travel-time information system on closed-toll highways with the existing surveillance equipment, suggesting that highway operators might provide...

  14. Estimating the effect of a rare time-dependent treatment on the recurrent event rate.

    Science.gov (United States)

    Smith, Abigail R; Zhu, Danting; Goodrich, Nathan P; Merion, Robert M; Schaubel, Douglas E

    2018-05-30

    In many observational studies, the objective is to estimate the effect of treatment or state-change on the recurrent event rate. If treatment is assigned after the start of follow-up, traditional methods (eg, adjustment for baseline-only covariates or fully conditional adjustment for time-dependent covariates) may give biased results. We propose a two-stage modeling approach using the method of sequential stratification to accurately estimate the effect of a time-dependent treatment on the recurrent event rate. At the first stage, we estimate the pretreatment recurrent event trajectory using a proportional rates model censored at the time of treatment. Prognostic scores are estimated from the linear predictor of this model and used to match treated patients to as yet untreated controls based on prognostic score at the time of treatment for the index patient. The final model is stratified on matched sets and compares the posttreatment recurrent event rate to the recurrent event rate of the matched controls. We demonstrate through simulation that bias due to dependent censoring is negligible, provided the treatment frequency is low, and we investigate a threshold at which correction for dependent censoring is needed. The method is applied to liver transplant (LT), where we estimate the effect of development of post-LT End Stage Renal Disease (ESRD) on rate of days hospitalized. Copyright © 2018 John Wiley & Sons, Ltd.

  15. Travel time estimation using Bluetooth.

    Science.gov (United States)

    2015-06-01

    The objective of this study was to investigate the feasibility of using a Bluetooth Probe Detection System (BPDS) to : estimate travel time in an urban area. Specifically, the study investigated the possibility of measuring overall congestion, the : ...

  16. Estimation of absorbed dose in cell nuclei due to DNA-bound /sup 3/H

    Energy Technology Data Exchange (ETDEWEB)

    Saito, M; Ishida, M R; Streffer, C; Molls, M

    1985-04-01

    The average absorbed dose due to DNA-bound /sup 3/H in a cell nucleus was estimated by a Monte Carlo simulation for a model nucleus which was assumed to be spheroidal. The volume of the cell nucleus was the major dose-determining factor for cell nuclei which have the same DNA content and the same specific activity of DNA. This result was applied to estimating the accumulated dose in the cell nuclei of organs of young mice born from mother mice which ingested /sup 3/H-thymidine with drinking water during pregnancy. The values of dose-modifying factors for the accumulated dose due to DNA-bound /sup 3/H compared to the dose due to an assumed homogenous distribution of /sup 3/H in organ were found to be between about 2 and 6 for the various organs.

  17. Fading Kalman filter-based real-time state of charge estimation in LiFePO_4 battery-powered electric vehicles

    International Nuclear Information System (INIS)

    Lim, KaiChin; Bastawrous, Hany Ayad; Duong, Van-Huan; See, Khay Wai; Zhang, Peng; Dou, Shi Xue

    2016-01-01

    Highlights: • Real-time battery model parameters and SoC estimation with novel method is proposed. • Cascading filtering stages are used for parameters identification and SoC estimation. • Optimized fading Kalman filter is implemented for SoC estimation. • Accurate SoC estimation is validated in UDDS load profile experiment. • This approach is suitable for BMS in EV applications due to its simplicity. - Abstract: A novel online estimation technique for estimating the state of charge (SoC) of a lithium iron phosphate (LiFePO_4) battery has been developed. Based on a simplified model, the open circuit voltage (OCV) of the battery is estimated through two cascaded linear filtering stages. A recursive least squares filter is employed in the first stage to dynamically estimate the battery model parameters in real-time, and then, a fading Kalman filter (FKF) is used to estimate the OCV from these parameters. FKF can avoid the possibility of large estimation errors, which may occur with a conventional Kalman filter, due to its capability to compensate any modeling error through a fading factor. By optimizing the value of the fading factor in the set of recursion equations of FKF with genetic algorithms, the errors in estimating the battery’s SoC in urban dynamometer driving schedules-based experiments and real vehicle driving cycle experiments were below 3% compared to more than 9% in the case of using an ordinary Kalman filter. The proposed method with its simplified model provides the simplicity and feasibility required for real-time application with highly accurate SoC estimation.

  18. Estimating the operator's performance time of emergency procedural tasks based on a task complexity measure

    International Nuclear Information System (INIS)

    Jung, Won Dae; Park, Jink Yun

    2012-01-01

    It is important to understand the amount of time required to execute an emergency procedural task in a high-stress situation for managing human performance under emergencies in a nuclear power plant. However, the time to execute an emergency procedural task is highly dependent upon expert judgment due to the lack of actual data. This paper proposes an analytical method to estimate the operator's performance time (OPT) of a procedural task, which is based on a measure of the task complexity (TACOM). The proposed method for estimating an OPT is an equation that uses the TACOM as a variable, and the OPT of a procedural task can be calculated if its relevant TACOM score is available. The validity of the proposed equation is demonstrated by comparing the estimated OPTs with the observed OPTs for emergency procedural tasks in a steam generator tube rupture scenario.

  19. Estimation of economic losses due to Peste de Petits Ruminants in small ruminants in India

    Directory of Open Access Journals (Sweden)

    B. Singh

    2014-04-01

    Full Text Available Aim: To develop a simple mathematical model to assess the losses due to peste des petits ruminants (PPR in small ruminants in India. Materials and Methods: The study was based on cases and deaths in goats and sheep due to PPR from the average combined data on ovine/caprine as published by Government of India for the last 5 years (2008-2012. All possible direct and indirect losses due to the disease, viz. mortality losses, losses due to direct reduction in milk/wool yield, losses due to reproduction failure, body weight losses, treatment costs and opportunity costs, were considered to provide estimate of annual economic losses due to PPR in sheep and goats in India. Based on cases and deaths as reported in sample survey studies, the annual economic loss was also estimated. Results: On the basis of data reported by Government of India, the study has shown average annual economic loss of Rs. 167.83 lacs, of which Rs. 125.67 lacs and Rs. 42.16 lacs respectively are due to the incidence of the disease in goats and sheep. Morbidity losses constituted the greater share of the total loss in both goats and sheep (56.99% and 61.34%, respectively. Among different components of morbidity loss, direct body weight loss was the most significant in both goats and sheep. Based on cases and deaths as reported in sample survey studies, the estimated annual economic loss due to PPR in goats and sheep is Rs. 8895.12 crores, of which Rs. 5477.48 and Rs. 3417.64 crores respectively are due to the disease in goats and sheep. Conclusion: The low economic losses as reported based on Government of India data points towards underreporting of cases and deaths due to the disease. The study thus revealed a significant loss due to PPR in small ruminants on a large scale.

  20. Nonparametric volatility density estimation for discrete time models

    NARCIS (Netherlands)

    Es, van Bert; Spreij, P.J.C.; Zanten, van J.H.

    2005-01-01

    We consider discrete time models for asset prices with a stationary volatility process. We aim at estimating the multivariate density of this process at a set of consecutive time instants. A Fourier-type deconvolution kernel density estimator based on the logarithm of the squared process is proposed

  1. Space-Time Smoothing of Complex Survey Data: Small Area Estimation for Child Mortality.

    Science.gov (United States)

    Mercer, Laina D; Wakefield, Jon; Pantazis, Athena; Lutambi, Angelina M; Masanja, Honorati; Clark, Samuel

    2015-12-01

    Many people living in low and middle-income countries are not covered by civil registration and vital statistics systems. Consequently, a wide variety of other types of data including many household sample surveys are used to estimate health and population indicators. In this paper we combine data from sample surveys and demographic surveillance systems to produce small area estimates of child mortality through time. Small area estimates are necessary to understand geographical heterogeneity in health indicators when full-coverage vital statistics are not available. For this endeavor spatio-temporal smoothing is beneficial to alleviate problems of data sparsity. The use of conventional hierarchical models requires careful thought since the survey weights may need to be considered to alleviate bias due to non-random sampling and non-response. The application that motivated this work is estimation of child mortality rates in five-year time intervals in regions of Tanzania. Data come from Demographic and Health Surveys conducted over the period 1991-2010 and two demographic surveillance system sites. We derive a variance estimator of under five years child mortality that accounts for the complex survey weighting. For our application, the hierarchical models we consider include random effects for area, time and survey and we compare models using a variety of measures including the conditional predictive ordinate (CPO). The method we propose is implemented via the fast and accurate integrated nested Laplace approximation (INLA).

  2. DEMO maintenance scenarios: scheme for time estimations and preliminary estimates for blankets arranged in multi-module-segments

    International Nuclear Information System (INIS)

    Nagy, D.

    2007-01-01

    Previous conceptual studies made clear that the ITER blanket concept and segmentation is not suitable for the environment of a potential fusion power plant (DEMO). One promising concept to be used instead is the so-called Multi-Module-Segment (MMS) concept. Each MMS consists of a number of blankets arranged on a strong back plate thus forming ''banana'' shaped in-board (IB) and out-board (OB) segments. With respect to port size, weight, or other limiting aspects the IB and OB MMS are segmented in toroidal direction. The number of segments to be replaced would be below 100. For this segmentation concept a new maintenance scenario had to be worked out. The aim of this paper is to present a promising MMS maintenance scenario, a flexible scheme for time estimations under varying boundary conditions and preliminary time estimates. According to the proposed scenario two upper, vertical arranged maintenance ports have to be opened for blanket maintenance on opposite sides of the tokamak. Both ports are central to a 180 degree sector and the MMS are removed and inserted through both ports. In-vessel machines are operating to transport the elements in toroidal direction and also to insert and attach the MMS to the shield. Outside the vessel the elements have to be transported between the tokamak and the hot cell to be refurbished. Calculating the maintenance time for such a scenario is rather challenging due to the numerous parallel processes involved. For this reason a flexible, multi-level calculation scheme has been developed in which the operations are organized into three levels: At the lowest level the basic maintenance steps are determined. These are organized into maintenance sequences that take into account parallelisms in the system. Several maintenance sequences constitute the maintenance phases which correspond to a certain logistics scenario. By adding the required times of the maintenance phases the total maintenance time is obtained. The paper presents

  3. Anthropogenic CO2 in the oceans estimated using transit time distributions

    International Nuclear Information System (INIS)

    Waugh, D.W.; McNeil, B.I.

    2006-01-01

    The distribution of anthropogenic carbon (Cant) in the oceans is estimated using the transit time distribution (TTD) method applied to global measurements of chlorofluorocarbon-12 (CFC12). Unlike most other inference methods, the TTD method does not assume a single ventilation time and avoids the large uncertainty incurred by attempts to correct for the large natural carbon background in dissolved inorganic carbon measurements. The highest concentrations and deepest penetration of anthropogenic carbon are found in the North Atlantic and Southern Oceans. The estimated total inventory in 1994 is 134 Pg-C. To evaluate uncertainties the TTD method is applied to output from an ocean general circulation model (OGCM) and compared the results to the directly simulated Cant. Outside of the Southern Ocean the predicted Cant closely matches the directly simulated distribution, but in the Southern Ocean the TTD concentrations are biased high due to the assumption of 'constant disequilibrium'. The net result is a TTD overestimate of the global inventory by about 20%. Accounting for this bias and other centred uncertainties, an inventory range of 94-121 Pg-C is obtained. This agrees with the inventory of Sabine et al., who applied the DeltaC* method to the same data. There are, however, significant differences in the spatial distributions: The TTD estimates are smaller than DeltaC* in the upper ocean and larger at depth, consistent with biases expected in DeltaC* given its assumption of a single parcel ventilation time

  4. Automated Land Cover Change Detection and Mapping from Hidden Parameter Estimates of Normalized Difference Vegetation Index (NDVI) Time-Series

    Science.gov (United States)

    Chakraborty, S.; Banerjee, A.; Gupta, S. K. S.; Christensen, P. R.; Papandreou-Suppappola, A.

    2017-12-01

    Multitemporal observations acquired frequently by satellites with short revisit periods such as the Moderate Resolution Imaging Spectroradiometer (MODIS), is an important source for modeling land cover. Due to the inherent seasonality of the land cover, harmonic modeling reveals hidden state parameters characteristic to it, which is used in classifying different land cover types and in detecting changes due to natural or anthropogenic factors. In this work, we use an eight day MODIS composite to create a Normalized Difference Vegetation Index (NDVI) time-series of ten years. Improved hidden parameter estimates of the nonlinear harmonic NDVI model are obtained using the Particle Filter (PF), a sequential Monte Carlo estimator. The nonlinear estimation based on PF is shown to improve parameter estimation for different land cover types compared to existing techniques that use the Extended Kalman Filter (EKF), due to linearization of the harmonic model. As these parameters are representative of a given land cover, its applicability in near real-time detection of land cover change is also studied by formulating a metric that captures parameter deviation due to change. The detection methodology is evaluated by considering change as a rare class problem. This approach is shown to detect change with minimum delay. Additionally, the degree of change within the change perimeter is non-uniform. By clustering the deviation in parameters due to change, this spatial variation in change severity is effectively mapped and validated with high spatial resolution change maps of the given regions.

  5. Effect of parameter calculation in direct estimation of the Lyapunov exponent in short time series

    Directory of Open Access Journals (Sweden)

    A. M. López Jiménez

    2002-01-01

    Full Text Available The literature about non-linear dynamics offers a few recommendations, which sometimes are divergent, about the criteria to be used in order to select the optimal calculus parameters in the estimation of Lyapunov exponents by direct methods. These few recommendations are circumscribed to the analysis of chaotic systems. We have found no recommendation for the estimation of λ starting from the time series of classic systems. The reason for this is the interest in distinguishing variability due to a chaotic behavior of determinist dynamic systems of variability caused by white noise or linear stochastic processes, and less in the identification of non-linear terms from the analysis of time series. In this study we have centered in the dependence of the Lyapunov exponent, obtained by means of direct estimation, of the initial distance and the time evolution. We have used generated series of chaotic systems and generated series of classic systems with varying complexity. To generate the series we have used the logistic map.

  6. Tsunami Amplitude Estimation from Real-Time GNSS.

    Science.gov (United States)

    Jeffries, C.; MacInnes, B. T.; Melbourne, T. I.

    2017-12-01

    Tsunami early warning systems currently comprise modeling of observations from the global seismic network, deep-ocean DART buoys, and a global distribution of tide gauges. While these tools work well for tsunamis traveling teleseismic distances, saturation of seismic magnitude estimation in the near field can result in significant underestimation of tsunami excitation for local warning. Moreover, DART buoy and tide gauge observations cannot be used to rectify the underestimation in the available time, typically 10-20 minutes, before local runup occurs. Real-time GNSS measurements of coseismic offsets may be used to estimate finite faulting within 1-2 minutes and, in turn, tsunami excitation for local warning purposes. We describe here a tsunami amplitude estimation algorithm; implemented for the Cascadia subduction zone, that uses continuous GNSS position streams to estimate finite faulting. The system is based on a time-domain convolution of fault slip that uses a pre-computed catalog of hydrodynamic Green's functions generated with the GeoClaw shallow-water wave simulation software and maps seismic slip along each section of the fault to points located off the Cascadia coast in 20m of water depth and relies on the principle of the linearity in tsunami wave propagation. The system draws continuous slip estimates from a message broker, convolves the slip with appropriate Green's functions which are then superimposed to produce wave amplitude at each coastal location. The maximum amplitude and its arrival time are then passed into a database for subsequent monitoring and display. We plan on testing this system using a suite of synthetic earthquakes calculated for Cascadia whose ground motions are simulated at 500 existing Cascadia GPS sites, as well as real earthquakes for which we have continuous GNSS time series and surveyed runup heights, including Maule, Chile 2010 and Tohoku, Japan 2011. This system has been implemented in the CWU Geodesy Lab for the Cascadia

  7. Bias Errors due to Leakage Effects When Estimating Frequency Response Functions

    Directory of Open Access Journals (Sweden)

    Andreas Josefsson

    2012-01-01

    Full Text Available Frequency response functions are often utilized to characterize a system's dynamic response. For a wide range of engineering applications, it is desirable to determine frequency response functions for a system under stochastic excitation. In practice, the measurement data is contaminated by noise and some form of averaging is needed in order to obtain a consistent estimator. With Welch's method, the discrete Fourier transform is used and the data is segmented into smaller blocks so that averaging can be performed when estimating the spectrum. However, this segmentation introduces leakage effects. As a result, the estimated frequency response function suffers from both systematic (bias and random errors due to leakage. In this paper the bias error in the H1 and H2-estimate is studied and a new method is proposed to derive an approximate expression for the relative bias error at the resonance frequency with different window functions. The method is based on using a sum of real exponentials to describe the window's deterministic autocorrelation function. Simple expressions are derived for a rectangular window and a Hanning window. The theoretical expressions are verified with numerical simulations and a very good agreement is found between the results from the proposed bias expressions and the empirical results.

  8. Smooth time-dependent receiver operating characteristic curve estimators.

    Science.gov (United States)

    Martínez-Camblor, Pablo; Pardo-Fernández, Juan Carlos

    2018-03-01

    The receiver operating characteristic curve is a popular graphical method often used to study the diagnostic capacity of continuous (bio)markers. When the considered outcome is a time-dependent variable, two main extensions have been proposed: the cumulative/dynamic receiver operating characteristic curve and the incident/dynamic receiver operating characteristic curve. In both cases, the main problem for developing appropriate estimators is the estimation of the joint distribution of the variables time-to-event and marker. As usual, different approximations lead to different estimators. In this article, the authors explore the use of a bivariate kernel density estimator which accounts for censored observations in the sample and produces smooth estimators of the time-dependent receiver operating characteristic curves. The performance of the resulting cumulative/dynamic and incident/dynamic receiver operating characteristic curves is studied by means of Monte Carlo simulations. Additionally, the influence of the choice of the required smoothing parameters is explored. Finally, two real-applications are considered. An R package is also provided as a complement to this article.

  9. Radiation dose estimation from foods due to the accident of TEPCO Fukushima Daiichi Nuclear Power Station

    International Nuclear Information System (INIS)

    Yamaguchi, Ichiro

    2012-01-01

    Explained are the purpose of dose assessment, its methods, actual radionuclide levels in food, amounts of food intake, dose estimated hitherto, dose in the future, dose estimated by total food studies, and problems of assessing the dose from food, all of which Tokyo Electric Power Company (TEPCO) Power Station Accident has raised. Dose derived from food can be estimated by the radioactivity measured in each food material and in its combined amounts or in actually cooked food. Amounts of radioactive materials ingested in the body can be measured externally or by bioassay. Japan MHLW published levels of radioactivity in vegetables', fruits, marine products and meats from Mar. 2011, of which time course pattern has been found different each other within and between month(s). Dose due to early exposure in the Accident can be estimated by the radioactivity levels above and data concerning the amounts of food intake summarized by National Institute of Health and Nutrition in 2010 and other institutions. For instance, the thyroid tissue equivalent dose by I-131 in a 1 year old child is estimated to be 1.1-5 mSv depending on the assumed data for calculation, in the first month after the Accident when ICRP tissue equivalent dose coefficient 3.7 x 10-6 Sv/Bq is used. In the future (later than Apr. 2012), new standard limits of radiocesium levels in milk/its products and foods for infant and in other general foods are to be defined 50 and 100 Bq/kg, respectively. The distribution of committed effective doses by radiocesium (mSv/y food intake) are presented as an instance, where it is estimated by 1 million stochastic simulations using 2 covariates of Cs-134, -137 levels (as representative nuclides under regulation) in food and of daily food intake. In dose prediction, conjecturing the behavior of environmental radionuclides and the time of resume of primary industries would be necessary. (T.T.)

  10. Consequences of Secondary Calibrations on Divergence Time Estimates.

    Directory of Open Access Journals (Sweden)

    John J Schenk

    Full Text Available Secondary calibrations (calibrations based on the results of previous molecular dating studies are commonly applied in divergence time analyses in groups that lack fossil data; however, the consequences of applying secondary calibrations in a relaxed-clock approach are not fully understood. I tested whether applying the posterior estimate from a primary study as a prior distribution in a secondary study results in consistent age and uncertainty estimates. I compared age estimates from simulations with 100 randomly replicated secondary trees. On average, the 95% credible intervals of node ages for secondary estimates were significantly younger and narrower than primary estimates. The primary and secondary age estimates were significantly different in 97% of the replicates after Bonferroni corrections. Greater error in magnitude was associated with deeper than shallower nodes, but the opposite was found when standardized by median node age, and a significant positive relationship was determined between the number of tips/age of secondary trees and the total amount of error. When two secondary calibrated nodes were analyzed, estimates remained significantly different, and although the minimum and median estimates were associated with less error, maximum age estimates and credible interval widths had greater error. The shape of the prior also influenced error, in which applying a normal, rather than uniform, prior distribution resulted in greater error. Secondary calibrations, in summary, lead to a false impression of precision and the distribution of age estimates shift away from those that would be inferred by the primary analysis. These results suggest that secondary calibrations should not be applied as the only source of calibration in divergence time analyses that test time-dependent hypotheses until the additional error associated with secondary calibrations is more properly modeled to take into account increased uncertainty in age estimates.

  11. Global Population Density Grid Time Series Estimates

    Data.gov (United States)

    National Aeronautics and Space Administration — Global Population Density Grid Time Series Estimates provide a back-cast time series of population density grids based on the year 2000 population grid from SEDAC's...

  12. Estimation of time-varying growth, uptake and excretion rates from dynamic metabolomics data.

    Science.gov (United States)

    Cinquemani, Eugenio; Laroute, Valérie; Cocaign-Bousquet, Muriel; de Jong, Hidde; Ropers, Delphine

    2017-07-15

    Technological advances in metabolomics have made it possible to monitor the concentration of extracellular metabolites over time. From these data, it is possible to compute the rates of uptake and excretion of the metabolites by a growing cell population, providing precious information on the functioning of intracellular metabolism. The computation of the rate of these exchange reactions, however, is difficult to achieve in practice for a number of reasons, notably noisy measurements, correlations between the concentration profiles of the different extracellular metabolites, and discontinuties in the profiles due to sudden changes in metabolic regime. We present a method for precisely estimating time-varying uptake and excretion rates from time-series measurements of extracellular metabolite concentrations, specifically addressing all of the above issues. The estimation problem is formulated in a regularized Bayesian framework and solved by a combination of extended Kalman filtering and smoothing. The method is shown to improve upon methods based on spline smoothing of the data. Moreover, when applied to two actual datasets, the method recovers known features of overflow metabolism in Escherichia coli and Lactococcus lactis , and provides evidence for acetate uptake by L. lactis after glucose exhaustion. The results raise interesting perspectives for further work on rate estimation from measurements of intracellular metabolites. The Matlab code for the estimation method is available for download at https://team.inria.fr/ibis/rate-estimation-software/ , together with the datasets. eugenio.cinquemani@inria.fr. Supplementary data are available at Bioinformatics online. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com

  13. Induced voltage due to time-dependent magnetisation textures

    International Nuclear Information System (INIS)

    Kudtarkar, Santosh Kumar; Dhadwal, Renu

    2010-01-01

    We determine the induced voltage generated by spatial and temporal magnetisation textures (inhomogeneities) in metallic ferromagnets due to the spin diffusion of non-equilibrium electrons. Using time dependent semi-classical theory as formulated in Zhang and Li and the drift-diffusion model of transport it is shown that the voltage generated depends critically on the difference in the diffusion constants of up and down spins. Including spin relaxation results in a crucial contribution to the induced voltage. We also show that the presence of magnetisation textures results in the modification of the conductivity of the system. As an illustration, we calculate the voltage generated due to a time dependent field driven helimagnet by solving the Landau-Lifshitz equation with Gilbert damping and explicitly calculate the dependence on the relaxation and damping parameters.

  14. Investigation of metabolites for estimating blood deposition time.

    Science.gov (United States)

    Lech, Karolina; Liu, Fan; Davies, Sarah K; Ackermann, Katrin; Ang, Joo Ern; Middleton, Benita; Revell, Victoria L; Raynaud, Florence J; Hoveijn, Igor; Hut, Roelof A; Skene, Debra J; Kayser, Manfred

    2018-01-01

    Trace deposition timing reflects a novel concept in forensic molecular biology involving the use of rhythmic biomarkers for estimating the time within a 24-h day/night cycle a human biological sample was left at the crime scene, which in principle allows verifying a sample donor's alibi. Previously, we introduced two circadian hormones for trace deposition timing and recently demonstrated that messenger RNA (mRNA) biomarkers significantly improve time prediction accuracy. Here, we investigate the suitability of metabolites measured using a targeted metabolomics approach, for trace deposition timing. Analysis of 171 plasma metabolites collected around the clock at 2-h intervals for 36 h from 12 male participants under controlled laboratory conditions identified 56 metabolites showing statistically significant oscillations, with peak times falling into three day/night time categories: morning/noon, afternoon/evening and night/early morning. Time prediction modelling identified 10 independently contributing metabolite biomarkers, which together achieved prediction accuracies expressed as AUC of 0.81, 0.86 and 0.90 for these three time categories respectively. Combining metabolites with previously established hormone and mRNA biomarkers in time prediction modelling resulted in an improved prediction accuracy reaching AUCs of 0.85, 0.89 and 0.96 respectively. The additional impact of metabolite biomarkers, however, was rather minor as the previously established model with melatonin, cortisol and three mRNA biomarkers achieved AUC values of 0.88, 0.88 and 0.95 for the same three time categories respectively. Nevertheless, the selected metabolites could become practically useful in scenarios where RNA marker information is unavailable such as due to RNA degradation. This is the first metabolomics study investigating circulating metabolites for trace deposition timing, and more work is needed to fully establish their usefulness for this forensic purpose.

  15. Estimation of organ and effective dose due to Compton backscatter security scans

    International Nuclear Information System (INIS)

    Hoppe, Michael E.; Schmidt, Taly Gilat

    2012-01-01

    Purpose: To estimate organ and effective radiation doses due to backscatter security scanners using Monte Carlo simulations and a voxelized phantom set. Methods: Voxelized phantoms of male and female adults and children were used with the GEANT4 toolkit to simulate a backscatter security scan. The backscatter system was modeled based on specifications available in the literature. The simulations modeled a 50 kVp spectrum with 1.0 mm-aluminum-equivalent filtration and a previously measured exposure of approximately 4.6 μR at 30 cm from the source. Photons and secondary interactions were tracked from the source until they reached zero kinetic energy or exited from the simulation’s boundaries. The energy deposited in the phantoms’ respective organs was tallied and used to calculate total organ dose and total effective dose for frontal, rear, and full scans with subjects located 30 and 75 cm from the source. Results: For a full screen, all phantoms’ total effective doses were below the established 0.25 μSv standard, with an estimated maximum total effective dose of 0.07 μSv for full screen of a male child. The estimated maximum organ dose due to a full screen was 1.03 μGy, deposited in the adipose tissue of the male child phantom when located 30 cm from the source. All organ dose estimates had a coefficient of variation of less than 3% for a frontal scan and less than 11% for a rear scan. Conclusions: Backscatter security scanners deposit dose in organs beyond the skin. The effective dose is below recommended standards set by the Health Physics Society (HPS) and the American National Standards Institute (ANSI) assuming the system provides a maximum exposure of approximately 4.6 μR at 30 cm.

  16. Template-Based Estimation of Time-Varying Tempo

    Directory of Open Access Journals (Sweden)

    Peeters Geoffroy

    2007-01-01

    Full Text Available We present a novel approach to automatic estimation of tempo over time. This method aims at detecting tempo at the tactus level for percussive and nonpercussive audio. The front-end of our system is based on a proposed reassigned spectral energy flux for the detection of musical events. The dominant periodicities of this flux are estimated by a proposed combination of discrete Fourier transform and frequency-mapped autocorrelation function. The most likely meter, beat, and tatum over time are then estimated jointly using proposed meter/beat subdivision templates and a Viterbi decoding algorithm. The performances of our system have been evaluated on four different test sets among which three were used during the ISMIR 2004 tempo induction contest. The performances obtained are close to the best results of this contest.

  17. Estimation of health damage due to emission of air pollutants by cars: the canyon effect

    Energy Technology Data Exchange (ETDEWEB)

    Spadaro, J.V. [Ecole des Mines, Centre d' Energetique, Paris, 75 (France); Rabl, A.

    1999-07-01

    Since current epidemiological evidence suggests that air pollution has harmful effects even at typical ambient concentrations and the dispersion is significant over hundreds to thousands of km, the estimation of total health damage involves consideration of local and regional effects. In recent years, several estimates have been published of health damage due to air pollution from cars, in particular by Delucchi et al of UC Davis and by the ExternE Project of the European Commission. To capture the geographic extent of pollutant dispersion, local and regional models have been used in combination. The present paper addresses a potentially significant contribution of the total damage, not yet taken into account in these studies: the increased concentration of pollutants inside urban street canyons. This canyon effect is appreciable only for primary pollutants, the time constants for the formation of secondary pollutants being long compared to the residence time in the canyon. We assumed linearity of incremental health impact with incremental concentration, in view of the lack of epidemiological evidence for no-effect thresholds or significant deviations from linearity at typical ambient concentrations; therefore, only long term average concentrations matter. We use the FLUENT software to model the dispersion inside a street canyon for a wide range of rectangular geometries and wind velocities. Our results suggest that the canyon effect is of marginal significance for total damages, the contribution of the canyon effect being roughly 10 to 20% of the total. The relative importance of the canyon effect is, of course, highly variable with local conditions; it could be much smaller but it is unlikely to add more than 100% to the flat terrain estimate. (Author)

  18. Dynamic travel time estimation using regression trees.

    Science.gov (United States)

    2008-10-01

    This report presents a methodology for travel time estimation by using regression trees. The dissemination of travel time information has become crucial for effective traffic management, especially under congested road conditions. In the absence of c...

  19. Time Estimation Deficits in Childhood Mathematics Difficulties

    Science.gov (United States)

    Hurks, Petra P. M.; van Loosbroek, Erik

    2014-01-01

    Time perception has not been comprehensively examined in mathematics difficulties (MD). Therefore, verbal time estimation, production, and reproduction were tested in 13 individuals with MD and 16 healthy controls, matched for age, sex, and intellectual skills. Individuals with MD performed comparably to controls in time reproduction, but showed a…

  20. Mode choice endogeneity in value of travel time estimation

    DEFF Research Database (Denmark)

    Mabit, Stefan Lindhard; Fosgerau, Mogens

    The current way to estimate value of travel time is to use a mode-specific sample and hence to estimate mode-specific value of travel times. This approach raises certain questions concerning how to generalise the values to a population. A problem would be if there is an uncontrolled sample...... selection mechanism. This is the case if there is correlation between mode choice and the value of travel time that is not controlled for by explanatory variables. What could confuse the estimated values is the difficulty to separate mode effects from user effect. An example would be the effect of income...... of travel time we use a stated choice dataset. These data include binary choice within mode for car and bus. The first approach is to use a probit model to model mode choice using instruments and then use this in the estimation of the value of travel time. The second approach is based on the use of a very...

  1. So many genes, so little time: A practical approach to divergence-time estimation in the genomic era.

    Science.gov (United States)

    Smith, Stephen A; Brown, Joseph W; Walker, Joseph F

    2018-01-01

    lognormal (UCLN) models. However, this overlap was often due to imprecise estimates from the UCLN model. We find that "gene shopping" can be an efficient approach to divergence-time inference for phylogenomic datasets that may otherwise be characterized by extensive gene tree heterogeneity.

  2. Robust Fault Estimation Design for Discrete-Time Nonlinear Systems via A Modified Fuzzy Fault Estimation Observer.

    Science.gov (United States)

    Xie, Xiang-Peng; Yue, Dong; Park, Ju H

    2018-02-01

    The paper provides relaxed designs of fault estimation observer for nonlinear dynamical plants in the Takagi-Sugeno form. Compared with previous theoretical achievements, a modified version of fuzzy fault estimation observer is implemented with the aid of the so-called maximum-priority-based switching law. Given each activated switching status, the appropriate group of designed matrices can be provided so as to explore certain key properties of the considered plants by means of introducing a set of matrix-valued variables. Owing to the reason that more abundant information of the considered plants can be updated in due course and effectively exploited for each time instant, the conservatism of the obtained result is less than previous theoretical achievements and thus the main defect of those existing methods can be overcome to some extent in practice. Finally, comparative simulation studies on the classical nonlinear truck-trailer model are given to certify the benefits of the theoretic achievement which is obtained in our study. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.

  3. Reliability of Bluetooth Technology for Travel Time Estimation

    DEFF Research Database (Denmark)

    Araghi, Bahar Namaki; Olesen, Jonas Hammershøj; Krishnan, Rajesh

    2015-01-01

    . However, their corresponding impacts on accuracy and reliability of estimated travel time have not been evaluated. In this study, a controlled field experiment is conducted to collect both Bluetooth and GPS data for 1000 trips to be used as the basis for evaluation. Data obtained by GPS logger is used...... to calculate actual travel time, referred to as ground truth, and to geo-code the Bluetooth detection events. In this setting, reliability is defined as the percentage of devices captured per trip during the experiment. It is found that, on average, Bluetooth-enabled devices will be detected 80% of the time......-range antennae detect Bluetooth-enabled devices in a closer location to the sensor, thus providing a more accurate travel time estimate. However, the smaller the size of the detection zone, the lower the penetration rate, which could itself influence the accuracy of estimates. Therefore, there has to be a trade...

  4. Health-related effects of early part-time sick leave due to musculoskeletal disorders: a randomized controlled trial.

    Science.gov (United States)

    Shiri, Rahman; Kausto, Johanna; Martimo, Kari-Pekka; Kaila-Kangas, Leena; Takala, Esa-Pekka; Viikari-Juntura, Eira

    2013-01-01

    Previously we reported that early part-time sick leave enhances return to work (RTW) among employees with musculoskeletal disorders (MSD). This paper assesses the health-related effects of this intervention. Patients aged 18-60 years who were unable to perform their regular work due to MSD were randomized to part- or full-time sick leave groups. In the former, workload was reduced by halving working time. Using validated questionnaires, we assessed pain intensity and interference with work and sleep, region-specific disability due to MSD, self-rated general health, health-related quality of life (measured via EuroQol), productivity loss, depression, and sleep disturbance at baseline, 1, 3, 8, 12, and 52 weeks. We analyzed the repeated measures data (171-356 observations) with the generalized estimating equation approach. The intervention (part-time sick leave) and control (full-time sick leave) groups did not differ with regard to pain intensity, pain interference with work and sleep, region-specific disability, productivity loss, depression, or sleep disturbance. The intervention group reported better self-rated general health (adjusted P=0.07) and health-related quality of life (adjusted P=0.02) than the control group. In subgroup analyses, the intervention was more effective among the patients whose current problem began occurring part-time sick leave did not exacerbate pain-related symptoms and functional disability, but improved self-rated general health and health-related quality of life in the early stage of work disability due to MSD.

  5. A method for the estimation of the probability of damage due to earthquakes

    International Nuclear Information System (INIS)

    Alderson, M.A.H.G.

    1979-07-01

    The available information on seismicity within the United Kingdom has been combined with building damage data from the United States to produce a method of estimating the probability of damage to structures due to the occurrence of earthquakes. The analysis has been based on the use of site intensity as the major damage producing parameter. Data for structural, pipework and equipment items have been assumed and the overall probability of damage calculated as a function of the design level. Due account is taken of the uncertainties of the seismic data. (author)

  6. Time allocation shifts and pollutant exposure due to traffic congestion: an analysis using the national human activity pattern survey.

    Science.gov (United States)

    Zhang, Kai; Batterman, Stuart A

    2009-10-15

    Traffic congestion increases air pollutant exposures of commuters and urban populations due to the increased time spent in traffic and the increased vehicular emissions that occur in congestion, especially "stop-and-go" traffic. Increased time in traffic also decreases time in other microenvironments, a trade-off that has not been considered in previous time activity pattern (TAP) analyses conducted for exposure assessment purposes. This research investigates changes in time allocations and exposures that result from traffic congestion. Time shifts were derived using data from the National Human Activity Pattern Survey (NHAPS), which was aggregated to nine microenvironments (six indoor locations, two outdoor locations and one transport location). After imputing missing values, handling outliers, and conducting other quality checks, these data were stratified by respondent age, employment status and period (weekday/weekend). Trade-offs or time-shift coefficients between time spent in vehicles and the eight other microenvironments were then estimated using robust regression. For children and retirees, congestion primarily reduced the time spent at home; for older children and working adults, congestion shifted the time spent at home as well as time in schools, public buildings, and other indoor environments. Changes in benzene and PM(2.5) exposure were estimated for the current average travel delay in the U.S. (9 min day(-1)) and other scenarios using the estimated time shifts coefficients, concentrations in key microenvironments derived from the literature, and a probabilistic analysis. Changes in exposures depended on the duration of the congestion and the pollutant. For example, a 30 min day(-1) travel delay was determined to account for 21+/-12% of current exposure to benzene and 14+/-8% of PM(2.5) exposure. The time allocation shifts and the dynamic approach to TAPs improve estimates of exposure impacts from congestion and other recurring events.

  7. Estimating a population cumulative incidence under calendar time trends

    DEFF Research Database (Denmark)

    Hansen, Stefan N; Overgaard, Morten; Andersen, Per K

    2017-01-01

    BACKGROUND: The risk of a disease or psychiatric disorder is frequently measured by the age-specific cumulative incidence. Cumulative incidence estimates are often derived in cohort studies with individuals recruited over calendar time and with the end of follow-up governed by a specific date...... by calendar time trends, the total sample Kaplan-Meier and Aalen-Johansen estimators do not provide useful estimates of the general risk in the target population. We present some alternatives to this type of analysis. RESULTS: We show how a proportional hazards model may be used to extrapolate disease risk...... estimates if proportionality is a reasonable assumption. If not reasonable, we instead advocate that a more useful description of the disease risk lies in the age-specific cumulative incidence curves across strata given by time of entry or perhaps just the end of follow-up estimates across all strata...

  8. Estimating Time-to-Collision with Retinitis Pigmentosa

    Science.gov (United States)

    Jones, Tim

    2006-01-01

    This article reports on the ability of observers who are sighted and those with low vision to make time-to-collision (TTC) estimations using video. The TTC estimations made by the observers with low vision were comparable to those made by the sighted observers, and both groups made underestimation errors that were similar to those that were…

  9. A soft-computing methodology for noninvasive time-spatial temperature estimation.

    Science.gov (United States)

    Teixeira, César A; Ruano, Maria Graça; Ruano, António E; Pereira, Wagner C A

    2008-02-01

    The safe and effective application of thermal therapies is restricted due to lack of reliable noninvasive temperature estimators. In this paper, the temporal echo-shifts of backscattered ultrasound signals, collected from a gel-based phantom, were tracked and assigned with the past temperature values as radial basis functions neural networks input information. The phantom was heated using a piston-like therapeutic ultrasound transducer. The neural models were assigned to estimate the temperature at different intensities and points arranged across the therapeutic transducer radial line (60 mm apart from the transducer face). Model inputs, as well as the number of neurons were selected using the multiobjective genetic algorithm (MOGA). The best attained models present, in average, a maximum absolute error less than 0.5 degrees C, which is pointed as the borderline between a reliable and an unreliable estimator in hyperthermia/diathermia. In order to test the spatial generalization capacity, the best models were tested using spatial points not yet assessed, and some of them presented a maximum absolute error inferior to 0.5 degrees C, being "elected" as the best models. It should be also stressed that these best models present implementational low-complexity, as desired for real-time applications.

  10. H∞ state estimation for discrete-time memristive recurrent neural networks with stochastic time-delays

    Science.gov (United States)

    Liu, Hongjian; Wang, Zidong; Shen, Bo; Alsaadi, Fuad E.

    2016-07-01

    This paper deals with the robust H∞ state estimation problem for a class of memristive recurrent neural networks with stochastic time-delays. The stochastic time-delays under consideration are governed by a Bernoulli-distributed stochastic sequence. The purpose of the addressed problem is to design the robust state estimator such that the dynamics of the estimation error is exponentially stable in the mean square, and the prescribed ? performance constraint is met. By utilizing the difference inclusion theory and choosing a proper Lyapunov-Krasovskii functional, the existence condition of the desired estimator is derived. Based on it, the explicit expression of the estimator gain is given in terms of the solution to a linear matrix inequality. Finally, a numerical example is employed to demonstrate the effectiveness and applicability of the proposed estimation approach.

  11. Variation in the estimations of ETo and crop water use due to the sensor accuracy of the meteorological variables

    Directory of Open Access Journals (Sweden)

    R. Moratiel

    2013-06-01

    Full Text Available In agricultural ecosystems the use of evapotranspiration (ET to improve irrigation water management is generally widespread. Commonly, the crop ET (ETc is estimated by multiplying the reference crop evapotranspiration (ETo by a crop coefficient (Kc. Accurate estimation of ETo is critical because it is the main factor affecting the calculation of crop water use and water management. The ETo is generally estimated from recorded meteorological variables at reference weather stations. The main objective of this paper was assessing the effect of the uncertainty due to random noise in the sensors used for measurement of meteorological variables on the estimation of ETo, crop ET and net irrigation requirements of grain corn and alfalfa in three irrigation districts of the middle Ebro River basin. Five scenarios were simulated, four of them individually considering each recorded meteorological variable (temperature, relative humidity, solar radiation and wind speed and a fifth scenario combining together the uncertainty of all sensors. The uncertainty in relative humidity for irrigation districts Riegos del Alto Aragón (RAA and Bardenas (BAR, and temperature for irrigation district Canal de Aragón y Cataluña (CAC, were the two most important factors affecting the estimation of ETo, corn ET (ETc_corn, alfalfa ET (ETc_alf, net corn irrigation water requirements (IRncorn and net alfalfa irrigation water requirements (IRnalf. Nevertheless, this effect was never greater than ±0.5% over annual scale time. The wind speed variable (Scenario 3 was the third variable more influential in the fluctuations (± of evapotranspiration, followed by solar radiation. Considering the accuracy for all sensors over annual scale time, the variation was about ±1% of ETo, ETc_corn, ETc_alf, IRncorn, and IRnalf. The fluctuations of evapotranspiration were higher at shorter time scale. ETo daily fluctuation remained lower than 5 % during the growing season of corn and

  12. Similarity estimators for irregular and age uncertain time series

    Science.gov (United States)

    Rehfeld, K.; Kurths, J.

    2013-09-01

    Paleoclimate time series are often irregularly sampled and age uncertain, which is an important technical challenge to overcome for successful reconstruction of past climate variability and dynamics. Visual comparison and interpolation-based linear correlation approaches have been used to infer dependencies from such proxy time series. While the first is subjective, not measurable and not suitable for the comparison of many datasets at a time, the latter introduces interpolation bias, and both face difficulties if the underlying dependencies are nonlinear. In this paper we investigate similarity estimators that could be suitable for the quantitative investigation of dependencies in irregular and age uncertain time series. We compare the Gaussian-kernel based cross correlation (gXCF, Rehfeld et al., 2011) and mutual information (gMI, Rehfeld et al., 2013) against their interpolation-based counterparts and the new event synchronization function (ESF). We test the efficiency of the methods in estimating coupling strength and coupling lag numerically, using ensembles of synthetic stalagmites with short, autocorrelated, linear and nonlinearly coupled proxy time series, and in the application to real stalagmite time series. In the linear test case coupling strength increases are identified consistently for all estimators, while in the nonlinear test case the correlation-based approaches fail. The lag at which the time series are coupled is identified correctly as the maximum of the similarity functions in around 60-55% (in the linear case) to 53-42% (for the nonlinear processes) of the cases when the dating of the synthetic stalagmite is perfectly precise. If the age uncertainty increases beyond 5% of the time series length, however, the true coupling lag is not identified more often than the others for which the similarity function was estimated. Age uncertainty contributes up to half of the uncertainty in the similarity estimation process. Time series irregularity

  13. Similarity estimators for irregular and age-uncertain time series

    Science.gov (United States)

    Rehfeld, K.; Kurths, J.

    2014-01-01

    Paleoclimate time series are often irregularly sampled and age uncertain, which is an important technical challenge to overcome for successful reconstruction of past climate variability and dynamics. Visual comparison and interpolation-based linear correlation approaches have been used to infer dependencies from such proxy time series. While the first is subjective, not measurable and not suitable for the comparison of many data sets at a time, the latter introduces interpolation bias, and both face difficulties if the underlying dependencies are nonlinear. In this paper we investigate similarity estimators that could be suitable for the quantitative investigation of dependencies in irregular and age-uncertain time series. We compare the Gaussian-kernel-based cross-correlation (gXCF, Rehfeld et al., 2011) and mutual information (gMI, Rehfeld et al., 2013) against their interpolation-based counterparts and the new event synchronization function (ESF). We test the efficiency of the methods in estimating coupling strength and coupling lag numerically, using ensembles of synthetic stalagmites with short, autocorrelated, linear and nonlinearly coupled proxy time series, and in the application to real stalagmite time series. In the linear test case, coupling strength increases are identified consistently for all estimators, while in the nonlinear test case the correlation-based approaches fail. The lag at which the time series are coupled is identified correctly as the maximum of the similarity functions in around 60-55% (in the linear case) to 53-42% (for the nonlinear processes) of the cases when the dating of the synthetic stalagmite is perfectly precise. If the age uncertainty increases beyond 5% of the time series length, however, the true coupling lag is not identified more often than the others for which the similarity function was estimated. Age uncertainty contributes up to half of the uncertainty in the similarity estimation process. Time series irregularity

  14. Worldwide incidence of malaria in 2009: estimates, time trends, and a critique of methods.

    Directory of Open Access Journals (Sweden)

    Richard E Cibulskis

    2011-12-01

    Full Text Available BACKGROUND: Measuring progress towards Millennium Development Goal 6, including estimates of, and time trends in, the number of malaria cases, has relied on risk maps constructed from surveys of parasite prevalence, and on routine case reports compiled by health ministries. Here we present a critique of both methods, illustrated with national incidence estimates for 2009. METHODS AND FINDINGS: We compiled information on the number of cases reported by National Malaria Control Programs in 99 countries with ongoing malaria transmission. For 71 countries we estimated the total incidence of Plasmodium falciparum and P. vivax by adjusting the number of reported cases using data on reporting completeness, the proportion of suspects that are parasite-positive, the proportion of confirmed cases due to each Plasmodium species, and the extent to which patients use public sector health facilities. All four factors varied markedly among countries and regions. For 28 African countries with less reliable routine surveillance data, we estimated the number of cases from model-based methods that link measures of malaria transmission with case incidence. In 2009, 98% of cases were due to P. falciparum in Africa and 65% in other regions. There were an estimated 225 million malaria cases (5th-95th centiles, 146-316 million worldwide, 176 (110-248 million in the African region, and 49 (36-68 million elsewhere. Our estimates are lower than other published figures, especially survey-based estimates for non-African countries. CONCLUSIONS: Estimates of malaria incidence derived from routine surveillance data were typically lower than those derived from surveys of parasite prevalence. Carefully interpreted surveillance data can be used to monitor malaria trends in response to control efforts, and to highlight areas where malaria programs and health information systems need to be strengthened. As malaria incidence declines around the world, evaluation of control efforts

  15. Studies on the radioactive contamination due to nuclear detonations III. On the method of estimating the probable time of nuclear detonation from the measurements of gross-activity

    Energy Technology Data Exchange (ETDEWEB)

    Nishiwaki, Yasushi [Nuclear Reactor Laboratory, Tokyo Institute of Technology, Tokyo (Japan); Nuclear Reactor Laboratoroy, Kinki University, Fuse City, Osaka Precture (Japan)

    1961-11-25

    Since it has been observed in Spring of 1954 that a considerable amount of fission products mixture fell with the rain following a large scale nuclear detonation conducted in Bikini area in the South Pacific by the United States Atomic Energy Commission, it has become important, especially from the health physics standpoint, to estimate the effective average age of the fission products mixture after the nuclear detonation. If the energy transferred to the atmospheric air at the time of nuclear detonation is large enough (order of megaton at the distance of about 4000 km), the probable time and test site of nuclear detonation may be estimated with considerable accuracy, from the records of the pressure wave caused by the detonation in the microbarographs at different meteorological stations. Even in this case, in order to estimate the possible correlation between the artificial radioactivity observed in the rain and the probable detonation, it is often times desirable to estimate the effective age of the fission products mixture in the rain from the decay measurement of the radioactivity.

  16. Studies on the radioactive contamination due to nuclear detonations III. On the method of estimating the probable time of nuclear detonation from the measurements of gross-activity

    International Nuclear Information System (INIS)

    Nishiwaki, Yasushi

    1961-01-01

    Since it has been observed in Spring of 1954 that a considerable amount of fission products mixture fell with the rain following a large scale nuclear detonation conducted in Bikini area in the South Pacific by the United States Atomic Energy Commission, it has become important, especially from the health physics standpoint, to estimate the effective average age of the fission products mixture after the nuclear detonation. If the energy transferred to the atmospheric air at the time of nuclear detonation is large enough (order of megaton at the distance of about 4000 km), the probable time and test site of nuclear detonation may be estimated with considerable accuracy, from the records of the pressure wave caused by the detonation in the microbarographs at different meteorological stations. Even in this case, in order to estimate the possible correlation between the artificial radioactivity observed in the rain and the probable detonation, it is often times desirable to estimate the effective age of the fission products mixture in the rain from the decay measurement of the radioactivity

  17. Evapotranspiration estimates and consequences due to errors in the determination of the net radiation and advective effects

    International Nuclear Information System (INIS)

    Oliveira, G.M. de; Leitao, M. de M.V.B.R.

    2000-01-01

    The objective of this study was to analyze the consequences in the evapotranspiration estimates (ET) during the growing cycle of a peanut crop due to the errors committed in the determination of the radiation balance (Rn), as well as those caused by the advective effects. This research was conducted at the Experimental Station of CODEVASF in an irrigated perimeter located in the city of Rodelas, BA, during the period of September to December of 1996. The results showed that errors of the order of 2.2 MJ m -2 d -1 in the calculation of Rn, and consequently in the estimate of ET, can occur depending on the time considered for the daily total of Rn. It was verified that the surrounding areas of the experimental field, as well as the areas of exposed soil within the field, contributed significantly to the generation of local advection of sensible heat, which resulted in the increase of the evapotranspiration [pt

  18. Beyond Newton's Law of Cooling--Estimation of Time since Death

    Science.gov (United States)

    Leinbach, Carl

    2011-01-01

    The estimate of the time since death and, thus, the time of death is strictly that, an estimate. However, the time of death can be an important piece of information in some coroner's cases, especially those that involve criminal or insurance investigations. It has been known almost from the beginning of time that bodies cool after the internal…

  19. Time Skew Estimator for Dual-Polarization QAM Transmitters

    DEFF Research Database (Denmark)

    Medeiros Diniz, Júlio César; Da Ros, Francesco; Jones, Rasmus Thomas

    2017-01-01

    A simple method for joint estimation of transmitter’s in-phase/quadrature and inter-polarization time skew is proposed and experimentally demonstrated. The method is based on clock tone extraction of a photodetected signal and genetic algorithm. The maximum estimation error was 0.5 ps....

  20. Doses and risk estimates to the human conceptus due to internal prenatal exposure to radioactive caesium

    International Nuclear Information System (INIS)

    Kalef-Ezra, J.A.

    1997-01-01

    The 1986 nuclear reactor accident at Chernobyl resulted in widespread internal contamination by radioactive caesium. The aim of the present study was to estimate the doses to embryos/fetus in Greece attributed to maternal 134 Cs and 137 Cs intake and the consequent health risks to their offspring. In pregnant women the concentration of total-body caesium (TBCs) was lower than in age-matched non-pregnant women measured during the same month. A detailed study of intake and retention in the members of one family carried out during the three years that followed the accident indicated that the biological half-time of caesium in the women decreased by a factor of two shortly after conception. Then at partus, there was an increase in the biological half-time, reaching a value similar to that before conception. The total-body potassium concentration was constant over the entire period. Doses to the embryo/fetus due to maternal intake was estimated to be about 150 μGy maximally in those conceived between November 1986 and March 1987. When conception took place later, the prenatal dose followed an exponential reduction with a half-time of about 170 d. These prenatal doses do not exceed the doses from either the natural internal potassium, or from the usual external background sources. The risks attributed to maternal 134 Cs and 137 Cs intake were considerably lower than levels that would justify consideration of termination of a pregnancy. In the absence of these data however, 2500 otherwise wanted pregnancies in Greece were terminated following the Chernobyl accident. (author)

  1. Limitations of the time slide method of background estimation

    International Nuclear Information System (INIS)

    Was, Michal; Bizouard, Marie-Anne; Brisson, Violette; Cavalier, Fabien; Davier, Michel; Hello, Patrice; Leroy, Nicolas; Robinet, Florent; Vavoulidis, Miltiadis

    2010-01-01

    Time shifting the output of gravitational wave detectors operating in coincidence is a convenient way of estimating the background in a search for short-duration signals. In this paper, we show how non-stationary data affect the background estimation precision. We present a method of measuring the fluctuations of the data and computing its effects on a coincident search. In particular, we show that for fluctuations of moderate amplitude, time slides larger than the fluctuation time scales can be used. We also recall how the false alarm variance saturates with the number of time shifts.

  2. Limitations of the time slide method of background estimation

    Energy Technology Data Exchange (ETDEWEB)

    Was, Michal; Bizouard, Marie-Anne; Brisson, Violette; Cavalier, Fabien; Davier, Michel; Hello, Patrice; Leroy, Nicolas; Robinet, Florent; Vavoulidis, Miltiadis, E-mail: mwas@lal.in2p3.f [LAL, Universite Paris-Sud, CNRS/IN2P3, Orsay (France)

    2010-10-07

    Time shifting the output of gravitational wave detectors operating in coincidence is a convenient way of estimating the background in a search for short-duration signals. In this paper, we show how non-stationary data affect the background estimation precision. We present a method of measuring the fluctuations of the data and computing its effects on a coincident search. In particular, we show that for fluctuations of moderate amplitude, time slides larger than the fluctuation time scales can be used. We also recall how the false alarm variance saturates with the number of time shifts.

  3. Estimation of absorbed doses in humans due to intravenous administration of fluorine-18-fluorodeoxyglucose in PET studies

    International Nuclear Information System (INIS)

    Mejia, A.A.; Nakamura, T.; Masatoshi, I.; Hatazawa, J.; Masaki, M.; Watanuki, S.

    1991-01-01

    Radiation absorbed doses due to intravenous administration of fluorine-18-fluorodeoxyglucose in positron emission tomography (PET) studies were estimated in normal volunteers. The time-activity curves were obtained for seven human organs (brain, heart, kidney, liver, lung, pancreas, and spleen) by using dynamic PET scans and for bladder content by using a single detector. These time-activity curves were used for the calculation of the cumulative activity in these organs. Absorbed doses were calculated by the MIRD method using the absorbed dose per unit of cumulated activity, 'S' value, transformed for the Japanese physique and the organ masses of the Japanese reference man. The bladder wall and the heart were the organs receiving higher doses of 1.2 x 10(-1) and 4.5 x 10(-2) mGy/MBq, respectively. The brain received a dose of 2.9 x 10(-2) mGy/MBq, and other organs received doses between 1.0 x 10(-2) and 3.0 x 10(-2) mGy/MBq. The effective dose equivalent was estimated to be 2.4 x 10(-2) mSv/MBq. These results were comparable to values of absorbed doses reported by other authors on the radiation dosimetry of this radiopharmaceutical

  4. Estimation of sojourn time in chronic disease screening without data on interval cases.

    Science.gov (United States)

    Chen, T H; Kuo, H S; Yen, M F; Lai, M S; Tabar, L; Duffy, S W

    2000-03-01

    Estimation of the sojourn time on the preclinical detectable period in disease screening or transition rates for the natural history of chronic disease usually rely on interval cases (diagnosed between screens). However, to ascertain such cases might be difficult in developing countries due to incomplete registration systems and difficulties in follow-up. To overcome this problem, we propose three Markov models to estimate parameters without using interval cases. A three-state Markov model, a five-state Markov model related to regional lymph node spread, and a five-state Markov model pertaining to tumor size are applied to data on breast cancer screening in female relatives of breast cancer cases in Taiwan. Results based on a three-state Markov model give mean sojourn time (MST) 1.90 (95% CI: 1.18-4.86) years for this high-risk group. Validation of these models on the basis of data on breast cancer screening in the age groups 50-59 and 60-69 years from the Swedish Two-County Trial shows the estimates from a three-state Markov model that does not use interval cases are very close to those from previous Markov models taking interval cancers into account. For the five-state Markov model, a reparameterized procedure using auxiliary information on clinically detected cancers is performed to estimate relevant parameters. A good fit of internal and external validation demonstrates the feasibility of using these models to estimate parameters that have previously required interval cancers. This method can be applied to other screening data in which there are no data on interval cases.

  5. Estimation of the collective dose in the Portuguese population due to medical procedures in 2010

    International Nuclear Information System (INIS)

    Teles, Pedro; Vaz, Pedro; Sousa, M. Carmen de; Paulo, Graciano; Santos, Joana; Pascoal, Ana; Cardoso, Gabriela; Santos, Ana Isabel; Lanca, Isabel; Matela, Nuno; Janeiro, Luis; Sousa, Patrick; Carvoeiras, Pedro; Parafita, Rui; Simaozinho, Paula

    2013-01-01

    In a wide range of medical fields, technological advancements have led to an increase in the average collective dose in national populations worldwide. Periodic estimations of the average collective population dose due to medical exposure is, therefore of utmost importance, and is now mandatory in countries within the European Union (article 12 of EURATOM directive 97/ 43). Presented in this work is a report on the estimation of the collective dose in the Portuguese population due to nuclear medicine diagnostic procedures and the Top 20 diagnostic radiology examinations, which represent the 20 exams that contribute the most to the total collective dose in diagnostic radiology and interventional procedures in Europe. This work involved the collaboration of a multidisciplinary taskforce comprising representatives of all major Portuguese stakeholders (universities, research institutions, public and private health care providers, administrative services of the National Healthcare System, scientific and professional associations and private service providers). This allowed us to gather a comprehensive amount of data necessary for a robust estimation of the collective effective dose to the Portuguese population. The methodology used for data collection and dose estimation was based on European Commission recommendations, as this work was performed in the framework of the European wide Dose Datamed II project. This is the first study estimating the collective dose for the population in Portugal, considering such a wide national coverage and range of procedures and consisting of important baseline reference data. The taskforce intends to continue developing periodic collective dose estimations in the future. The estimated annual average effective dose for the Portuguese population was of 0.080±0.017 mSv caput -1 for nuclear medicine exams and of 0.96±0.68 mSv caput -1 for the Top 20 diagnostic radiology exams. (authors)

  6. Single-machine common/slack due window assignment problems with linear decreasing processing times

    Science.gov (United States)

    Zhang, Xingong; Lin, Win-Chin; Wu, Wen-Hsiang; Wu, Chin-Chia

    2017-08-01

    This paper studies linear non-increasing processing times and the common/slack due window assignment problems on a single machine, where the actual processing time of a job is a linear non-increasing function of its starting time. The aim is to minimize the sum of the earliness cost, tardiness cost, due window location and due window size. Some optimality results are discussed for the common/slack due window assignment problems and two O(n log n) time algorithms are presented to solve the two problems. Finally, two examples are provided to illustrate the correctness of the corresponding algorithms.

  7. Estimating bus passenger waiting times from incomplete bus arrivals data

    OpenAIRE

    McLeod, F.N.

    2007-01-01

    This paper considers the problem of estimating bus passenger waiting times at bus stops using incomplete bus arrivals data. This is of importance to bus operators and regulators as passenger waiting time is a key performance measure. Average waiting times are usually estimated from bus headways, that is, time gaps between buses. It is both time-consuming and expensive to measure bus arrival times manually so methods using automatic vehicle location systems are attractive; however, these syste...

  8. Multi-processor system for real-time deconvolution and flow estimation in medical ultrasound

    DEFF Research Database (Denmark)

    Jensen, Jesper Lomborg; Jensen, Jørgen Arendt; Stetson, Paul F.

    1996-01-01

    of the algorithms. Many of the algorithms can only be properly evaluated in a clinical setting with real-time processing, which generally cannot be done with conventional equipment. This paper therefore presents a multi-processor system capable of performing 1.2 billion floating point operations per second on RF...... filter is used with a second time-reversed recursive estimation step. Here it is necessary to perform about 70 arithmetic operations per RF sample or about 1 billion operations per second for real-time deconvolution. Furthermore, these have to be floating point operations due to the adaptive nature...... interfaced to our previously-developed real-time sampling system that can acquire RF data at a rate of 20 MHz and simultaneously transmit the data at 20 MHz to the processing system via several parallel channels. These two systems can, thus, perform real-time processing of ultrasound data. The advantage...

  9. A Dynamic Travel Time Estimation Model Based on Connected Vehicles

    Directory of Open Access Journals (Sweden)

    Daxin Tian

    2015-01-01

    Full Text Available With advances in connected vehicle technology, dynamic vehicle route guidance models gradually become indispensable equipment for drivers. Traditional route guidance models are designed to direct a vehicle along the shortest path from the origin to the destination without considering the dynamic traffic information. In this paper a dynamic travel time estimation model is presented which can collect and distribute traffic data based on the connected vehicles. To estimate the real-time travel time more accurately, a road link dynamic dividing algorithm is proposed. The efficiency of the model is confirmed by simulations, and the experiment results prove the effectiveness of the travel time estimation method.

  10. Response of orthotropic micropolar elastic medium due to time ...

    Indian Academy of Sciences (India)

    R. Narasimhan (Krishtel eMaging) 1461 1996 Oct 15 13:05:22

    namic response of anisotropic continuum has received the attention of ... linear theory of micropolar elasticity and bending of orthotropic micropolar ... medium due to time harmonic concentrated load, the continuum is divided into two half-.

  11. Life time estimation of SSCs for decommissioning safety of nuclear facilities

    International Nuclear Information System (INIS)

    Jeong, Kwan-Seong; Lee, Kune-Woo; Moon, Jei-Kwon; Jeong, Seong-Young; Lee, Jung-Jun; Kim, Geun-Ho; Choi, Byung-Seon

    2012-01-01

    Highlights: ► This paper suggests the expectation algorithm of SSCs life time for decommissioning safety of nuclear facilities. ► The life time of SSCs can be estimated by using fuzzy theory. ► The estimated results depend on the membership functions and performance characteristic functions. - Abstract: This paper suggests the estimation algorithm for life time of structure, system and components (SSCs) for decommissioning safety of nuclear facilities using the performance data of linguistic languages and fuzzy theory. The fuzzy estimation algorithm of life time can be easily applicable but the estimated results depend on the relevant membership functions and performance characteristic functions. This method will be expected to be very useful for maintenance and decommissioning of nuclear facilities’ SSCs as a safety assessment tool.

  12. Noise level estimation in weakly nonlinear slowly time-varying systems

    International Nuclear Information System (INIS)

    Aerts, J R M; Dirckx, J J J; Lataire, J; Pintelon, R

    2008-01-01

    Recently, a method using multisine excitation was proposed for estimating the frequency response, the nonlinear distortions and the disturbing noise of weakly nonlinear time-invariant systems. This method has been demonstrated on the measurement of nonlinear distortions in the vibration of acoustically driven systems such as a latex membrane, which is a good example of a time-invariant system [1]. However, not all systems are perfectly time invariant, e.g. biomechanical systems. This time variation can be misinterpreted as an elevated noise floor, and the classical noise estimation method gives a wrong result. Two improved methods to retrieve the correct noise information from the measurements are presented. Both of them make use of multisine excitations. First, it is demonstrated that the improved methods give the same result as the classical noise estimation method when applied to a time-invariant system (high-quality microphone membrane). Next, it is demonstrated that the new methods clearly give an improved estimate of the noise level on time-varying systems. As an application example results for the vibration response of an eardrum are shown

  13. Generalized synchronization-based multiparameter estimation in modulated time-delayed systems

    Science.gov (United States)

    Ghosh, Dibakar; Bhattacharyya, Bidyut K.

    2011-09-01

    We propose a nonlinear active observer based generalized synchronization scheme for multiparameter estimation in time-delayed systems with periodic time delay. A sufficient condition for parameter estimation is derived using Krasovskii-Lyapunov theory. The suggested tool proves to be globally and asymptotically stable by means of Krasovskii-Lyapunov method. With this effective method, parameter identification and generalized synchronization of modulated time-delayed systems with all the system parameters unknown, can be achieved simultaneously. We restrict our study for multiple parameter estimation in modulated time-delayed systems with single state variable only. Theoretical proof and numerical simulation demonstrate the effectiveness and feasibility of the proposed technique. The block diagram of electronic circuit for multiple time delay system shows that the method is easily applicable in practical communication problems.

  14. Compressed air leak detection based on time delay estimation using a portable multi-sensor ultrasonic detector

    International Nuclear Information System (INIS)

    Liao, Pingping; Cai, Maolin; Shi, Yan; Fan, Zichuan

    2013-01-01

    The conventional ultrasonic method for compressed air leak detection utilizes a directivity-based ultrasonic leak detector (DULD) to locate the leak. The location accuracy of this method is low due to the limit of the nominal frequency and the size of the ultrasonic sensor. In order to overcome this deficiency, a method based on time delay estimation (TDE) is proposed. The method utilizes three ultrasonic sensors arranged in an equilateral triangle to simultaneously receive the ultrasound generated by the leak. The leak can be located according to time delays between every two sensor signals. The theoretical accuracy of the method is analyzed, and it is found that the location error increases linearly with delay estimation error and the distance from the leak to the sensor plane, and the location error decreases with the distance between sensors. The average square difference function delay estimator with parabolic fitting is used and two practical techniques are devised to remove the anomalous delay estimates. Experimental results indicate that the location accuracy using the TDE-based ultrasonic leak detector is 6.5–8.3 times as high as that using the DULD. By adopting the proposed method, the leak can be located more accurately and easily, and then the detection efficiency is improved. (paper)

  15. Origin of Human Losses due to the Emilia Romagna, Italy, M5.9 Earthquake of 20 May 2012 and their Estimate in Real Time

    Science.gov (United States)

    Wyss, M.

    2012-12-01

    Estimating human losses within less than an hour worldwide requires assumptions and simplifications. Earthquake for which losses are accurately recorded after the event provide clues concerning the influence of error sources. If final observations and real time estimates differ significantly, data and methods to calculate losses may be modified or calibrated. In the case of the earthquake in the Emilia Romagna region with M5.9 on May 20th, the real time epicenter estimates of the GFZ and the USGS differed from the ultimate location by the INGV by 6 and 9 km, respectively. Fatalities estimated within an hour of the earthquake by the loss estimating tool QLARM, based on these two epicenters, numbered 20 and 31, whereas 7 were reported in the end, and 12 would have been calculated if the ultimate epicenter released by INGV had been used. These four numbers being small, do not differ statistically. Thus, the epicenter errors in this case did not appreciably influence the results. The QUEST team of INGV has reported intensities with I ≥ 5 at 40 locations with accuracies of 0.5 units and QLARM estimated I > 4.5 at 224 locations. The differences between the observed and calculated values at the 23 common locations show that the calculation in the 17 instances with significant differences were too high on average by one unit. By assuming higher than average attenuation within standard bounds for worldwide loss estimates, the calculated intensities model the observed ones better: For 57% of the locations, the difference was not significant; for the others, the calculated intensities were still somewhat higher than the observed ones. Using a generic attenuation law with higher than average attenuation, but not tailored to the region, the number of estimated fatalities becomes 12 compared to 7 reported ones. Thus, attenuation in this case decreased the discrepancy between observed and reported death by approximately a factor of two. The source of the fatalities is

  16. Aircraft Fault Detection Using Real-Time Frequency Response Estimation

    Science.gov (United States)

    Grauer, Jared A.

    2016-01-01

    A real-time method for estimating time-varying aircraft frequency responses from input and output measurements was demonstrated. The Bat-4 subscale airplane was used with NASA Langley Research Center's AirSTAR unmanned aerial flight test facility to conduct flight tests and collect data for dynamic modeling. Orthogonal phase-optimized multisine inputs, summed with pilot stick and pedal inputs, were used to excite the responses. The aircraft was tested in its normal configuration and with emulated failures, which included a stuck left ruddervator and an increased command path latency. No prior knowledge of a dynamic model was used or available for the estimation. The longitudinal short period dynamics were investigated in this work. Time-varying frequency responses and stability margins were tracked well using a 20 second sliding window of data, as compared to a post-flight analysis using output error parameter estimation and a low-order equivalent system model. This method could be used in a real-time fault detection system, or for other applications of dynamic modeling such as real-time verification of stability margins during envelope expansion tests.

  17. Production loss due to new subclinical mastitis in Dutch dairy cows estimated iwth a test-day model

    NARCIS (Netherlands)

    Halasa, T.; Nielen, M.; Roos, de S.; Hoorne, van R.; Jong, de G.; Lam, T.J.G.M.; Werven, van T.; Hogeveen, H.

    2009-01-01

    Milk, fat, and protein loss due to a new subclinical mastitis case may be economically important, and the objective of this study was to estimate this loss. The loss was estimated based on test-day (TD) cow records collected over a 1-yr period from 400 randomly selected Dutch dairy herds. After

  18. Single Machine Scheduling and Due Date Assignment with Past-Sequence-Dependent Setup Time and Position-Dependent Processing Time

    Directory of Open Access Journals (Sweden)

    Chuan-Li Zhao

    2014-01-01

    Full Text Available This paper considers single machine scheduling and due date assignment with setup time. The setup time is proportional to the length of the already processed jobs; that is, the setup time is past-sequence-dependent (p-s-d. It is assumed that a job's processing time depends on its position in a sequence. The objective functions include total earliness, the weighted number of tardy jobs, and the cost of due date assignment. We analyze these problems with two different due date assignment methods. We first consider the model with job-dependent position effects. For each case, by converting the problem to a series of assignment problems, we proved that the problems can be solved in On4 time. For the model with job-independent position effects, we proved that the problems can be solved in On3 time by providing a dynamic programming algorithm.

  19. Empirical Study of Travel Time Estimation and Reliability

    OpenAIRE

    Li, Ruimin; Chai, Huajun; Tang, Jin

    2013-01-01

    This paper explores the travel time distribution of different types of urban roads, the link and path average travel time, and variance estimation methods by analyzing the large-scale travel time dataset detected from automatic number plate readers installed throughout Beijing. The results show that the best-fitting travel time distribution for different road links in 15 min time intervals differs for different traffic congestion levels. The average travel time for all links on all days can b...

  20. Estimation of unemployment rates using small area estimation model by combining time series and cross-sectional data

    Science.gov (United States)

    Muchlisoh, Siti; Kurnia, Anang; Notodiputro, Khairil Anwar; Mangku, I. Wayan

    2016-02-01

    Labor force surveys conducted over time by the rotating panel design have been carried out in many countries, including Indonesia. Labor force survey in Indonesia is regularly conducted by Statistics Indonesia (Badan Pusat Statistik-BPS) and has been known as the National Labor Force Survey (Sakernas). The main purpose of Sakernas is to obtain information about unemployment rates and its changes over time. Sakernas is a quarterly survey. The quarterly survey is designed only for estimating the parameters at the provincial level. The quarterly unemployment rate published by BPS (official statistics) is calculated based on only cross-sectional methods, despite the fact that the data is collected under rotating panel design. The study purpose to estimate a quarterly unemployment rate at the district level used small area estimation (SAE) model by combining time series and cross-sectional data. The study focused on the application and comparison between the Rao-Yu model and dynamic model in context estimating the unemployment rate based on a rotating panel survey. The goodness of fit of both models was almost similar. Both models produced an almost similar estimation and better than direct estimation, but the dynamic model was more capable than the Rao-Yu model to capture a heterogeneity across area, although it was reduced over time.

  1. An Embedded Device for Real-Time Noninvasive Intracranial Pressure Estimation.

    Science.gov (United States)

    Matthews, Jonathan M; Fanelli, Andrea; Heldt, Thomas

    2018-01-01

    The monitoring of intracranial pressure (ICP) is indicated for diagnosing and guiding therapy in many neurological conditions. Current monitoring methods, however, are highly invasive, limiting their use to the most critically ill patients only. Our goal is to develop and test an embedded device that performs all necessary mathematical operations in real-time for noninvasive ICP (nICP) estimation based on a previously developed model-based approach that uses cerebral blood flow velocity (CBFV) and arterial blood pressure (ABP) waveforms. The nICP estimation algorithm along with the required preprocessing steps were implemented on an NXP LPC4337 microcontroller unit (MCU). A prototype device using the MCU was also developed, complete with display, recording functionality, and peripheral interfaces for ABP and CBFV monitoring hardware. The device produces an estimate of mean ICP once per minute and performs the necessary computations in 410 ms, on average. Real-time nICP estimates differed from the original batch-mode MATLAB implementation of theestimation algorithm by 0.63 mmHg (root-mean-square error). We have demonstrated that real-time nICP estimation is possible on a microprocessor platform, which offers the advantages of low cost, small size, and product modularity over a general-purpose computer. These attributes take a step toward the goal of real-time nICP estimation at the patient's bedside in a variety of clinical settings.

  2. Radiological Characterization of cuban coffee and estimation of the doses received by the population due to coffee consumption

    International Nuclear Information System (INIS)

    Fernandez Gomez, I.M.; Rodriguez Castro, G.; Perez Sanchez, D.

    1996-01-01

    The purpose of this paper is to study the radioactivity levels in the coffee produced and consumed in our country and to estimate the doses received by the Cuban people due to its consumption. As a results of this study the most relevant radionuclide was K 40 , due to the concentration levels found only considering it when estimating the annual committed dose that is received thought this via. The K 40 concentration value present in the infusion represented a dose for the consumer of 15.6 u Sv/year

  3. The current duration design for estimating the time to pregnancy distribution

    DEFF Research Database (Denmark)

    Gasbarra, Dario; Arjas, Elja; Vehtari, Aki

    2015-01-01

    This paper was inspired by the studies of Niels Keiding and co-authors on estimating the waiting time-to-pregnancy (TTP) distribution, and in particular on using the current duration design in that context. In this design, a cross-sectional sample of women is collected from those who are currently...... attempting to become pregnant, and then by recording from each the time she has been attempting. Our aim here is to study the identifiability and the estimation of the waiting time distribution on the basis of current duration data. The main difficulty in this stems from the fact that very short waiting...... times are only rarely selected into the sample of current durations, and this renders their estimation unstable. We introduce here a Bayesian method for this estimation problem, prove its asymptotic consistency, and compare the method to some variants of the non-parametric maximum likelihood estimators...

  4. Estimating mortality due to cigarette smoking

    DEFF Research Database (Denmark)

    Brønnum-Hansen, H; Juel, K

    2000-01-01

    We estimated the mortality from various diseases caused by cigarette smoking using two methods and compared the results. In one method, the "Prevent" model is used to simulate the effect on mortality of the prevalence of cigarette smoking derived retrospectively. The other method, suggested by R....... Peto et al (Lancet 1992;339:1268-1278), requires data on mortality from lung cancer among people who have never smoked and among smokers, but it does not require data on the prevalence of smoking. In the Prevent model, 33% of deaths among men and 23% of those among women in 1993 from lung cancer...... are small and appear to be explicable. The Prevent model can be used for more general scenarios of effective health promotion, but it requires more data than the Peto et al method, which can be used only to estimate mortality related to smoking....

  5. Time-Distance Helioseismology: Noise Estimation

    Science.gov (United States)

    Gizon, L.; Birch, A. C.

    2004-10-01

    As in global helioseismology, the dominant source of noise in time-distance helioseismology measurements is realization noise due to the stochastic nature of the excitation mechanism of solar oscillations. Characterizing noise is important for the interpretation and inversion of time-distance measurements. In this paper we introduce a robust definition of travel time that can be applied to very noisy data. We then derive a simple model for the full covariance matrix of the travel-time measurements. This model depends only on the expectation value of the filtered power spectrum and assumes that solar oscillations are stationary and homogeneous on the solar surface. The validity of the model is confirmed through comparison with SOHO MDI measurements in a quiet-Sun region. We show that the correlation length of the noise in the travel times is about half the dominant wavelength of the filtered power spectrum. We also show that the signal-to-noise ratio in quiet-Sun travel-time maps increases roughly as the square root of the observation time and is at maximum for a distance near half the length scale of supergranulation.

  6. Fuzzy logic estimator of rotor time constant in induction motors

    Energy Technology Data Exchange (ETDEWEB)

    Alminoja, J. [Tampere University of Technology (Finland). Control Engineering Laboratory; Koivo, H. [Helsinki University of Technology, Otaniemi (Finland). Control Engineering Laboratory

    1997-12-31

    Vector control of AC machines is a well-known and widely used technique in induction machine control. It offers an exact method for speed control of induction motors, but it is also sensitive to the changes in machine parameters. E.g. rotor time constant has a strong dependence on temperature. In this paper a fuzzy logic estimator is developed, with which the rotor time constant can be estimated when the machine has a load. It is more simple than the estimators proposed in the literature. The fuzzy estimator is tested by simulation when step-wise abrupt changes and slow drifting occurs. (orig.) 7 refs.

  7. Radiation dose estimates due to air particulate emissions from selected phosphate industry operations

    International Nuclear Information System (INIS)

    Partridge, J.E.; Horton, T.R.; Sensintaffar, E.L.; Boysen, G.A.

    1978-06-01

    The EPA Office of Radiation Programs has conducted a series of studies to determine the radiological impact of the phosphate mining and milling industry. This report describes the efforts to estimate the radiation doses due to airborne emissions of particulates from selected phosphate milling operations in Florida. Two wet process phosphoric acid plants and one ore drying facility were selected for this study. The 1976 Annual Operations/Emissions Report, submitted by each facility to the Florida Department of Environmental Regulation, and a field survey trip by EPA personnel to each facility were used to develop data for dose calculations. The field survey trip included sampling for stack emissions and ambient air samples collected in the general vicinity of each plant. Population and individual radiation dose estimates are made based on these sources of data

  8. Methodology for Time-Domain Estimation of Storm-Time Electric Fields Using the 3D Earth Impedance

    Science.gov (United States)

    Kelbert, A.; Balch, C. C.; Pulkkinen, A. A.; Egbert, G. D.; Love, J. J.; Rigler, E. J.; Fujii, I.

    2016-12-01

    Magnetic storms can induce geoelectric fields in the Earth's electrically conducting interior, interfering with the operations of electric-power grid industry. The ability to estimate these electric fields at Earth's surface in close to real-time and to provide local short-term predictions would improve the ability of the industry to protect their operations. At any given time, the electric field at the Earth's surface is a function of the time-variant magnetic activity (driven by the solar wind), and the local electrical conductivity structure of the Earth's crust and mantle. For this reason, implementation of an operational electric field estimation service requires an interdisciplinary, collaborative effort between space science, real-time space weather operations, and solid Earth geophysics. We highlight in this talk an ongoing collaboration between USGS, NOAA, NASA, Oregon State University, and the Japan Meteorological Agency, to develop algorithms that can be used for scenario analyses and which might be implemented in a real-time, operational setting. We discuss the development of a time domain algorithm that employs discrete time domain representation of the impedance tensor for a realistic 3D Earth, known as the discrete time impulse response (DTIR), convolved with the local magnetic field time series, to estimate the local electric field disturbances. The algorithm is validated against measured storm-time electric field data collected in the United States and Japan. We also discuss our plans for operational real-time electric field estimation using 3D Earth impedances.

  9. Propensity score estimation to address calendar time-specific channeling in comparative effectiveness research of second generation antipsychotics.

    Directory of Open Access Journals (Sweden)

    Stacie B Dusetzina

    estimates, possibly due to unmeasured confounding. Although calendar time-specific propensity scores appear to improve covariate balance, the impact on comparative effectiveness results is limited in this setting.

  10. Estimating time-varying RSA to examine psychophysiological linkage of marital dyads.

    Science.gov (United States)

    Gates, Kathleen M; Gatzke-Kopp, Lisa M; Sandsten, Maria; Blandon, Alysia Y

    2015-08-01

    One of the primary tenets of polyvagal theory dictates that parasympathetic influence on heart rate, often estimated by respiratory sinus arrhythmia (RSA), shifts rapidly in response to changing environmental demands. The current standard analytic approach of aggregating RSA estimates across time to arrive at one value fails to capture this dynamic property within individuals. By utilizing recent methodological developments that enable precise RSA estimates at smaller time intervals, we demonstrate the utility of computing time-varying RSA for assessing psychophysiological linkage (or synchrony) in husband-wife dyads using time-locked data collected in a naturalistic setting. © 2015 Society for Psychophysiological Research.

  11. Real time bayesian estimation of the epidemic potential of emerging infectious diseases.

    Directory of Open Access Journals (Sweden)

    Luís M A Bettencourt

    Full Text Available BACKGROUND: Fast changes in human demographics worldwide, coupled with increased mobility, and modified land uses make the threat of emerging infectious diseases increasingly important. Currently there is worldwide alert for H5N1 avian influenza becoming as transmissible in humans as seasonal influenza, and potentially causing a pandemic of unprecedented proportions. Here we show how epidemiological surveillance data for emerging infectious diseases can be interpreted in real time to assess changes in transmissibility with quantified uncertainty, and to perform running time predictions of new cases and guide logistics allocations. METHODOLOGY/PRINCIPAL FINDINGS: We develop an extension of standard epidemiological models, appropriate for emerging infectious diseases, that describes the probabilistic progression of case numbers due to the concurrent effects of (incipient human transmission and multiple introductions from a reservoir. The model is cast in terms of surveillance observables and immediately suggests a simple graphical estimation procedure for the effective reproductive number R (mean number of cases generated by an infectious individual of standard epidemics. For emerging infectious diseases, which typically show large relative case number fluctuations over time, we develop a bayesian scheme for real time estimation of the probability distribution of the effective reproduction number and show how to use such inferences to formulate significance tests on future epidemiological observations. CONCLUSIONS/SIGNIFICANCE: Violations of these significance tests define statistical anomalies that may signal changes in the epidemiology of emerging diseases and should trigger further field investigation. We apply the methodology to case data from World Health Organization reports to place bounds on the current transmissibility of H5N1 influenza in humans and establish a statistical basis for monitoring its evolution in real time.

  12. Challenges in automated estimation of capillary refill time in dogs

    Science.gov (United States)

    Cugmas, Blaž; Spigulis, Janis

    2018-02-01

    Capillary refill time (CRT) is a part of the cardiorespiratory examination in dogs. Changes in CRT can reflect pathological conditions like shock or anemia. Visual CRT estimation has low repeatability; therefore, optical systems for automated estimation have recently appeared. Since existing systems are unsuitable for use in dogs, we designed a simple, small and portable device, which could be easily used at veterinary clinic. The device was preliminarily tested on several measurement sites in two dogs. Not all measurement sites were suitable for CRT measurements due to underlying tissue optical and mechanical properties. The CRT measurements were possible on the labial mucosa, above the sternum and on the digit where CRT was in the range of values, retrieved from the color video of the visual CRT measurement. It seems that light penetration predominantly governs tissue optical response when the pressure is applied. Therefore, it is important to select a proper light, which reaches only superficial capillaries and does not penetrate deeper. Blue or green light is probably suitable for light skin or mucosa, on the other hand, red or near-infrared might be used for skin with pigmented or thick epidermis. Additionally, further improvements of the device design are considered, like adding a calibrated spring, which would insure application of consistent pressure.

  13. Accurate Lithium-ion battery parameter estimation with continuous-time system identification methods

    International Nuclear Information System (INIS)

    Xia, Bing; Zhao, Xin; Callafon, Raymond de; Garnier, Hugues; Nguyen, Truong; Mi, Chris

    2016-01-01

    Highlights: • Continuous-time system identification is applied in Lithium-ion battery modeling. • Continuous-time and discrete-time identification methods are compared in detail. • The instrumental variable method is employed to further improve the estimation. • Simulations and experiments validate the advantages of continuous-time methods. - Abstract: The modeling of Lithium-ion batteries usually utilizes discrete-time system identification methods to estimate parameters of discrete models. However, in real applications, there is a fundamental limitation of the discrete-time methods in dealing with sensitivity when the system is stiff and the storage resolutions are limited. To overcome this problem, this paper adopts direct continuous-time system identification methods to estimate the parameters of equivalent circuit models for Lithium-ion batteries. Compared with discrete-time system identification methods, the continuous-time system identification methods provide more accurate estimates to both fast and slow dynamics in battery systems and are less sensitive to disturbances. A case of a 2"n"d-order equivalent circuit model is studied which shows that the continuous-time estimates are more robust to high sampling rates, measurement noises and rounding errors. In addition, the estimation by the conventional continuous-time least squares method is further improved in the case of noisy output measurement by introducing the instrumental variable method. Simulation and experiment results validate the analysis and demonstrate the advantages of the continuous-time system identification methods in battery applications.

  14. Estimating PM2.5-associated mortality increase in California due to the Volkswagen emission control defeat device

    Science.gov (United States)

    Wang, Tianyang; Jerrett, Michael; Sinsheimer, Peter; Zhu, Yifang

    2016-11-01

    The Volkswagen Group of America (VW) was found by the US Environmental Protection Agency (EPA) and the California Air Resources Board (CARB) to have installed "defeat devices" and emit more oxides of nitrogen (NOx) than permitted under current EPA standards. In this paper, we quantify the hidden NOx emissions from this so-called VW scandal and the resulting public health impacts in California. The NOx emissions are calculated based on VW road test data and the CARB Emission Factors (EMFAC) model. Cumulative hidden NOx emissions from 2009 to 2015 were estimated to be over 3500 tons. Adult mortality changes were estimated based on ambient fine particulate matter (PM2.5) change due to secondary nitrate formation and the related concentration-response functions. We estimated that hidden NOx emissions from 2009 to 2015 have resulted in a total of 12 PM2.5-associated adult mortality increases in California. Most of the mortality increase happened in metropolitan areas, due to their high population and vehicle density.

  15. Joint Estimation and Decoding of Space-Time Trellis Codes

    Directory of Open Access Journals (Sweden)

    Zhang Jianqiu

    2002-01-01

    Full Text Available We explore the possibility of using an emerging tool in statistical signal processing, sequential importance sampling (SIS, for joint estimation and decoding of space-time trellis codes (STTC. First, we provide background on SIS, and then we discuss its application to space-time trellis code (STTC systems. It is shown through simulations that SIS is suitable for joint estimation and decoding of STTC with time-varying flat-fading channels when phase ambiguity is avoided. We used a design criterion for STTCs and temporally correlated channels that combats phase ambiguity without pilot signaling. We have shown by simulations that the design is valid.

  16. Time improvement of photoelectric effect calculation for absorbed dose estimation

    International Nuclear Information System (INIS)

    Massa, J M; Wainschenker, R S; Doorn, J H; Caselli, E E

    2007-01-01

    Ionizing radiation therapy is a very useful tool in cancer treatment. It is very important to determine absorbed dose in human tissue to accomplish an effective treatment. A mathematical model based on affected areas is the most suitable tool to estimate the absorbed dose. Lately, Monte Carlo based techniques have become the most reliable, but they are time expensive. Absorbed dose calculating programs using different strategies have to choose between estimation quality and calculating time. This paper describes an optimized method for the photoelectron polar angle calculation in photoelectric effect, which is significant to estimate deposited energy in human tissue. In the case studies, time cost reduction nearly reached 86%, meaning that the time needed to do the calculation is approximately 1/7 th of the non optimized approach. This has been done keeping precision invariant

  17. Maxima estimate of non gaussian process from observation of time history samples

    International Nuclear Information System (INIS)

    Borsoi, L.

    1987-01-01

    The problem constitutes a formidable task but is essential for industrial applications: extreme value design, fatigue analysis, etc. Even for the linear Gaussian case, the process ergodicity does not prevent the observation duration to be long enough to make reliable estimates. As well known, this duration is closely related to the process autocorrelation. A subterfuge, which distorts a little the problem, consists in considering periodic random process and in adjusting the observation duration to a complete period. In the nonlinear case, the stated problem is as much important as time history simulation is presently the only practicable way for analysing structures. Thus it is always interesting to adjust a tractable model to rough time history observations. In some cases this can be done with a Gumble-Poisson model. Then the difficulty is to make reliable estimates of the parameters involved in the model. Unfortunately it seems that even the use of sophisticated Bayesian method does not permit to reduce as wanted the necessary observation duration. One of the difficulties lies in process ergodicity which is often assumed to be based on physical considerations but which is not always rigorously stated. An other difficulty is the confusion between hidden informations - which can be extracted - and missing informations - which cannot be extracted. Finally it must be recalled that the obligation of considering time histories long enough is not always embarrassing due to the current computer cost reduction. (orig./HP)

  18. Continuous Fine-Fault Estimation with Real-Time GNSS

    Science.gov (United States)

    Norford, B. B.; Melbourne, T. I.; Szeliga, W. M.; Santillan, V. M.; Scrivner, C.; Senko, J.; Larsen, D.

    2017-12-01

    Thousands of real-time telemetered GNSS stations operate throughout the circum-Pacific that may be used for rapid earthquake characterization and estimation of local tsunami excitation. We report on the development of a GNSS-based finite-fault inversion system that continuously estimates slip using real-time GNSS position streams from the Cascadia subduction zone and which is being expanded throughout the circum-Pacific. The system uses 1 Hz precise point position streams computed in the ITRF14 reference frame using clock and satellite orbit corrections from the IGS. The software is implemented as seven independent modules that filter time series using Kalman filters, trigger and estimate coseismic offsets, invert for slip using a non-negative least squares method developed by Lawson and Hanson (1974) and elastic half-space Green's Functions developed by Okada (1985), smooth the results temporally and spatially, and write the resulting streams of time-dependent slip to a RabbitMQ messaging server for use by downstream modules such as tsunami excitation modules. Additional fault models can be easily added to the system for other circum-Pacific subduction zones as additional real-time GNSS data become available. The system is currently being tested using data from well-recorded earthquakes including the 2011 Tohoku earthquake, the 2010 Maule earthquake, the 2015 Illapel earthquake, the 2003 Tokachi-oki earthquake, the 2014 Iquique earthquake, the 2010 Mentawai earthquake, the 2016 Kaikoura earthquake, the 2016 Ecuador earthquake, the 2015 Gorkha earthquake, and others. Test data will be fed to the system and the resultant earthquake characterizations will be compared with published earthquake parameters. Seismic events will be assumed to occur on major faults, so, for example, only the San Andreas fault will be considered in Southern California, while the hundreds of other faults in the region will be ignored. Rake will be constrained along each subfault to be

  19. Advances in Time Estimation Methods for Molecular Data.

    Science.gov (United States)

    Kumar, Sudhir; Hedges, S Blair

    2016-04-01

    Molecular dating has become central to placing a temporal dimension on the tree of life. Methods for estimating divergence times have been developed for over 50 years, beginning with the proposal of molecular clock in 1962. We categorize the chronological development of these methods into four generations based on the timing of their origin. In the first generation approaches (1960s-1980s), a strict molecular clock was assumed to date divergences. In the second generation approaches (1990s), the equality of evolutionary rates between species was first tested and then a strict molecular clock applied to estimate divergence times. The third generation approaches (since ∼2000) account for differences in evolutionary rates across the tree by using a statistical model, obviating the need to assume a clock or to test the equality of evolutionary rates among species. Bayesian methods in the third generation require a specific or uniform prior on the speciation-process and enable the inclusion of uncertainty in clock calibrations. The fourth generation approaches (since 2012) allow rates to vary from branch to branch, but do not need prior selection of a statistical model to describe the rate variation or the specification of speciation model. With high accuracy, comparable to Bayesian approaches, and speeds that are orders of magnitude faster, fourth generation methods are able to produce reliable timetrees of thousands of species using genome scale data. We found that early time estimates from second generation studies are similar to those of third and fourth generation studies, indicating that methodological advances have not fundamentally altered the timetree of life, but rather have facilitated time estimation by enabling the inclusion of more species. Nonetheless, we feel an urgent need for testing the accuracy and precision of third and fourth generation methods, including their robustness to misspecification of priors in the analysis of large phylogenies and data

  20. Near-real-time and scenario earthquake loss estimates for Mexico

    Science.gov (United States)

    Wyss, M.; Zuñiga, R.

    2017-12-01

    The large earthquakes of 8 September 2017, M8.1, and 19 September 2017, M7.1 have focused attention on the dangers of Mexican seismicity. The near-real-time alerts by QLARM estimated 10 to 300 fatalities and 0 to 200 fatalities, respectively. At the time of this submission the reported death tolls are 96 and 226, respectively. These alerts were issued within 96 and 57 minutes of the occurrence times. For the M8.1 earthquake the losses due to a line model could be calculated. The line with length L=110 km extended from the initial epicenter to the NE, where the USGS had reported aftershocks. On September 19, no aftershocks were available in near-real-time, so a point source had to be used for the quick calculation of likely casualties. In both cases, the casualties were at least an order of magnitude smaller than what they could have been because on 8 September the source was relatively far offshore and on 19 September the hypocenter was relatively deep. The largest historic earthquake in Mexico occurred on 28 March 1787 and likely had a rupture length of 450 km and M8.6. Based on this event, and after verifying our tool for Mexico, we estimated the order of magnitude of a disaster, given the current population, in a maximum credible earthquake along the Pacific coast. In the countryside along the coast we expect approximately 27,000 fatalities and 480,000 injured. In the special case of Mexico City the casualties in a worst possible earthquake along the Pacific plate boundary would likely be counted as five digit numbers. The large agglomerate of the capital with its lake bed soil attracts most attention. Nevertheless, one should pay attention to the fact that the poor, rural segment of society, living in buildings of weak resistance to shaking, are likely to sustain a mortality rate about 20% larger than the population in cities on average soil.

  1. The fossilized birth–death process for coherent calibration of divergence-time estimates

    Science.gov (United States)

    Heath, Tracy A.; Huelsenbeck, John P.; Stadler, Tanja

    2014-01-01

    Time-calibrated species phylogenies are critical for addressing a wide range of questions in evolutionary biology, such as those that elucidate historical biogeography or uncover patterns of coevolution and diversification. Because molecular sequence data are not informative on absolute time, external data—most commonly, fossil age estimates—are required to calibrate estimates of species divergence dates. For Bayesian divergence time methods, the common practice for calibration using fossil information involves placing arbitrarily chosen parametric distributions on internal nodes, often disregarding most of the information in the fossil record. We introduce the “fossilized birth–death” (FBD) process—a model for calibrating divergence time estimates in a Bayesian framework, explicitly acknowledging that extant species and fossils are part of the same macroevolutionary process. Under this model, absolute node age estimates are calibrated by a single diversification model and arbitrary calibration densities are not necessary. Moreover, the FBD model allows for inclusion of all available fossils. We performed analyses of simulated data and show that node age estimation under the FBD model results in robust and accurate estimates of species divergence times with realistic measures of statistical uncertainty, overcoming major limitations of standard divergence time estimation methods. We used this model to estimate the speciation times for a dataset composed of all living bears, indicating that the genus Ursus diversified in the Late Miocene to Middle Pliocene. PMID:25009181

  2. Time-to-contact estimation modulated by implied friction.

    Science.gov (United States)

    Yamada, Yuki; Sasaki, Kyoshiro; Miura, Kayo

    2014-01-01

    The present study demonstrated that friction cues for target motion affect time-to-contact (TTC) estimation. A circular target moved in a linear path with a constant velocity and was gradually occluded by a static rectangle. The target moved with forward and backward spins or without spin. Observers were asked to respond at the time when the moving target appeared to pass the occluder. The results showed that TTC was significantly longer in the backward spin condition than in the forward and without-spin conditions. Moreover, similar results were obtained when a sound was used to imply friction. Our findings indicate that the observer's experiential knowledge of motion coupled with friction intuitively modulated their TTC estimation.

  3. Bayesian Nonparametric Model for Estimating Multistate Travel Time Distribution

    Directory of Open Access Journals (Sweden)

    Emmanuel Kidando

    2017-01-01

    Full Text Available Multistate models, that is, models with more than two distributions, are preferred over single-state probability models in modeling the distribution of travel time. Literature review indicated that the finite multistate modeling of travel time using lognormal distribution is superior to other probability functions. In this study, we extend the finite multistate lognormal model of estimating the travel time distribution to unbounded lognormal distribution. In particular, a nonparametric Dirichlet Process Mixture Model (DPMM with stick-breaking process representation was used. The strength of the DPMM is that it can choose the number of components dynamically as part of the algorithm during parameter estimation. To reduce computational complexity, the modeling process was limited to a maximum of six components. Then, the Markov Chain Monte Carlo (MCMC sampling technique was employed to estimate the parameters’ posterior distribution. Speed data from nine links of a freeway corridor, aggregated on a 5-minute basis, were used to calculate the corridor travel time. The results demonstrated that this model offers significant flexibility in modeling to account for complex mixture distributions of the travel time without specifying the number of components. The DPMM modeling further revealed that freeway travel time is characterized by multistate or single-state models depending on the inclusion of onset and offset of congestion periods.

  4. Consumption estimation in system design for real time embedded applications; Estimation de la consommation dans la conception systeme des applications embarquees temps reel

    Energy Technology Data Exchange (ETDEWEB)

    Laurent, J.

    2002-12-15

    Today, power and energy consumption have become, as time and area, an important constraint when you design a system. Indeed modern applications use more and more processing and memory resources so these lead a significant increase of consumption. Furthermore, embedded software impact is preponderant in real time system so the code optimisation has a great impact onto the consumption constraint. Several research teams have already developed estimation methodologies for processor but almost are at instruction level (ILPA). With this kind of method you have to measure the consumption of each instruction of the instruction set and also the inter-instruction consumption overhead. For complex architecture, this kind of methodology is not adapted due to the prohibitive number of consumption measures. So the characterisation time of this kind of architecture is too important furthermore with this method is very difficult to take into account the external environment. For actual architecture another method is needed to reduce the characterisation time while preserving the accuracy. The reduction of the characterisation time have to be realized by increasing the abstraction level. So, we propose here a new approach based on a functional and architectural analysis of the target in consumption point of view (FLPA). Our methodology has two steps: the first one is a modeling step and the second is estimation step. (author)

  5. Nonlinear estimation of ring-down time for a Fabry-Perot optical cavity.

    Science.gov (United States)

    Kallapur, Abhijit G; Boyson, Toby K; Petersen, Ian R; Harb, Charles C

    2011-03-28

    This paper discusses the application of a discrete-time extended Kalman filter (EKF) to the problem of estimating the decay time constant for a Fabry-Perot optical cavity for cavity ring-down spectroscopy (CRDS). The data for the estimation process is obtained from a CRDS experimental setup in terms of the light intensity at the output of the cavity. The cavity is held in lock with the input laser frequency by controlling the distance between the mirrors within the cavity by means of a proportional-integral (PI) controller. The cavity is purged with nitrogen and placed under vacuum before chopping the incident light at 25 KHz and recording the light intensity at its output. In spite of beginning the EKF estimation process with uncertainties in the initial value for the decay time constant, its estimates converge well within a small neighborhood of the expected value for the decay time constant of the cavity within a few ring-down cycles. Also, the EKF estimation results for the decay time constant are compared to those obtained using the Levenberg-Marquardt estimation scheme.

  6. Canadian Estimate of Bird Mortality Due to Collisions and Direct Habitat Loss Associated with Wind Turbine Developments

    Directory of Open Access Journals (Sweden)

    J. Ryan. Zimmerling

    2013-12-01

    Full Text Available We estimated impacts on birds from the development and operation of wind turbines in Canada considering both mortality due to collisions and loss of nesting habitat. We estimated collision mortality using data from carcass searches for 43 wind farms, incorporating correction factors for scavenger removal, searcher efficiency, and carcasses that fell beyond the area searched. On average, 8.2 ± 1.4 birds (95% C.I. were killed per turbine per year at these sites, although the numbers at individual wind farms varied from 0 - 26.9 birds per turbine per year. Based on 2955 installed turbines (the number installed in Canada by December 2011, an estimated 23,300 birds (95% C.I. 20,000 - 28,300 would be killed from collisions with turbines each year. We estimated direct habitat loss based on data from 32 wind farms in Canada. On average, total habitat loss per turbine was 1.23 ha, which corresponds to an estimated total habitat loss due to wind farms nationwide of 3635 ha. Based on published estimates of nest density, this could represent habitat for ~5700 nests of all species. Assuming nearby habitats are saturated, and 2 adults displaced per nest site, effects of direct habitat loss are less than that of direct mortality. Installed wind capacity is growing rapidly, and is predicted to increase more than 10-fold over the next 10-15 years, which could lead to direct mortality of approximately 233,000 birds / year, and displacement of 57,000 pairs. Despite concerns about the impacts of biased correction factors on the accuracy of mortality estimates, these values are likely much lower than those from collisions with some other anthropogenic sources such as windows, vehicles, or towers, or habitat loss due to many other forms of development. Species composition data suggest that < 0.2% of the population of any species is currently affected by mortality or displacement from wind turbine development. Therefore, population level impacts are unlikely

  7. Real-time estimation of TP load in a Mississippi Delta Stream using a dynamic data driven application system

    Science.gov (United States)

    Ying Ouyang; Theodor D. Leininger; Jeff Hatten

    2013-01-01

    Elevated phosphorus (P) in surface waters can cause eutrophication of aquatic ecosystems and can impair water for drinking, industry, agriculture, and recreation. Currently, no effort has been devoted to estimating real-time variation and load of total P (TP) in surface waters due to the lack of suitable and/or cost-effective wireless sensors. However, when considering...

  8. Real-time airborne gamma-ray background estimation using NASVD with MLE and radiation transport for calibration

    Energy Technology Data Exchange (ETDEWEB)

    Kulisek, J.A., E-mail: Jonathan.Kulisek@pnnl.gov; Schweppe, J.E.; Stave, S.C.; Bernacki, B.E.; Jordan, D.V.; Stewart, T.N.; Seifert, C.E.; Kernan, W.J.

    2015-06-01

    Helicopter-mounted gamma-ray detectors can provide law enforcement officials the means to quickly and accurately detect, identify, and locate radiological threats over a wide geographical area. The ability to accurately distinguish radiological threat-generated gamma-ray signatures from background gamma radiation in real time is essential in order to realize this potential. This problem is non-trivial, especially in urban environments for which the background may change very rapidly during flight. This exacerbates the challenge of estimating background due to the poor counting statistics inherent in real-time airborne gamma-ray spectroscopy measurements. To address this challenge, we have developed a new technique for real-time estimation of background gamma radiation from aerial measurements without the need for human analyst intervention. The method can be calibrated using radiation transport simulations along with data from previous flights over areas for which the isotopic composition need not be known. Over the examined measured and simulated data sets, the method generated accurate background estimates even in the presence of a strong, {sup 60}Co source. The potential to track large and abrupt changes in background spectral shape and magnitude was demonstrated. The method can be implemented fairly easily in most modern computing languages and environments.

  9. Splitting Travel Time Based on AFC Data: Estimating Walking, Waiting, Transfer, and In-Vehicle Travel Times in Metro System

    Directory of Open Access Journals (Sweden)

    Yong-Sheng Zhang

    2015-01-01

    Full Text Available The walking, waiting, transfer, and delayed in-vehicle travel times mainly contribute to route’s travel time reliability in the metro system. The automatic fare collection (AFC system provides huge amounts of smart card records which can be used to estimate all these times distributions. A new estimation model based on Bayesian inference formulation is proposed in this paper by integrating the probability measurement of the OD pair with only one effective route, in which all kinds of times follow the truncated normal distributions. Then, Markov Chain Monte Carlo method is designed to estimate all parameters endogenously. Finally, based on AFC data in Guangzhou Metro, the estimations show that all parameters can be estimated endogenously and identifiably. Meanwhile, the truncated property of the travel time is significant and the threshold tested by the surveyed data is reliable. Furthermore, the superiority of the proposed model over the existing model in estimation and forecasting accuracy is also demonstrated.

  10. Estimated number of preterm births and low birth weight children born in the United States due to maternal binge drinking.

    Science.gov (United States)

    Truong, Khoa D; Reifsnider, Odette S; Mayorga, Maria E; Spitler, Hugh

    2013-05-01

    The objective of this study was to estimate the aggregate burden of maternal binge drinking on preterm birth (PTB) and low birth weight (LBW) across American sociodemographic groups in 2008. To estimate the aggregate burden of maternal binge drinking on preterm birth (PTB) and low birth weight (LBW) across American sociodemographic groups in 2008. A simulation model was developed to estimate the number of PTB and LBW cases due to maternal binge drinking. Data inputs for the model included number of births and rates of preterm and LBW from the National Center for Health Statistics; female population by childbearing age groups from the U.S. Census; increased relative risks of preterm and LBW deliveries due to maternal binge drinking extracted from the literature; and adjusted prevalence of binge drinking among pregnant women estimated in a multivariate logistic regression model using Behavioral Risk Factor Surveillance System survey. The most conservative estimates attributed maternal binge drinking to 8,701 (95% CI: 7,804-9,598) PTBs (1.75% of all PTBs) and 5,627 (95% CI 5,121-6,133) LBW deliveries in 2008, with 3,708 (95% CI: 3,375-4,041) cases of both PTB and LBW. The estimated rate of PTB due to maternal binge drinking was 1.57% among all PTBs to White women, 0.69% among Black women, 3.31% among Hispanic women, and 2.35% among other races. Compared to other age groups, women ages 40-44 had the highest adjusted binge drinking rate and highest PTB rate due to maternal binge drinking (4.33%). Maternal binge drinking contributed significantly to PTB and LBW differentially across sociodemographic groups.

  11. Polynomial Phase Estimation Based on Adaptive Short-Time Fourier Transform.

    Science.gov (United States)

    Jing, Fulong; Zhang, Chunjie; Si, Weijian; Wang, Yu; Jiao, Shuhong

    2018-02-13

    Polynomial phase signals (PPSs) have numerous applications in many fields including radar, sonar, geophysics, and radio communication systems. Therefore, estimation of PPS coefficients is very important. In this paper, a novel approach for PPS parameters estimation based on adaptive short-time Fourier transform (ASTFT), called the PPS-ASTFT estimator, is proposed. Using the PPS-ASTFT estimator, both one-dimensional and multi-dimensional searches and error propagation problems, which widely exist in PPSs field, are avoided. In the proposed algorithm, the instantaneous frequency (IF) is estimated by S-transform (ST), which can preserve information on signal phase and provide a variable resolution similar to the wavelet transform (WT). The width of the ASTFT analysis window is equal to the local stationary length, which is measured by the instantaneous frequency gradient (IFG). The IFG is calculated by the principal component analysis (PCA), which is robust to the noise. Moreover, to improve estimation accuracy, a refinement strategy is presented to estimate signal parameters. Since the PPS-ASTFT avoids parameter search, the proposed algorithm can be computed in a reasonable amount of time. The estimation performance, computational cost, and implementation of the PPS-ASTFT are also analyzed. The conducted numerical simulations support our theoretical results and demonstrate an excellent statistical performance of the proposed algorithm.

  12. Estimated value of insurance premium due to Citarum River flood by using Bayesian method

    Science.gov (United States)

    Sukono; Aisah, I.; Tampubolon, Y. R. H.; Napitupulu, H.; Supian, S.; Subiyanto; Sidi, P.

    2018-03-01

    Citarum river flood in South Bandung, West Java Indonesia, often happens every year. It causes property damage, producing economic loss. The risk of loss can be mitigated by following the flood insurance program. In this paper, we discussed about the estimated value of insurance premiums due to Citarum river flood by Bayesian method. It is assumed that the risk data for flood losses follows the Pareto distribution with the right fat-tail. The estimation of distribution model parameters is done by using Bayesian method. First, parameter estimation is done with assumption that prior comes from Gamma distribution family, while observation data follow Pareto distribution. Second, flood loss data is simulated based on the probability of damage in each flood affected area. The result of the analysis shows that the estimated premium value of insurance based on pure premium principle is as follows: for the loss value of IDR 629.65 million of premium IDR 338.63 million; for a loss of IDR 584.30 million of its premium IDR 314.24 million; and the loss value of IDR 574.53 million of its premium IDR 308.95 million. The premium value estimator can be used as neither a reference in the decision of reasonable premium determination, so as not to incriminate the insured, nor it result in loss of the insurer.

  13. Online Synchrophasor-Based Dynamic State Estimation using Real-Time Digital Simulator

    DEFF Research Database (Denmark)

    Khazraj, Hesam; Adewole, Adeyemi Charles; Udaya, Annakkage

    2018-01-01

    Dynamic state estimation is a very important control center application used in the dynamic monitoring of state variables. This paper presents and validates a time-synchronized phasor measurement unit (PMU)-based for dynamic state estimation by unscented Kalman filter (UKF) method using the real-...... using the RTDS (real-time digital simulator). The dynamic state variables of multi-machine systems are monitored and measured for the study on the transient behavior of power systems.......Dynamic state estimation is a very important control center application used in the dynamic monitoring of state variables. This paper presents and validates a time-synchronized phasor measurement unit (PMU)-based for dynamic state estimation by unscented Kalman filter (UKF) method using the real......-time digital simulator (RTDS). The dynamic state variables of the system are the rotor angle and speed of the generators. The performance of the UKF method is tested with PMU measurements as inputs using the IEEE 14-bus test system. This test system was modeled in the RSCAD software and tested in real time...

  14. A theory of timing in scintillation counters based on maximum likelihood estimation

    International Nuclear Information System (INIS)

    Tomitani, Takehiro

    1982-01-01

    A theory of timing in scintillation counters based on the maximum likelihood estimation is presented. An optimum filter that minimizes the variance of timing is described. A simple formula to estimate the variance of timing is presented as a function of photoelectron number, scintillation decay constant and the single electron transit time spread in the photomultiplier. The present method was compared with the theory by E. Gatti and V. Svelto. The proposed method was applied to two simple models and rough estimations of potential time resolution of several scintillators are given. The proposed method is applicable to the timing in Cerenkov counters and semiconductor detectors as well. (author)

  15. Estimation of integrity of cast-iron cask against impact due to free drop test, (1)

    International Nuclear Information System (INIS)

    Itoh, Chihiro

    1988-01-01

    Ductile cast iron is examined to use for shipping and storage cask from a economic point of view. However, ductile cast iron is considered to be a brittle material in general. Therefore, it is very important to estimate the integrity of cast iron cask against brittle failure due to impact load at 9 m drop test and 1 m derop test on to pin. So, the F.E.M. analysis which takes nonlinearity of materials into account and the estimation against brittle failure by the method which is proposed in this report were carried out. From the analysis, it is made clear that critical flaw depth (the minimum depth to initiate the brittle failure) is 21.1 mm and 13.1 mm in the case of 9 m drop test and 1 m drop test on to pin respectively. These flaw depth can be detected by ultrasonic test. Then, the cask is assured against brittle failure due to impact load at 9 m drop test and 1 m drop test on to pin. (author)

  16. Soft sensor for real-time cement fineness estimation.

    Science.gov (United States)

    Stanišić, Darko; Jorgovanović, Nikola; Popov, Nikola; Čongradac, Velimir

    2015-03-01

    This paper describes the design and implementation of soft sensors to estimate cement fineness. Soft sensors are mathematical models that use available data to provide real-time information on process variables when the information, for whatever reason, is not available by direct measurement. In this application, soft sensors are used to provide information on process variable normally provided by off-line laboratory tests performed at large time intervals. Cement fineness is one of the crucial parameters that define the quality of produced cement. Providing real-time information on cement fineness using soft sensors can overcome limitations and problems that originate from a lack of information between two laboratory tests. The model inputs were selected from candidate process variables using an information theoretic approach. Models based on multi-layer perceptrons were developed, and their ability to estimate cement fineness of laboratory samples was analyzed. Models that had the best performance, and capacity to adopt changes in the cement grinding circuit were selected to implement soft sensors. Soft sensors were tested using data from a continuous cement production to demonstrate their use in real-time fineness estimation. Their performance was highly satisfactory, and the sensors proved to be capable of providing valuable information on cement grinding circuit performance. After successful off-line tests, soft sensors were implemented and installed in the control room of a cement factory. Results on the site confirm results obtained by tests conducted during soft sensor development. Copyright © 2014 ISA. Published by Elsevier Ltd. All rights reserved.

  17. Estimation of time averages from irregularly spaced observations - With application to coastal zone color scanner estimates of chlorophyll concentration

    Science.gov (United States)

    Chelton, Dudley B.; Schlax, Michael G.

    1991-01-01

    The sampling error of an arbitrary linear estimate of a time-averaged quantity constructed from a time series of irregularly spaced observations at a fixed located is quantified through a formalism. The method is applied to satellite observations of chlorophyll from the coastal zone color scanner. The two specific linear estimates under consideration are the composite average formed from the simple average of all observations within the averaging period and the optimal estimate formed by minimizing the mean squared error of the temporal average based on all the observations in the time series. The resulting suboptimal estimates are shown to be more accurate than composite averages. Suboptimal estimates are also found to be nearly as accurate as optimal estimates using the correct signal and measurement error variances and correlation functions for realistic ranges of these parameters, which makes it a viable practical alternative to the composite average method generally employed at present.

  18. Estimating time to pregnancy from current durations in a cross-sectional sample

    DEFF Research Database (Denmark)

    Keiding, Niels; Kvist, Kajsa; Hartvig, Helle

    2002-01-01

    A new design for estimating the distribution of time to pregnancy is proposed and investigated. The design is based on recording current durations in a cross-sectional sample of women, leading to statistical problems similar to estimating renewal time distributions from backward recurrence times....

  19. Nonlinear systems time-varying parameter estimation: Application to induction motors

    Energy Technology Data Exchange (ETDEWEB)

    Kenne, Godpromesse [Laboratoire d' Automatique et d' Informatique Appliquee (LAIA), Departement de Genie Electrique, IUT FOTSO Victor, Universite de Dschang, B.P. 134 Bandjoun (Cameroon); Ahmed-Ali, Tarek [Ecole Nationale Superieure des Ingenieurs des Etudes et Techniques d' Armement (ENSIETA), 2 Rue Francois Verny, 29806 Brest Cedex 9 (France); Lamnabhi-Lagarrigue, F. [Laboratoire des Signaux et Systemes (L2S), C.N.R.S-SUPELEC, Universite Paris XI, 3 Rue Joliot Curie, 91192 Gif-sur-Yvette (France); Arzande, Amir [Departement Energie, Ecole Superieure d' Electricite-SUPELEC, 3 Rue Joliot Curie, 91192 Gif-sur-Yvette (France)

    2008-11-15

    In this paper, an algorithm for time-varying parameter estimation for a large class of nonlinear systems is presented. The proof of the convergence of the estimates to their true values is achieved using Lyapunov theories and does not require that the classical persistent excitation condition be satisfied by the input signal. Since the induction motor (IM) is widely used in several industrial sectors, the algorithm developed is potentially useful for adjusting the controller parameters of variable speed drives. The method proposed is simple and easily implementable in real-time. The application of this approach to on-line estimation of the rotor resistance of IM shows a rapidly converging estimate in spite of measurement noise, discretization effects, parameter uncertainties (e.g. inaccuracies on motor inductance values) and modeling inaccuracies. The robustness analysis for this IM application also revealed that the proposed scheme is insensitive to the stator resistance variations within a wide range. The merits of the proposed algorithm in the case of on-line time-varying rotor resistance estimation are demonstrated via experimental results in various operating conditions of the induction motor. The experimental results obtained demonstrate that the application of the proposed algorithm to update on-line the parameters of an adaptive controller (e.g. IM and synchronous machines adaptive control) can improve the efficiency of the industrial process. The other interesting features of the proposed method include fault detection/estimation and adaptive control of IM and synchronous machines. (author)

  20. time of arrival 3-d position estimation using minimum ads-b receiver ...

    African Journals Online (AJOL)

    HOD

    The location from which a signal is transmitted can be estimated using the time it takes to be detected at a receiver. The difference between transmission time and the detection time is known as time of arrival (TOA). In this work, an algorithm for 3-dimensional (3-D) position estimation (PE) of an emitter using the minimum ...

  1. The generalized correlation method for estimation of time delay in power plants

    International Nuclear Information System (INIS)

    Kostic, Lj.

    1981-01-01

    The generalized correlation estimation is developed for determining time delay between signals received at two spatially separated sensors in the presence of uncorrelated noise in a power plant. This estimator can be realized as a pair of receiver prefilters followed by a cross correlator. The time argument at which the correlator achieves a maximum is the delay estimate. (author)

  2. On estimation of time-dependent attributable fraction from population-based case-control studies.

    Science.gov (United States)

    Zhao, Wei; Chen, Ying Qing; Hsu, Li

    2017-09-01

    Population attributable fraction (PAF) is widely used to quantify the disease burden associated with a modifiable exposure in a population. It has been extended to a time-varying measure that provides additional information on when and how the exposure's impact varies over time for cohort studies. However, there is no estimation procedure for PAF using data that are collected from population-based case-control studies, which, because of time and cost efficiency, are commonly used for studying genetic and environmental risk factors of disease incidences. In this article, we show that time-varying PAF is identifiable from a case-control study and develop a novel estimator of PAF. Our estimator combines odds ratio estimates from logistic regression models and density estimates of the risk factor distribution conditional on failure times in cases from a kernel smoother. The proposed estimator is shown to be consistent and asymptotically normal with asymptotic variance that can be estimated empirically from the data. Simulation studies demonstrate that the proposed estimator performs well in finite sample sizes. Finally, the method is illustrated by a population-based case-control study of colorectal cancer. © 2017, The International Biometric Society.

  3. REAL TIME SPEED ESTIMATION FROM MONOCULAR VIDEO

    Directory of Open Access Journals (Sweden)

    M. S. Temiz

    2012-07-01

    Full Text Available In this paper, detailed studies have been performed for developing a real time system to be used for surveillance of the traffic flow by using monocular video cameras to find speeds of the vehicles for secure travelling are presented. We assume that the studied road segment is planar and straight, the camera is tilted downward a bridge and the length of one line segment in the image is known. In order to estimate the speed of a moving vehicle from a video camera, rectification of video images is performed to eliminate the perspective effects and then the interest region namely the ROI is determined for tracking the vehicles. Velocity vectors of a sufficient number of reference points are identified on the image of the vehicle from each video frame. For this purpose sufficient number of points from the vehicle is selected, and these points must be accurately tracked on at least two successive video frames. In the second step, by using the displacement vectors of the tracked points and passed time, the velocity vectors of those points are computed. Computed velocity vectors are defined in the video image coordinate system and displacement vectors are measured by the means of pixel units. Then the magnitudes of the computed vectors in the image space are transformed to the object space to find the absolute values of these magnitudes. The accuracy of the estimated speed is approximately ±1 – 2 km/h. In order to solve the real time speed estimation problem, the authors have written a software system in C++ programming language. This software system has been used for all of the computations and test applications.

  4. Real-time gaze estimation via pupil center tracking

    Directory of Open Access Journals (Sweden)

    Cazzato Dario

    2018-02-01

    Full Text Available Automatic gaze estimation not based on commercial and expensive eye tracking hardware solutions can enable several applications in the fields of human computer interaction (HCI and human behavior analysis. It is therefore not surprising that several related techniques and methods have been investigated in recent years. However, very few camera-based systems proposed in the literature are both real-time and robust. In this work, we propose a real-time user-calibration-free gaze estimation system that does not need person-dependent calibration, can deal with illumination changes and head pose variations, and can work with a wide range of distances from the camera. Our solution is based on a 3-D appearance-based method that processes the images from a built-in laptop camera. Real-time performance is obtained by combining head pose information with geometrical eye features to train a machine learning algorithm. Our method has been validated on a data set of images of users in natural environments, and shows promising results. The possibility of a real-time implementation, combined with the good quality of gaze tracking, make this system suitable for various HCI applications.

  5. Method to estimate the internal stresses due to moisture in wood using transmission properties of microwaves

    International Nuclear Information System (INIS)

    Takemura, T.; Taniguchi, T.

    2004-01-01

    The purpose of this paper is to offer a new method for detecting stress in wood due to moisture along the lines of a theory reported previously. According to the theory, the stress in wood could be estimated from the moisture content of the wood and the power voltage of a microwave moisture meter (i.e., attenuation of the projected microwave). This seems to suggest a possibility of utilizing microwaves in the field of stress detection. To develop such an idea, the stress formulas were initially modified to the form of an uni-variable function of power voltage, and the application method of the formulas to detection was tried. Finally, these results were applied to the data of sugi (Cryptomeria japonica) lumber in the previous experiment. The estimated strains showed fairly good agreement with those observed. It could be concluded from this study that the proposed method might be available for detecting stress in wood due to moisture

  6. Dose estimation from food intake due to the Fukushima Daiichi nuclear power plant accident

    International Nuclear Information System (INIS)

    Yamaguchi, Ichiro; Terada, Hiroshi; Kunugita, Naoki; Takahashi, Kunihiko

    2013-01-01

    Since the Fukushima Daiichi nuclear power plant accident, concerns have arisen about the radiation safety of food raised at home and abroad. Therefore, many measures have been taken to address this. To evaluate the effectiveness of these measures, dose estimation due to food consumption has been attempted by various methods. In this paper, we show the results of dose estimation based on the monitoring data of radioactive materials in food published by the Ministry of Health, Labour and Welfare. The Radioactive Material Response Working Group in the Food Sanitation Subcommittee of the Pharmaceutical Affairs and Food Sanitation Council reported such dose estimation results on October 31, 2011 using monitoring data from immediately after the accident through September, 2011. Our results presented in this paper were the effective dose and thyroid equivalent dose integrated up to December 2012 from immediately after the accident. The estimated results of committed effective dose by age group derived from the radioiodine and radiocesium in food after the Fukushima Daiichi nuclear power plant accident showed the highest median value (0.19 mSv) in children 13-18 years of age. The highest 95% tile value, 0.33 mSv, was shown in the 1-6 years age range. These dose estimations from food can be useful for evaluation of radiation risk for individuals or populations and for radiation protection measures. It would also be helpful for the study of risk management of food in the future. (author)

  7. Efficient Implementation of a Symbol Timing Estimator for Broadband PLC.

    Science.gov (United States)

    Nombela, Francisco; García, Enrique; Mateos, Raúl; Hernández, Álvaro

    2015-08-21

    Broadband Power Line Communications (PLC) have taken advantage of the research advances in multi-carrier modulations to mitigate frequency selective fading, and their adoption opens up a myriad of applications in the field of sensory and automation systems, multimedia connectivity or smart spaces. Nonetheless, the use of these multi-carrier modulations, such as Wavelet-OFDM, requires a highly accurate symbol timing estimation for reliably recovering of transmitted data. Furthermore, the PLC channel presents some particularities that prevent the direct use of previous synchronization algorithms proposed in wireless communication systems. Therefore more research effort should be involved in the design and implementation of novel and robust synchronization algorithms for PLC, thus enabling real-time synchronization. This paper proposes a symbol timing estimator for broadband PLC based on cross-correlation with multilevel complementary sequences or Zadoff-Chu sequences and its efficient implementation in a FPGA; the obtained results show a 90% of success rate in symbol timing estimation for a certain PLC channel model and a reduced resource consumption for its implementation in a Xilinx Kyntex FPGA.

  8. Efficient Implementation of a Symbol Timing Estimator for Broadband PLC

    Directory of Open Access Journals (Sweden)

    Francisco Nombela

    2015-08-01

    Full Text Available Broadband Power Line Communications (PLC have taken advantage of the research advances in multi-carrier modulations to mitigate frequency selective fading, and their adoption opens up a myriad of applications in the field of sensory and automation systems, multimedia connectivity or smart spaces. Nonetheless, the use of these multi-carrier modulations, such as Wavelet-OFDM, requires a highly accurate symbol timing estimation for reliably recovering of transmitted data. Furthermore, the PLC channel presents some particularities that prevent the direct use of previous synchronization algorithms proposed in wireless communication systems. Therefore more research effort should be involved in the design and implementation of novel and robust synchronization algorithms for PLC, thus enabling real-time synchronization. This paper proposes a symbol timing estimator for broadband PLC based on cross-correlation with multilevel complementary sequences or Zadoff-Chu sequences and its efficient implementation in a FPGA; the obtained results show a 90% of success rate in symbol timing estimation for a certain PLC channel model and a reduced resource consumption for its implementation in a Xilinx Kyntex FPGA.

  9. Ab initio quantum-enhanced optical phase estimation using real-time feedback control

    DEFF Research Database (Denmark)

    Berni, Adriano; Gehring, Tobias; Nielsen, Bo Melholt

    2015-01-01

    of a quantum-enhanced and fully deterministic ab initio phase estimation protocol based on real-time feedback control. Using robust squeezed states of light combined with a real-time Bayesian adaptive estimation algorithm, we demonstrate deterministic phase estimation with a precision beyond the quantum shot...... noise limit. The demonstrated protocol opens up new opportunities for quantum microscopy, quantum metrology and quantum information processing....

  10. Time-dependent inversion of surface subsidence due to dynamic reservoir compaction

    NARCIS (Netherlands)

    Muntendam-Bos, A.G.; Kroon, I.C.; Fokker, P.A.

    2008-01-01

    We introduce a novel, time-dependent inversion scheme for resolving temporal reservoir pressure drop from surface subsidence observations (from leveling or GPS data, InSAR, tiltmeter monitoring) in a single procedure. The theory is able to accommodate both the absence of surface subsidence estimates

  11. Estimating the level of dynamical noise in time series by using fractal dimensions

    Energy Technology Data Exchange (ETDEWEB)

    Sase, Takumi, E-mail: sase@sat.t.u-tokyo.ac.jp [Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 153-8505 (Japan); Ramírez, Jonatán Peña [CONACYT Research Fellow, Center for Scientific Research and Higher Education at Ensenada (CICESE), Carretera Ensenada-Tijuana No. 3918, Zona Playitas, C.P. 22860, Ensenada, Baja California (Mexico); Kitajo, Keiichi [BSI-Toyota Collaboration Center, RIKEN Brain Science Institute, Wako, Saitama 351-0198 (Japan); Aihara, Kazuyuki; Hirata, Yoshito [Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 153-8505 (Japan); Institute of Industrial Science, The University of Tokyo, Tokyo 153-8505 (Japan)

    2016-03-11

    We present a method for estimating the dynamical noise level of a ‘short’ time series even if the dynamical system is unknown. The proposed method estimates the level of dynamical noise by calculating the fractal dimensions of the time series. Additionally, the method is applied to EEG data to demonstrate its possible effectiveness as an indicator of temporal changes in the level of dynamical noise. - Highlights: • A dynamical noise level estimator for time series is proposed. • The estimator does not need any information about the dynamics generating the time series. • The estimator is based on a novel definition of time series dimension (TSD). • It is demonstrated that there exists a monotonic relationship between the • TSD and the level of dynamical noise. • We apply the proposed method to human electroencephalographic data.

  12. Estimating the level of dynamical noise in time series by using fractal dimensions

    International Nuclear Information System (INIS)

    Sase, Takumi; Ramírez, Jonatán Peña; Kitajo, Keiichi; Aihara, Kazuyuki; Hirata, Yoshito

    2016-01-01

    We present a method for estimating the dynamical noise level of a ‘short’ time series even if the dynamical system is unknown. The proposed method estimates the level of dynamical noise by calculating the fractal dimensions of the time series. Additionally, the method is applied to EEG data to demonstrate its possible effectiveness as an indicator of temporal changes in the level of dynamical noise. - Highlights: • A dynamical noise level estimator for time series is proposed. • The estimator does not need any information about the dynamics generating the time series. • The estimator is based on a novel definition of time series dimension (TSD). • It is demonstrated that there exists a monotonic relationship between the • TSD and the level of dynamical noise. • We apply the proposed method to human electroencephalographic data.

  13. Improving The Accuracy Of Bluetooth Based Travel Time Estimation Using Low-Level Sensor Data

    DEFF Research Database (Denmark)

    Araghi, Bahar Namaki; Tørholm Christensen, Lars; Krishnan, Rajesh

    2013-01-01

    triggered by a single device. This could lead to location ambiguity and reduced accuracy of travel time estimation. Therefore, the accuracy of travel time estimations by Bluetooth Technology (BT) depends upon how location ambiguity is handled by the estimation method. The issue of multiple detection events...... in the context of travel time estimation by BT has been considered by various researchers. However, treatment of this issue has remained simplistic so far. Most previous studies simply used the first detection event (Enter-Enter) as the best estimate. No systematic analysis for exploring the most accurate method...... of estimating travel time using multiple detection events has been conducted. In this study different aspects of BT detection zone, including size and its impact on the accuracy of travel time estimation, are discussed. Moreover, four alternative methods are applied; namely, Enter-Enter, Leave-Leave, Peak...

  14. Optimal and sub-optimal post-detection timing estimators for PET

    International Nuclear Information System (INIS)

    Hero, A.O.; Antoniadis, N.; Clinthorne, N.; Rogers, W.L.; Hutchins, G.D.

    1990-01-01

    In this paper the authors derive linear and non-linear approximations to the post-detection likelihood function for scintillator interaction time in nuclear particle detection systems. The likelihood function is the optimal statistic for performing detection and estimation of scintillator events and event times. The authors derive the likelihood function approximations from a statistical model for the post-detection waveform which is common in the optical communications literature and takes account of finite detector bandwidth, random gains, and thermal noise. They then present preliminary simulation results for the associated approximate maximum likelihood timing estimators which indicate that significant MSE improvements may be achieved for low post-detection signal-to-noise ratio

  15. Estimating a population cumulative incidence under calendar time trends

    DEFF Research Database (Denmark)

    Hansen, Stefan N; Overgaard, Morten; Andersen, Per K

    2017-01-01

    BACKGROUND: The risk of a disease or psychiatric disorder is frequently measured by the age-specific cumulative incidence. Cumulative incidence estimates are often derived in cohort studies with individuals recruited over calendar time and with the end of follow-up governed by a specific date....... It is common practice to apply the Kaplan-Meier or Aalen-Johansen estimator to the total sample and report either the estimated cumulative incidence curve or just a single point on the curve as a description of the disease risk. METHODS: We argue that, whenever the disease or disorder of interest is influenced...

  16. GNSS global real-time augmentation positioning: Real-time precise satellite clock estimation, prototype system construction and performance analysis

    Science.gov (United States)

    Chen, Liang; Zhao, Qile; Hu, Zhigang; Jiang, Xinyuan; Geng, Changjiang; Ge, Maorong; Shi, Chuang

    2018-01-01

    Lots of ambiguities in un-differenced (UD) model lead to lower calculation efficiency, which isn't appropriate for the high-frequency real-time GNSS clock estimation, like 1 Hz. Mixed differenced model fusing UD pseudo-range and epoch-differenced (ED) phase observations has been introduced into real-time clock estimation. In this contribution, we extend the mixed differenced model for realizing multi-GNSS real-time clock high-frequency updating and a rigorous comparison and analysis on same conditions are performed to achieve the best real-time clock estimation performance taking the efficiency, accuracy, consistency and reliability into consideration. Based on the multi-GNSS real-time data streams provided by multi-GNSS Experiment (MGEX) and Wuhan University, GPS + BeiDou + Galileo global real-time augmentation positioning prototype system is designed and constructed, including real-time precise orbit determination, real-time precise clock estimation, real-time Precise Point Positioning (RT-PPP) and real-time Standard Point Positioning (RT-SPP). The statistical analysis of the 6 h-predicted real-time orbits shows that the root mean square (RMS) in radial direction is about 1-5 cm for GPS, Beidou MEO and Galileo satellites and about 10 cm for Beidou GEO and IGSO satellites. Using the mixed differenced estimation model, the prototype system can realize high-efficient real-time satellite absolute clock estimation with no constant clock-bias and can be used for high-frequency augmentation message updating (such as 1 Hz). The real-time augmentation message signal-in-space ranging error (SISRE), a comprehensive accuracy of orbit and clock and effecting the users' actual positioning performance, is introduced to evaluate and analyze the performance of GPS + BeiDou + Galileo global real-time augmentation positioning system. The statistical analysis of real-time augmentation message SISRE is about 4-7 cm for GPS, whlile 10 cm for Beidou IGSO/MEO, Galileo and about 30 cm

  17. A Modularized Efficient Framework for Non-Markov Time Series Estimation

    Science.gov (United States)

    Schamberg, Gabriel; Ba, Demba; Coleman, Todd P.

    2018-06-01

    We present a compartmentalized approach to finding the maximum a-posteriori (MAP) estimate of a latent time series that obeys a dynamic stochastic model and is observed through noisy measurements. We specifically consider modern signal processing problems with non-Markov signal dynamics (e.g. group sparsity) and/or non-Gaussian measurement models (e.g. point process observation models used in neuroscience). Through the use of auxiliary variables in the MAP estimation problem, we show that a consensus formulation of the alternating direction method of multipliers (ADMM) enables iteratively computing separate estimates based on the likelihood and prior and subsequently "averaging" them in an appropriate sense using a Kalman smoother. As such, this can be applied to a broad class of problem settings and only requires modular adjustments when interchanging various aspects of the statistical model. Under broad log-concavity assumptions, we show that the separate estimation problems are convex optimization problems and that the iterative algorithm converges to the MAP estimate. As such, this framework can capture non-Markov latent time series models and non-Gaussian measurement models. We provide example applications involving (i) group-sparsity priors, within the context of electrophysiologic specrotemporal estimation, and (ii) non-Gaussian measurement models, within the context of dynamic analyses of learning with neural spiking and behavioral observations.

  18. A novel technique for real-time estimation of edge pedestal density gradients via reflectometer time delay data

    Energy Technology Data Exchange (ETDEWEB)

    Zeng, L., E-mail: zeng@fusion.gat.com; Doyle, E. J.; Rhodes, T. L.; Wang, G.; Sung, C.; Peebles, W. A. [Physics and Astronomy Department, University of California, Los Angeles, California 90095 (United States); Bobrek, M. [Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831-6006 (United States)

    2016-11-15

    A new model-based technique for fast estimation of the pedestal electron density gradient has been developed. The technique uses ordinary mode polarization profile reflectometer time delay data and does not require direct profile inversion. Because of its simple data processing, the technique can be readily implemented via a Field-Programmable Gate Array, so as to provide a real-time density gradient estimate, suitable for use in plasma control systems such as envisioned for ITER, and possibly for DIII-D and Experimental Advanced Superconducting Tokamak. The method is based on a simple edge plasma model with a linear pedestal density gradient and low scrape-off-layer density. By measuring reflectometer time delays for three adjacent frequencies, the pedestal density gradient can be estimated analytically via the new approach. Using existing DIII-D profile reflectometer data, the estimated density gradients obtained from the new technique are found to be in good agreement with the actual density gradients for a number of dynamic DIII-D plasma conditions.

  19. Estimates of Imaging Times for Conventional and Synchrotron X-Ray Sources

    CERN Document Server

    Kinney, J

    2003-01-01

    The following notes are to be taken as estimates of the time requirements for imaging NIF targets in three-dimensions with absorption contrast. The estimates ignore target geometry and detector inefficiency, and focus only on the statistical question of detecting compositional (structural) differences between adjacent volume elements in the presence of noise. The basic equations, from the classic reference by Grodzins, consider imaging times in terms of the required number of photons necessary to provide an image with given resolution and noise. The time estimates, therefore, have been based on the calculated x-ray fluxes from the proposed Advanced Light Source (ALS) imaging beamline, and from the calculated flux for a tungsten anode x-ray generator operated in a point focus mode.

  20. Heterogeneous Data Fusion Method to Estimate Travel Time Distributions in Congested Road Networks.

    Science.gov (United States)

    Shi, Chaoyang; Chen, Bi Yu; Lam, William H K; Li, Qingquan

    2017-12-06

    Travel times in congested urban road networks are highly stochastic. Provision of travel time distribution information, including both mean and variance, can be very useful for travelers to make reliable path choice decisions to ensure higher probability of on-time arrival. To this end, a heterogeneous data fusion method is proposed to estimate travel time distributions by fusing heterogeneous data from point and interval detectors. In the proposed method, link travel time distributions are first estimated from point detector observations. The travel time distributions of links without point detectors are imputed based on their spatial correlations with links that have point detectors. The estimated link travel time distributions are then fused with path travel time distributions obtained from the interval detectors using Dempster-Shafer evidence theory. Based on fused path travel time distribution, an optimization technique is further introduced to update link travel time distributions and their spatial correlations. A case study was performed using real-world data from Hong Kong and showed that the proposed method obtained accurate and robust estimations of link and path travel time distributions in congested road networks.

  1. Timing Metrics of Joint Timing and Carrier-Frequency Offset Estimation Algorithms for TDD-based OFDM systems

    NARCIS (Netherlands)

    Hoeksema, F.W.; Srinivasan, R.; Schiphorst, Roelof; Slump, Cornelis H.

    2004-01-01

    In joint timing and carrier offset estimation algorithms for Time Division Duplexing (TDD) OFDM systems, different timing metrics are proposed to determine the beginning of a burst or symbol. In this contribution we investigated the different timing metrics in order to establish their impact on the

  2. Estimating challenge load due to disease outbreaks and other challenges using reproduction records of sows.

    Science.gov (United States)

    Mathur, P K; Herrero-Medrano, J M; Alexandri, P; Knol, E F; ten Napel, J; Rashidi, H; Mulder, H A

    2014-12-01

    A method was developed and tested to estimate challenge load due to disease outbreaks and other challenges in sows using reproduction records. The method was based on reproduction records from a farm with known disease outbreaks. It was assumed that the reduction in weekly reproductive output within a farm is proportional to the magnitude of the challenge. As the challenge increases beyond certain threshold, it is manifested as an outbreak. The reproduction records were divided into 3 datasets. The first dataset called the Training dataset consisted of 57,135 reproduction records from 10,901 sows from 1 farm in Canada with several outbreaks of porcine reproductive and respiratory syndrome (PRRS). The known disease status of sows was regressed on the traits number born alive, number of losses as a combination of still birth and mummified piglets, and number of weaned piglets. The regression coefficients from this analysis were then used as weighting factors for derivation of an index measure called challenge load indicator. These weighting factors were derived with i) a two-step approach using residuals or year-week solutions estimated from a previous step, and ii) a single-step approach using the trait values directly. Two types of models were used for each approach: a logistic regression model and a general additive model. The estimates of challenge load indicator were then compared based on their ability to detect PRRS outbreaks in a Test dataset consisting of records from 65,826 sows from 15 farms in the Netherlands. These farms differed from the Canadian farm with respect to PRRS virus strains, severity and frequency of outbreaks. The single-step approach using a general additive model was best and detected 14 out of the 15 outbreaks. This approach was then further validated using the third dataset consisting of reproduction records of 831,855 sows in 431 farms located in different countries in Europe and America. A total of 41 out of 48 outbreaks detected

  3. Inferring Saving in Training Time From Effect Size Estimates

    National Research Council Canada - National Science Library

    Burright, Burke

    2000-01-01

    .... Students' time saving represents a major potential benefit of using them. This paper fills a methodology gap in estimating the students' timesaving benefit of asynchronous training technologies...

  4. Estimating and Analyzing Savannah Phenology with a Lagged Time Series Model

    DEFF Research Database (Denmark)

    Boke-Olen, Niklas; Lehsten, Veiko; Ardo, Jonas

    2016-01-01

    cycle due to their areal coverage and can have an effect on the food security in regions that depend on subsistence farming. In this study we investigate how soil moisture, mean annual precipitation, and day length control savannah phenology by developing a lagged time series model. The model uses...... climate data for 15 flux tower sites across four continents, and normalized difference vegetation index from satellite to optimize a statistical phenological model. We show that all three variables can be used to estimate savannah phenology on a global scale. However, it was not possible to create...... a simplified savannah model that works equally well for all sites on the global scale without inclusion of more site specific parameters. The simplified model showed no bias towards tree cover or between continents and resulted in a cross-validated r2 of 0.6 and root mean squared error of 0.1. We therefore...

  5. Analytical estimation show low depth-independent water loss due to vapor flux from deep aquifers

    Science.gov (United States)

    Selker, John S.

    2017-06-01

    Recent articles have provided estimates of evaporative flux from water tables in deserts that span 5 orders of magnitude. In this paper, we present an analytical calculation that indicates aquifer vapor flux to be limited to 0.01 mm/yr for sites where there is negligible recharge and the water table is well over 20 m below the surface. This value arises from the geothermal gradient, and therefore, is nearly independent of the actual depth of the aquifer. The value is in agreement with several numerical studies, but is 500 times lower than recently reported experimental values, and 100 times larger than an earlier analytical estimate.

  6. Time-scales for runoff and erosion estimates, with implications for spatial scaling

    Science.gov (United States)

    Kirkby, M. J.; Irvine, B. J.; Dalen, E. N.

    2009-04-01

    Using rainfall data at high temporal resolution, runoff may be estimated for every bucket-tip, or for aggregated hourly or daily periods. Although there is no doubt that finer resolution gives substantially better estimates, many models make use of coarser time steps because these data are more widely available. This paper makes comparisons between runoff estimates based on infiltration measurements used with high resolution rainfall data for SE Spain and theoretical work on improving the time resolution in the PESERA model from daily to hourly values, for areas where these are available. For a small plot at fine temporal scale, runoff responds to bursts of intense rainfall which, for the Guadalentin catchment, typically lasts for about 30 minutes. However, when a larger area is considered, the large and unstructured variability in infiltration capacity produces an aggregate runoff that differs substantially from estimates using average infiltration parameters (in the Green-Ampt equation). When these estimates are compared with estimates based on rainfall for aggregated hourly or daily periods, using a simpler infiltration model, it can be seen that there a substantial scatter, as expected, but that suitable parameterisation can provide reasonable average estimates. Similar conclusions may be drawn for erosion estimates, assuming that sediment transport is proportional to a power of runoff discharge.. The spatial implications of these estimates can be made explicit with fine time resolution, showing that, with observed low overland flow velocities, only a small fraction of the hillside is generally able to deliver runoff to the nearest channel before rainfall intensity drops and runoff re-infiltrates. For coarser time resolutions, this has to be parameterised as a delivery ratio, and we show that how this ratio can be rationally estimated from rainfall characteristics.

  7. ESTIMATING RELIABILITY OF DISTURBANCES IN SATELLITE TIME SERIES DATA BASED ON STATISTICAL ANALYSIS

    Directory of Open Access Journals (Sweden)

    Z.-G. Zhou

    2016-06-01

    Full Text Available Normally, the status of land cover is inherently dynamic and changing continuously on temporal scale. However, disturbances or abnormal changes of land cover — caused by such as forest fire, flood, deforestation, and plant diseases — occur worldwide at unknown times and locations. Timely detection and characterization of these disturbances is of importance for land cover monitoring. Recently, many time-series-analysis methods have been developed for near real-time or online disturbance detection, using satellite image time series. However, the detection results were only labelled with “Change/ No change” by most of the present methods, while few methods focus on estimating reliability (or confidence level of the detected disturbances in image time series. To this end, this paper propose a statistical analysis method for estimating reliability of disturbances in new available remote sensing image time series, through analysis of full temporal information laid in time series data. The method consists of three main steps. (1 Segmenting and modelling of historical time series data based on Breaks for Additive Seasonal and Trend (BFAST. (2 Forecasting and detecting disturbances in new time series data. (3 Estimating reliability of each detected disturbance using statistical analysis based on Confidence Interval (CI and Confidence Levels (CL. The method was validated by estimating reliability of disturbance regions caused by a recent severe flooding occurred around the border of Russia and China. Results demonstrated that the method can estimate reliability of disturbances detected in satellite image with estimation error less than 5% and overall accuracy up to 90%.

  8. Estimation Accuracy on Execution Time of Run-Time Tasks in a Heterogeneous Distributed Environment

    Directory of Open Access Journals (Sweden)

    Qi Liu

    2016-08-01

    Full Text Available Distributed Computing has achieved tremendous development since cloud computing was proposed in 2006, and played a vital role promoting rapid growth of data collecting and analysis models, e.g., Internet of things, Cyber-Physical Systems, Big Data Analytics, etc. Hadoop has become a data convergence platform for sensor networks. As one of the core components, MapReduce facilitates allocating, processing and mining of collected large-scale data, where speculative execution strategies help solve straggler problems. However, there is still no efficient solution for accurate estimation on execution time of run-time tasks, which can affect task allocation and distribution in MapReduce. In this paper, task execution data have been collected and employed for the estimation. A two-phase regression (TPR method is proposed to predict the finishing time of each task accurately. Detailed data of each task have drawn interests with detailed analysis report being made. According to the results, the prediction accuracy of concurrent tasks’ execution time can be improved, in particular for some regular jobs.

  9. Nonlinear Estimation of Discrete-Time Signals Under Random Observation Delay

    International Nuclear Information System (INIS)

    Caballero-Aguila, R.; Jimenez-Lopez, J. D.; Hermoso-Carazo, A.; Linares-Perez, J.; Nakamori, S.

    2008-01-01

    This paper presents an approximation to the nonlinear least-squares estimation problem of discrete-time stochastic signals using nonlinear observations with additive white noise which can be randomly delayed by one sampling time. The observation delay is modelled by a sequence of independent Bernoulli random variables whose values, zero or one, indicate that the real observation arrives on time or it is delayed and, hence, the available measurement to estimate the signal is not up-to-date. Assuming that the state-space model generating the signal is unknown and only the covariance functions of the processes involved in the observation equation are ready for use, a filtering algorithm based on linear approximations of the real observations is proposed.

  10. Fatigue life estimation on coke drum due to cycle optimization

    Science.gov (United States)

    Siahaan, Andrey Stephan; Ambarita, Himsar; Kawai, Hideki; Daimaruya, Masashi

    2018-04-01

    In the last decade, due to the increasing demand of petroleum product, the necessity for converting the heavy oil are increasing. Thus, demand for installing coke drum in whole world will be increase. The coke drum undergoes the cyclic high temperature and suddenly cooling but in fact is not designed to withstand that kind of cycle, thus the operational life of coke drum is much shorter in comparison to other equipment in oil refinery. Various factors determine in order to improve reliability and minimize the down time, and it is found that the cycle optimization due to cycle, temperature, and pressure have an important role. From this research it is found that the fatigue life of the short cycle is decrease by a half compare to the normal cycle. It also found that in the preheating stage, the stress peak is far exceed the yield strength of coke drum material and fall into plastic deformation. This is happened because of the temperature leap in the preheating stage that cause thermal shock in the upper part of the skirt of the coke drum.

  11. Density perturbations due to the inhomogeneous discrete spatial structure of space-time

    International Nuclear Information System (INIS)

    Wolf, C.

    1998-01-01

    For the case that space-time permits an inhomogeneous discrete spatial structure due to varying gravitational fields or a foam-like structure of space-time, it is demonstrated that thermodynamic reasoning implies that matter-density perturbations will arise in the early universe

  12. Multiple Estimation Architecture in Discrete-Time Adaptive Mixing Control

    Directory of Open Access Journals (Sweden)

    Simone Baldi

    2013-05-01

    Full Text Available Adaptive mixing control (AMC is a recently developed control scheme for uncertain plants, where the control action coming from a bank of precomputed controller is mixed based on the parameter estimates generated by an on-line parameter estimator. Even if the stability of the control scheme, also in the presence of modeling errors and disturbances, has been shown analytically, its transient performance might be sensitive to the initial conditions of the parameter estimator. In particular, for some initial conditions, transient oscillations may not be acceptable in practical applications. In order to account for such a possible phenomenon and to improve the learning capability of the adaptive scheme, in this paper a new mixing architecture is developed, involving the use of parallel parameter estimators, or multi-estimators, each one working on a small subset of the uncertainty set. A supervisory logic, using performance signals based on the past and present estimation error, selects the parameter estimate to determine the mixing of the controllers. The stability and robustness properties of the resulting approach, referred to as multi-estimator adaptive mixing control (Multi-AMC, are analytically established. Besides, extensive simulations demonstrate that the scheme improves the transient performance of the original AMC with a single estimator. The control scheme and the analysis are carried out in a discrete-time framework, for easier implementation of the method in digital control.

  13. Methodology for time-domain estimation of storm time geoelectric fields using the 3-D magnetotelluric response tensors

    Science.gov (United States)

    Kelbert, Anna; Balch, Christopher C.; Pulkkinen, Antti; Egbert, Gary D.; Love, Jeffrey J.; Rigler, E. Joshua; Fujii, Ikuko

    2017-07-01

    Geoelectric fields at the Earth's surface caused by magnetic storms constitute a hazard to the operation of electric power grids and related infrastructure. The ability to estimate these geoelectric fields in close to real time and provide local predictions would better equip the industry to mitigate negative impacts on their operations. Here we report progress toward this goal: development of robust algorithms that convolve a magnetic storm time series with a frequency domain impedance for a realistic three-dimensional (3-D) Earth, to estimate the local, storm time geoelectric field. Both frequency domain and time domain approaches are presented and validated against storm time geoelectric field data measured in Japan. The methods are then compared in the context of a real-time application.

  14. Heterogeneous Data Fusion Method to Estimate Travel Time Distributions in Congested Road Networks

    Directory of Open Access Journals (Sweden)

    Chaoyang Shi

    2017-12-01

    Full Text Available Travel times in congested urban road networks are highly stochastic. Provision of travel time distribution information, including both mean and variance, can be very useful for travelers to make reliable path choice decisions to ensure higher probability of on-time arrival. To this end, a heterogeneous data fusion method is proposed to estimate travel time distributions by fusing heterogeneous data from point and interval detectors. In the proposed method, link travel time distributions are first estimated from point detector observations. The travel time distributions of links without point detectors are imputed based on their spatial correlations with links that have point detectors. The estimated link travel time distributions are then fused with path travel time distributions obtained from the interval detectors using Dempster-Shafer evidence theory. Based on fused path travel time distribution, an optimization technique is further introduced to update link travel time distributions and their spatial correlations. A case study was performed using real-world data from Hong Kong and showed that the proposed method obtained accurate and robust estimations of link and path travel time distributions in congested road networks.

  15. Estimation of the internal radiation dose received by the adult population of Graz due to the contamination with radioactive iodine and cesium

    International Nuclear Information System (INIS)

    Rabitsch, H.; Kahr, G.

    1991-07-01

    During the first months following the fallout we have measured the activities of J-131 in some human thyroids. To study the long-term variation of radiocesium with time, we have observed the activity level of Cs-134 and Cs-137 in human muscle tissues over a period of 4 years. Simultaneously we have determined the activities of the naturally occuring potassium-40 in all samples, which were taken at forensic autopsies of persons deceased in the area of Graz. Comparisons of iodine and radiocesium activities measured in the samples with data obtained by other studies after nuclear weapon tests are given. Average individual thyroid dose was calculated to be 556.1 μSv. The main part of this thyroid dose is caused by the inhalation pathway. Effective individual dose equivalents originated by the radiocesium body content were calculated by means of time-integrated activities and the method of absorbed fractions. Dose estimations based on data of the Standard Man and a distribution factor of 0.7 was assumed with regards to the amount of radiocesium and K-40 in muscle mass. From the measurements, we have estimated a mean individual effective dose equivalent of 252.2 μSv due to internal exposure to radiocesium during the 4 years following the fallout. Estimated dose values are compared with predictions and the exposure caused by K-40. (Authors, shortened by Quittner)

  16. Estimation of fuel loss due to idling of vehicles at a signalized intersection in Chennai, India

    Science.gov (United States)

    Vasantha Kumar, S.; Gulati, Himanshu; Arora, Shivam

    2017-11-01

    The vehicles while waiting at signalized intersections are generally found to be in idling condition, i.e., not switching off their vehicles during red times. This phenomenon of idling of vehicles during red times at signalized intersections may lead to huge economic loss as lot of fuel is consumed by vehicles when they are in idling condition. The situation may even be worse in countries like India as different vehicle types consume varying amount of fuel. Only limited studies have been reported on estimation of fuel loss due to idling of vehicles in India. In the present study, one of the busy intersections in Chennai, namely, Tidel Park Junction in Rajiv Gandhi salai was considered. Data collection was carried out in one approach road of the intersection during morning and evening peak hours on a typical working day by manually noting down the red timings of each cycle and the corresponding number of two-wheelers, three-wheelers, passenger cars, light commercial vehicles (LCV) and heavy motorized vehicles (HMV) that were in idling mode. Using the fuel consumption values of various vehicles types suggested by Central Road Research Institute (CRRI), the total fuel loss during the study period was found to be Rs. 4,93,849/-. The installation of red timers, synchronization of signals, use of non-motorized transport for short trips and public awareness are some of the measures which government need to focus to save the fuel wasted at signalized intersections in major cities of India.

  17. Estimating evolutionary rates using time-structured data: a general comparison of phylogenetic methods.

    Science.gov (United States)

    Duchêne, Sebastián; Geoghegan, Jemma L; Holmes, Edward C; Ho, Simon Y W

    2016-11-15

    In rapidly evolving pathogens, including viruses and some bacteria, genetic change can accumulate over short time-frames. Accordingly, their sampling times can be used to calibrate molecular clocks, allowing estimation of evolutionary rates. Methods for estimating rates from time-structured data vary in how they treat phylogenetic uncertainty and rate variation among lineages. We compiled 81 virus data sets and estimated nucleotide substitution rates using root-to-tip regression, least-squares dating and Bayesian inference. Although estimates from these three methods were often congruent, this largely relied on the choice of clock model. In particular, relaxed-clock models tended to produce higher rate estimates than methods that assume constant rates. Discrepancies in rate estimates were also associated with high among-lineage rate variation, and phylogenetic and temporal clustering. These results provide insights into the factors that affect the reliability of rate estimates from time-structured sequence data, emphasizing the importance of clock-model testing. sduchene@unimelb.edu.au or garzonsebastian@hotmail.comSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  18. Adaptive divergence in flowering time among natural populations of Arabidopsis thaliana: Estimates of selection and QTL mapping.

    Science.gov (United States)

    Ågren, Jon; Oakley, Christopher G; Lundemo, Sverre; Schemske, Douglas W

    2017-03-01

    To identify the ecological and genetic mechanisms of local adaptation requires estimating selection on traits, identifying their genetic basis, and evaluating whether divergence in adaptive traits is due to conditional neutrality or genetic trade-offs. To this end, we conducted field experiments for three years using recombinant inbred lines (RILs) derived from two ecotypes of Arabidopsis thaliana (Italy, Sweden), and at each parental site examined selection on flowering time and mapped quantitative trait loci (QTL). There was strong selection for early flowering in Italy, but weak selection in Sweden. Eleven distinct flowering time QTL were detected, and for each the Italian genotype caused earlier flowering. Twenty-seven candidate genes were identified, two of which (FLC and VIN3) appear under major flowering time QTL in Italy. Seven of eight QTL in Italy with narrow credible intervals colocalized with previously reported fitness QTL, in comparison to three of four in Sweden. The results demonstrate that the magnitude of selection on flowering time differs strikingly between our study populations, that the genetic basis of flowering time variation is multigenic with some QTL of large effect, and suggest that divergence in flowering time between ecotypes is due mainly to conditional neutrality. © 2016 The Author(s). Evolution © 2016 The Society for the Study of Evolution.

  19. Influence of hypo- and hyperthermia on death time estimation - A simulation study.

    Science.gov (United States)

    Muggenthaler, H; Hubig, M; Schenkl, S; Mall, G

    2017-09-01

    Numerous physiological and pathological mechanisms can cause elevated or lowered body core temperatures. Deviations from the physiological level of about 37°C can influence temperature based death time estimations. However, it has not been investigated by means of thermodynamics, to which extent hypo- and hyperthermia bias death time estimates. Using numerical simulation, the present study investigates the errors inherent in temperature based death time estimation in case of elevated or lowered body core temperatures before death. The most considerable errors with regard to the normothermic model occur in the first few hours post-mortem. With decreasing body core temperature and increasing post-mortem time the error diminishes and stagnates at a nearly constant level. Copyright © 2017 Elsevier B.V. All rights reserved.

  20. Estimated time of arrival and debiasing the time saving bias.

    Science.gov (United States)

    Eriksson, Gabriella; Patten, Christopher J D; Svenson, Ola; Eriksson, Lars

    2015-01-01

    The time saving bias predicts that the time saved when increasing speed from a high speed is overestimated, and underestimated when increasing speed from a slow speed. In a questionnaire, time saving judgements were investigated when information of estimated time to arrival was provided. In an active driving task, an alternative meter indicating the inverted speed was used to debias judgements. The simulated task was to first drive a distance at a given speed, and then drive the same distance again at the speed the driver judged was required to gain exactly 3 min in travel time compared with the first drive. A control group performed the same task with a speedometer and saved less than the targeted 3 min when increasing speed from a high speed, and more than 3 min when increasing from a low speed. Participants in the alternative meter condition were closer to the target. The two studies corroborate a time saving bias and show that biased intuitive judgements can be debiased by displaying the inverted speed. Practitioner Summary: Previous studies have shown a cognitive bias in judgements of the time saved by increasing speed. This simulator study aims to improve driver judgements by introducing a speedometer indicating the inverted speed in active driving. The results show that the bias can be reduced by presenting the inverted speed and this finding can be used when designing in-car information systems.

  1. Using linear time-invariant system theory to estimate kinetic parameters directly from projection measurements

    International Nuclear Information System (INIS)

    Zeng, G.L.; Gullberg, G.T.

    1995-01-01

    It is common practice to estimate kinetic parameters from dynamically acquired tomographic data by first reconstructing a dynamic sequence of three-dimensional reconstructions and then fitting the parameters to time activity curves generated from the time-varying reconstructed images. However, in SPECT, the pharmaceutical distribution can change during the acquisition of a complete tomographic data set, which can bias the estimated kinetic parameters. It is hypothesized that more accurate estimates of the kinetic parameters can be obtained by fitting to the projection measurements instead of the reconstructed time sequence. Estimation from projections requires the knowledge of their relationship between the tissue regions of interest or voxels with particular kinetic parameters and the project measurements, which results in a complicated nonlinear estimation problem with a series of exponential factors with multiplicative coefficients. A technique is presented in this paper where the exponential decay parameters are estimated separately using linear time-invariant system theory. Once the exponential factors are known, the coefficients of the exponentials can be estimated using linear estimation techniques. Computer simulations demonstrate that estimation of the kinetic parameters directly from the projections is more accurate than the estimation from the reconstructed images

  2. Empirical estimation of the arrival time of ICME Shocks

    Science.gov (United States)

    Shaltout, Mosalam

    Empirical estimation of the arrival time of ICME Shocks Mosalam Shaltout1 ,M.Youssef 1and R.Mawad2 1 National Research Institute of Astronomy and Geophysics (NRIAG) ,Helwan -Cairo-Egypt Email: mosalamshaltout@hotmail.com 2 Faculty of Science-Monifiia University-Physics Department-Shiben Al-Koum -Monifiia-Egypt We are got the Data of the SSC events from Preliminary Reports of the ISGI (Institut de Physique du Globe, France) .Also we are selected the same CME interval 1996-2005 from SOHO/LASCO/C2.We have estimated the arrival time of ICME shocks during solar cycle 23rd (1996-2005), we take the Sudden storm commencement SSC as a indicator of the arrival of CMEs at the Earth's Magnetosphere (ICME).Under our model ,we selected 203 ICME shock-SSC associated events, we got an imperial relation between CME velocity and their travel time, from which we obtained high correlation between them, R=0.75.

  3. Estimation of Collective Effective Dose Due to Cosmic Ray Exposures to Members of The Public and to Airline Passenger

    International Nuclear Information System (INIS)

    Sayed, N.S.; Salah Eldin, T.; Gomaa, M.A.; El Dosoky, T.M.

    2011-01-01

    Using UNSCEAR 2000 report to United Nation General Assembly and its appendices, Annual collective dose to Egyptian members of the public (75097301). Was estimated to be 252.5 man Sv , hence the average collective effective dose to air line passenger for 10 million is estimated as 25.25 micro Sievert. Furthermore using hypothetical approach for Egyptian passengers who fly locally, regionally and internationally, the collective dose was estimated to be 252.5 man Sv , hence the average average collective effective dose for Egyptian passenger is due to Aviation is 3.36 micro Sievert

  4. Iterative Bayesian Estimation of Travel Times on Urban Arterials: Fusing Loop Detector and Probe Vehicle Data.

    Science.gov (United States)

    Liu, Kai; Cui, Meng-Ying; Cao, Peng; Wang, Jiang-Bo

    2016-01-01

    On urban arterials, travel time estimation is challenging especially from various data sources. Typically, fusing loop detector data and probe vehicle data to estimate travel time is a troublesome issue while considering the data issue of uncertain, imprecise and even conflicting. In this paper, we propose an improved data fusing methodology for link travel time estimation. Link travel times are simultaneously pre-estimated using loop detector data and probe vehicle data, based on which Bayesian fusion is then applied to fuse the estimated travel times. Next, Iterative Bayesian estimation is proposed to improve Bayesian fusion by incorporating two strategies: 1) substitution strategy which replaces the lower accurate travel time estimation from one sensor with the current fused travel time; and 2) specially-designed conditions for convergence which restrict the estimated travel time in a reasonable range. The estimation results show that, the proposed method outperforms probe vehicle data based method, loop detector based method and single Bayesian fusion, and the mean absolute percentage error is reduced to 4.8%. Additionally, iterative Bayesian estimation performs better for lighter traffic flows when the variability of travel time is practically higher than other periods.

  5. Autoregressive-model-based missing value estimation for DNA microarray time series data.

    Science.gov (United States)

    Choong, Miew Keen; Charbit, Maurice; Yan, Hong

    2009-01-01

    Missing value estimation is important in DNA microarray data analysis. A number of algorithms have been developed to solve this problem, but they have several limitations. Most existing algorithms are not able to deal with the situation where a particular time point (column) of the data is missing entirely. In this paper, we present an autoregressive-model-based missing value estimation method (ARLSimpute) that takes into account the dynamic property of microarray temporal data and the local similarity structures in the data. ARLSimpute is especially effective for the situation where a particular time point contains many missing values or where the entire time point is missing. Experiment results suggest that our proposed algorithm is an accurate missing value estimator in comparison with other imputation methods on simulated as well as real microarray time series datasets.

  6. NOTE ON TRAVEL TIME SHIFTS DUE TO AMPLITUDE MODULATION IN TIME-DISTANCE HELIOSEISMOLOGY MEASUREMENTS

    International Nuclear Information System (INIS)

    Nigam, R.; Kosovichev, A. G.

    2010-01-01

    Correct interpretation of acoustic travel times measured by time-distance helioseismology is essential to get an accurate understanding of the solar properties that are inferred from them. It has long been observed that sunspots suppress p-mode amplitude, but its implications on travel times have not been fully investigated so far. It has been found in test measurements using a 'masking' procedure, in which the solar Doppler signal in a localized quiet region of the Sun is artificially suppressed by a spatial function, and using numerical simulations that the amplitude modulations in combination with the phase-speed filtering may cause systematic shifts of acoustic travel times. To understand the properties of this procedure, we derive an analytical expression for the cross-covariance of a signal that has been modulated locally by a spatial function that has azimuthal symmetry and then filtered by a phase-speed filter typically used in time-distance helioseismology. Comparing this expression to the Gabor wavelet fitting formula without this effect, we find that there is a shift in the travel times that is introduced by the amplitude modulation. The analytical model presented in this paper can be useful also for interpretation of travel time measurements for the non-uniform distribution of oscillation amplitude due to observational effects.

  7. Nonparametric autocovariance estimation from censored time series by Gaussian imputation.

    Science.gov (United States)

    Park, Jung Wook; Genton, Marc G; Ghosh, Sujit K

    2009-02-01

    One of the most frequently used methods to model the autocovariance function of a second-order stationary time series is to use the parametric framework of autoregressive and moving average models developed by Box and Jenkins. However, such parametric models, though very flexible, may not always be adequate to model autocovariance functions with sharp changes. Furthermore, if the data do not follow the parametric model and are censored at a certain value, the estimation results may not be reliable. We develop a Gaussian imputation method to estimate an autocovariance structure via nonparametric estimation of the autocovariance function in order to address both censoring and incorrect model specification. We demonstrate the effectiveness of the technique in terms of bias and efficiency with simulations under various rates of censoring and underlying models. We describe its application to a time series of silicon concentrations in the Arctic.

  8. Time Estimation in Alzheimer's Disease and the Role of the Central Executive

    Science.gov (United States)

    Papagno, Costanza; Allegra, Adele; Cardaci, Maurizio

    2004-01-01

    The aim of this study was to evaluate the role of short-term memory and attention in time estimation. For this purpose we studied prospective time verbal estimation in 21 patients with Alzheimer's disease (AD), and compared their performance with that of 21 matched normal controls in two different conditions: during a digit span task and during an…

  9. Estimate of Passive Time Reversal Communication Performance in Shallow Water

    Directory of Open Access Journals (Sweden)

    Sunhyo Kim

    2017-12-01

    Full Text Available Time reversal processes have been used to improve communication performance in the severe underwater communication environment characterized by significant multipath channels by reducing inter-symbol interference and increasing signal-to-noise ratio. In general, the performance of the time reversal is strongly related to the behavior of the q -function, which is estimated by a sum of the autocorrelation of the channel impulse response for each channel in the receiver array. The q -function depends on the complexity of the communication channel, the number of channel elements and their spacing. A q -function with a high side-lobe level and a main-lobe width wider than the symbol duration creates a residual ISI (inter-symbol interference, which makes communication difficult even after time reversal is applied. In this paper, we propose a new parameter, E q , to describe the performance of time reversal communication. E q is an estimate of how much of the q -function lies within one symbol duration. The values of E q were estimated using communication data acquired at two different sites: one in which the sound speed ratio of sediment to water was less than unity and one where the ratio was higher than unity. Finally, the parameter E q was compared to the bit error rate and the output signal-to-noise ratio obtained after the time reversal operation. The results show that these parameters are strongly correlated to the parameter E q .

  10. Stationary echo canceling in velocity estimation by time-domain cross-correlation

    DEFF Research Database (Denmark)

    Jensen, Jørgen Arendt

    1993-01-01

    The application of stationary echo canceling to ultrasonic estimation of blood velocities using time-domain cross-correlation is investigated. Expressions are derived that show the influence from the echo canceler on the signals that enter the cross-correlation estimator. It is demonstrated...

  11. Estimating spatial travel times using automatic vehicle identification data

    Science.gov (United States)

    2001-01-01

    Prepared ca. 2001. The paper describes an algorithm that was developed for estimating reliable and accurate average roadway link travel times using Automatic Vehicle Identification (AVI) data. The algorithm presented is unique in two aspects. First, ...

  12. Own-wage labor supply elasticities: variation across time and estimation methods

    Directory of Open Access Journals (Sweden)

    Olivier Bargain

    2016-10-01

    Full Text Available Abstract There is a huge variation in the size of labor supply elasticities in the literature, which hampers policy analysis. While recent studies show that preference heterogeneity across countries explains little of this variation, we focus on two other important features: observation period and estimation method. We start with a thorough survey of existing evidence for both Western Europe and the USA, over a long period and from different empirical approaches. Then, our meta-analysis attempts to disentangle the role of time changes and estimation methods. We highlight the key role of time changes, documenting the incredible fall in labor supply elasticities since the 1980s not only for the USA but also in the EU. In contrast, we find no compelling evidence that the choice of estimation method explains variation in elasticity estimates. From our analysis, we derive important guidelines for policy simulations.

  13. Remaining Useful Life Estimation using Time Trajectory Tracking and Support Vector Machines

    International Nuclear Information System (INIS)

    Galar, D; Kumar, U; Lee, J; Zhao, W

    2012-01-01

    In this paper, a novel RUL prediction method inspired by feature maps and SVM classifiers is proposed. The historical instances of a system with life-time condition data are used to create a classification by SVM hyper planes. For a test instance of the same system, whose RUL is to be estimated, degradation speed is evaluated by computing the minimal distance defined based on the degradation trajectories, i.e. the approach of the system to the hyper plane that segregates good and bad condition data at different time horizon. Therefore, the final RUL of a specific component can be estimated and global RUL information can then be obtained by aggregating the multiple RUL estimations using a density estimation method.

  14. Estimation of traffic recovery time for different flow regimes on freeways.

    Science.gov (United States)

    2008-06-01

    This study attempts to estimate post-incident traffic recovery time along a freeway using Monte Carlo simulation techniques. It has been found that there is a linear relationship between post-incident traffic recovery time, and incident time and traf...

  15. Time Series Decomposition into Oscillation Components and Phase Estimation.

    Science.gov (United States)

    Matsuda, Takeru; Komaki, Fumiyasu

    2017-02-01

    Many time series are naturally considered as a superposition of several oscillation components. For example, electroencephalogram (EEG) time series include oscillation components such as alpha, beta, and gamma. We propose a method for decomposing time series into such oscillation components using state-space models. Based on the concept of random frequency modulation, gaussian linear state-space models for oscillation components are developed. In this model, the frequency of an oscillator fluctuates by noise. Time series decomposition is accomplished by this model like the Bayesian seasonal adjustment method. Since the model parameters are estimated from data by the empirical Bayes' method, the amplitudes and the frequencies of oscillation components are determined in a data-driven manner. Also, the appropriate number of oscillation components is determined with the Akaike information criterion (AIC). In this way, the proposed method provides a natural decomposition of the given time series into oscillation components. In neuroscience, the phase of neural time series plays an important role in neural information processing. The proposed method can be used to estimate the phase of each oscillation component and has several advantages over a conventional method based on the Hilbert transform. Thus, the proposed method enables an investigation of the phase dynamics of time series. Numerical results show that the proposed method succeeds in extracting intermittent oscillations like ripples and detecting the phase reset phenomena. We apply the proposed method to real data from various fields such as astronomy, ecology, tidology, and neuroscience.

  16. Estimation of dynamic flux profiles from metabolic time series data

    Directory of Open Access Journals (Sweden)

    Chou I-Chun

    2012-07-01

    Full Text Available Abstract Background Advances in modern high-throughput techniques of molecular biology have enabled top-down approaches for the estimation of parameter values in metabolic systems, based on time series data. Special among them is the recent method of dynamic flux estimation (DFE, which uses such data not only for parameter estimation but also for the identification of functional forms of the processes governing a metabolic system. DFE furthermore provides diagnostic tools for the evaluation of model validity and of the quality of a model fit beyond residual errors. Unfortunately, DFE works only when the data are more or less complete and the system contains as many independent fluxes as metabolites. These drawbacks may be ameliorated with other types of estimation and information. However, such supplementations incur their own limitations. In particular, assumptions must be made regarding the functional forms of some processes and detailed kinetic information must be available, in addition to the time series data. Results The authors propose here a systematic approach that supplements DFE and overcomes some of its shortcomings. Like DFE, the approach is model-free and requires only minimal assumptions. If sufficient time series data are available, the approach allows the determination of a subset of fluxes that enables the subsequent applicability of DFE to the rest of the flux system. The authors demonstrate the procedure with three artificial pathway systems exhibiting distinct characteristics and with actual data of the trehalose pathway in Saccharomyces cerevisiae. Conclusions The results demonstrate that the proposed method successfully complements DFE under various situations and without a priori assumptions regarding the model representation. The proposed method also permits an examination of whether at all, to what degree, or within what range the available time series data can be validly represented in a particular functional format of

  17. Only through perturbation can relaxation times be estimated

    Czech Academy of Sciences Publication Activity Database

    Ditlevsen, S.; Lánský, Petr

    2012-01-01

    Roč. 86, č. 5 (2012), 050102-5 ISSN 1539-3755 R&D Projects: GA ČR(CZ) GAP103/11/0282; GA ČR(CZ) GBP304/12/G069 Institutional support: RVO:67985823 Keywords : stochastic diffusion * parameter estimation * time constant Subject RIV: JD - Computer Applications, Robotics Impact factor: 2.313, year: 2012

  18. Use of Low-Level Sensor Data to Improve the Accuracy of Bluetooth-Based Travel Time Estimation

    DEFF Research Database (Denmark)

    Araghi, Bahar Namaki; Christensen, Lars Tørholm; Krishnan, Rajesh

    2013-01-01

    by a single device. The latter situation could lead to location ambiguity and could reduce the accuracy of travel time estimation. Therefore, the accuracy of travel time estimation by Bluetooth technology depends on how location ambiguity is handled by the estimation method. The issue of multiple detection...... events in the context of travel time estimation by Bluetooth technology has been considered by various researchers. However, treatment of this issue has been simplistic. Most previous studies have used the first detection event (enter-enter) as the best estimate. No systematic analysis has been conducted...... to explore the most accurate method of travel time estimation with multiple detection events. In this study, different aspects of the Bluetooth detection zone, including size and impact on the accuracy of travel time estimation, were discussed. Four methods were applied to estimate travel time: enter...

  19. Estimation of airway smooth muscle stiffness changes due to length oscillation using artificial neural network.

    Science.gov (United States)

    Al-Jumaily, Ahmed; Chen, Leizhi

    2012-10-07

    This paper presents a novel approach to estimate stiffness changes in airway smooth muscles due to external oscillation. Artificial neural networks are used to model the stiffness changes due to cyclic stretches of the smooth muscles. The nonlinear relationship between stiffness ratios and oscillation frequencies is modeled by a feed-forward neural network (FNN) model. The structure of the FNN is selected through the training and validation using literature data from 11 experiments with different muscle lengths, muscle masses, oscillation frequencies and amplitudes. Data pre-processing methods are used to improve the robustness of the neural network model to match the non-linearity. The validation results show that the FNN model can predict the stiffness ratio changes with a mean square error of 0.0042. Copyright © 2012 Elsevier Ltd. All rights reserved.

  20. Tightness of M-estimators for multiple linear regression in time series

    DEFF Research Database (Denmark)

    Johansen, Søren; Nielsen, Bent

    We show tightness of a general M-estimator for multiple linear regression in time series. The positive criterion function for the M-estimator is assumed lower semi-continuous and sufficiently large for large argument: Particular cases are the Huber-skip and quantile regression. Tightness requires...

  1. Optical losses due to tracking error estimation for a low concentrating solar collector

    International Nuclear Information System (INIS)

    Sallaberry, Fabienne; García de Jalón, Alberto; Torres, José-Luis; Pujol-Nadal, Ramón

    2015-01-01

    Highlights: • A solar thermal collector with low concentration and one-axis tracking was tested. • A quasi-dynamic testing procedure for IAM was defined for tracking collector. • The adequation between the concentrator optics and the tracking was checked. • The maximum and long-term optical losses due to tracking error were calculated. - Abstract: The determination of the accuracy of a solar tracker used in domestic hot water solar collectors is not yet standardized. However, while using optical concentration devices, it is important to use a solar tracker with adequate precision with regard to the specific optical concentration factor. Otherwise, the concentrator would sustain high optical losses due to the inadequate focusing of the solar radiation onto its receiver, despite having a good quality. This study is focused on the estimation of long-term optical losses due to the tracking error of a low-temperature collector using low-concentration optics. For this purpose, a testing procedure for the incidence angle modifier on the tracking plane is proposed to determinate the acceptance angle of its concentrator even with different longitudinal incidence angles along the focal line plane. Then, the impact of maximum tracking error angle upon the optical efficiency has been determined. Finally, the calculation of the long-term optical error due to the tracking errors, using the design angular tracking error declared by the manufacturer, is carried out. The maximum tracking error calculated for this collector imply an optical loss of about 8.5%, which is high, but the average long-term optical loss calculated for one year was about 1%, which is reasonable for such collectors used for domestic hot water

  2. Real-time soft tissue motion estimation for lung tumors during radiotherapy delivery.

    Science.gov (United States)

    Rottmann, Joerg; Keall, Paul; Berbeco, Ross

    2013-09-01

    To provide real-time lung tumor motion estimation during radiotherapy treatment delivery without the need for implanted fiducial markers or additional imaging dose to the patient. 2D radiographs from the therapy beam's-eye-view (BEV) perspective are captured at a frame rate of 12.8 Hz with a frame grabber allowing direct RAM access to the image buffer. An in-house developed real-time soft tissue localization algorithm is utilized to calculate soft tissue displacement from these images in real-time. The system is tested with a Varian TX linear accelerator and an AS-1000 amorphous silicon electronic portal imaging device operating at a resolution of 512 × 384 pixels. The accuracy of the motion estimation is verified with a dynamic motion phantom. Clinical accuracy was tested on lung SBRT images acquired at 2 fps. Real-time lung tumor motion estimation from BEV images without fiducial markers is successfully demonstrated. For the phantom study, a mean tracking error real-time markerless lung tumor motion estimation from BEV images alone. The described system can operate at a frame rate of 12.8 Hz and does not require prior knowledge to establish traceable landmarks for tracking on the fly. The authors show that the geometric accuracy is similar to (or better than) previously published markerless algorithms not operating in real-time.

  3. Estimation of Nuclei Cooling Time by Electrons in Superdense Nonequilibrium Plasma

    CERN Document Server

    Kostenko, B F

    2004-01-01

    Estimations of nuclei cooling time by electrons in superdense nonequilibrium plasma formed at cavitation bubble collapse in deuterated acetone have been carried out. The necessity of these computations was stipulated by using in the latest theoretical calculations of nuclear reaction rate in these processes one poorly grounded assumption that electron temperatures remain essentially lower than nuclei ones during thermonuclear synthesis time t_s. The estimations have shown that the initial electron temperatures at the moment of superdense plasma formation with \\rho =100 g/cm^3 turn out to be appreciably lower than the nuclear temperatures, while the nuclei cooling time is of the same order as t_s.

  4. Time Delay Estimation in Room Acoustic Environments: An Overview

    Directory of Open Access Journals (Sweden)

    Benesty Jacob

    2006-01-01

    Full Text Available Time delay estimation has been a research topic of significant practical importance in many fields (radar, sonar, seismology, geophysics, ultrasonics, hands-free communications, etc.. It is a first stage that feeds into subsequent processing blocks for identifying, localizing, and tracking radiating sources. This area has made remarkable advances in the past few decades, and is continuing to progress, with an aim to create processors that are tolerant to both noise and reverberation. This paper presents a systematic overview of the state-of-the-art of time-delay-estimation algorithms ranging from the simple cross-correlation method to the advanced blind channel identification based techniques. We discuss the pros and cons of each individual algorithm, and outline their inherent relationships. We also provide experimental results to illustrate their performance differences in room acoustic environments where reverberation and noise are commonly encountered.

  5. Patients with Testicular Cancer Undergoing CT Surveillance Demonstrate a Pitfall of Radiation-induced Cancer Risk Estimates: The Timing Paradox

    Science.gov (United States)

    Eisenberg, Jonathan D.; Lee, Richard J.; Gilmore, Michael E.; Turan, Ekin A.; Singh, Sarabjeet; Kalra, Mannudeep K.; Liu, Bob; Kong, Chung Yin; Gazelle, G. Scott

    2013-01-01

    Purpose: To demonstrate a limitation of lifetime radiation-induced cancer risk metrics in the setting of testicular cancer surveillance—in particular, their failure to capture the delayed timing of radiation-induced cancers over the course of a patient’s lifetime. Materials and Methods: Institutional review board approval was obtained for the use of computed tomographic (CT) dosimetry data in this study. Informed consent was waived. This study was HIPAA compliant. A Markov model was developed to project outcomes in patients with testicular cancer who were undergoing CT surveillance in the decade after orchiectomy. To quantify effects of early versus delayed risks, life expectancy losses and lifetime mortality risks due to testicular cancer were compared with life expectancy losses and lifetime mortality risks due to radiation-induced cancers from CT. Projections of life expectancy loss, unlike lifetime risk estimates, account for the timing of risks over the course of a lifetime, which enabled evaluation of the described limitation of lifetime risk estimates. Markov chain Monte Carlo methods were used to estimate the uncertainty of the results. Results: As an example of evidence yielded, 33-year-old men with stage I seminoma who were undergoing CT surveillance were projected to incur a slightly higher lifetime mortality risk from testicular cancer (598 per 100 000; 95% uncertainty interval [UI]: 302, 894) than from radiation-induced cancers (505 per 100 000; 95% UI: 280, 730). However, life expectancy loss attributable to testicular cancer (83 days; 95% UI: 42, 124) was more than three times greater than life expectancy loss attributable to radiation-induced cancers (24 days; 95% UI: 13, 35). Trends were consistent across modeled scenarios. Conclusion: Lifetime radiation risk estimates, when used for decision making, may overemphasize radiation-induced cancer risks relative to short-term health risks. © RSNA, 2012 Supplemental material: http

  6. Facing the estimation of effective population size based on molecular markers: comparison of estimators

    DEFF Research Database (Denmark)

    Jimenez Mena, Belen; Verrier, Etienne; Hospital, Frederic

    an increase in the variability of values over time. The distance from the mean and the median to the true Ne increased over time too. This was caused by the fixation of alleles through time due to genetic drift and the changes in the distribution of allele frequencies. We compared the three estimators of Ne...

  7. Travel Time Estimation on Urban Street Segment

    Directory of Open Access Journals (Sweden)

    Jelena Kajalić

    2018-02-01

    Full Text Available Level of service (LOS is used as the main indicator of transport quality on urban roads and it is estimated based on the travel speed. The main objective of this study is to determine which of the existing models for travel speed calculation is most suitable for local conditions. The study uses actual data gathered in travel time survey on urban streets, recorded by applying second by second GPS data. The survey is limited to traffic flow in saturated conditions. The RMSE method (Root Mean Square Error is used for research results comparison with relevant models: Akcelik, HCM (Highway Capacity Manual, Singapore model and modified BPR (the Bureau of Public Roads function (Dowling - Skabardonis. The lowest deviation in local conditions for urban streets with standardized intersection distance (400-500 m is demonstrated by Akcelik model. However, for streets with lower signal density (<1 signal/km the correlation between speed and degree of saturation is best presented by HCM and Singapore model. According to test results, Akcelik model was adopted for travel speed estimation which can be the basis for determining the level of service in urban streets with standardized intersection distance and coordinated signal timing under local conditions.

  8. The time aspect of bioenergy. Climate impacts of bioenergy due to differences in carbon uptake rates

    Energy Technology Data Exchange (ETDEWEB)

    Zetterberg, Lars [IVL Swedish Environmental Research Institute, Stockholm (Sweden); Chen, Deliang [Dept. of Earth Sciences, Univ. of Gothenburg, Gothenburg (Sweden)

    2011-07-01

    This paper investigates the climate impacts from bioenergy due to how they influence carbon stocks over time and more specifically how fast combustion related carbon emissions are compensated by uptake of atmospheric carbon. A set of fuel types representing different uptake rates are investigated, namely willow, branches and tops, stumps and coal. Net emissions are defined as emissions from utilizing the fuel minus emissions from a reference case of no utilisation. In the case of forest residues, the compensating 'uptake' is avoided emissions from the reference case of leaving the residues to decompose on the ground. Climate impacts are estimated using the measures radiative forcing and global average surface temperature, which have been calculated by an energy balance climate model. We conclude that there is a climate impact from using bioenergy due to how fast the emission pulse is compensated by uptake of atmospheric carbon (or avoided emissions). Biofuels with slower uptake rates have a stronger climate impact than fuels with a faster uptake rate, assuming all other parameters equal. The time perspective over which the analysis is done is crucial for the climate impact of biofuels. If only biogenic fluxes are considered, our results show that over a 100 year perspective branches and tops are better for climate mitigation than stumps which in turn are better than coal. Over a 20 year time perspective this conclusion holds, but the differences between these fuels are relatively smaller. Establishing willow on earlier crop land may reduce atmospheric carbon, provided new land is available. However, these results are inconclusive since we haven't considered the effects, if needed, of producing the traditional agricultural crops elsewhere. The analysis is not a life cycle assessment of different fuels and does therefore not consider the use of fossil fuels for logging, transportation and refining, other greenhouse gases than carbon or energy

  9. An anti-disturbing real time pose estimation method and system

    Science.gov (United States)

    Zhou, Jian; Zhang, Xiao-hu

    2011-08-01

    Pose estimation relating two-dimensional (2D) images to three-dimensional (3D) rigid object need some known features to track. In practice, there are many algorithms which perform this task in high accuracy, but all of these algorithms suffer from features lost. This paper investigated the pose estimation when numbers of known features or even all of them were invisible. Firstly, known features were tracked to calculate pose in the current and the next image. Secondly, some unknown but good features to track were automatically detected in the current and the next image. Thirdly, those unknown features which were on the rigid and could match each other in the two images were retained. Because of the motion characteristic of the rigid object, the 3D information of those unknown features on the rigid could be solved by the rigid object's pose at the two moment and their 2D information in the two images except only two case: the first one was that both camera and object have no relative motion and camera parameter such as focus length, principle point, and etc. have no change at the two moment; the second one was that there was no shared scene or no matched feature in the two image. Finally, because those unknown features at the first time were known now, pose estimation could go on in the followed images in spite of the missing of known features in the beginning by repeating the process mentioned above. The robustness of pose estimation by different features detection algorithms such as Kanade-Lucas-Tomasi (KLT) feature, Scale Invariant Feature Transform (SIFT) and Speed Up Robust Feature (SURF) were compared and the compact of the different relative motion between camera and the rigid object were discussed in this paper. Graphic Processing Unit (GPU) parallel computing was also used to extract and to match hundreds of features for real time pose estimation which was hard to work on Central Processing Unit (CPU). Compared with other pose estimation methods, this new

  10. PolyWaTT: A polynomial water travel time estimator based on Derivative Dynamic Time Warping and Perceptually Important Points

    Science.gov (United States)

    Claure, Yuri Navarro; Matsubara, Edson Takashi; Padovani, Carlos; Prati, Ronaldo Cristiano

    2018-03-01

    Traditional methods for estimating timing parameters in hydrological science require a rigorous study of the relations of flow resistance, slope, flow regime, watershed size, water velocity, and other local variables. These studies are mostly based on empirical observations, where the timing parameter is estimated using empirically derived formulas. The application of these studies to other locations is not always direct. The locations in which equations are used should have comparable characteristics to the locations from which such equations have been derived. To overcome this barrier, in this work, we developed a data-driven approach to estimate timing parameters such as travel time. Our proposal estimates timing parameters using historical data of the location without the need of adapting or using empirical formulas from other locations. The proposal only uses one variable measured at two different locations on the same river (for instance, two river-level measurements, one upstream and the other downstream on the same river). The recorded data from each location generates two time series. Our method aligns these two time series using derivative dynamic time warping (DDTW) and perceptually important points (PIP). Using data from timing parameters, a polynomial function generalizes the data by inducing a polynomial water travel time estimator, called PolyWaTT. To evaluate the potential of our proposal, we applied PolyWaTT to three different watersheds: a floodplain ecosystem located in the part of Brazil known as Pantanal, the world's largest tropical wetland area; and the Missouri River and the Pearl River, in United States of America. We compared our proposal with empirical formulas and a data-driven state-of-the-art method. The experimental results demonstrate that PolyWaTT showed a lower mean absolute error than all other methods tested in this study, and for longer distances the mean absolute error achieved by PolyWaTT is three times smaller than empirical

  11. Nonlinear response of vessel walls due to short-time thermomechanical loading

    International Nuclear Information System (INIS)

    Pfeiffer, P.A.; Kulak, R.F.

    1994-01-01

    Maintaining structural integrity of the reactor pressure vessel (RPV) during a postulated core melt accident is an important safety consideration in the design of the vessel. This study addresses the failure predictions of the vessel due to thermal and pressure loadings fro the molten core debris depositing on the lower head of the vessel. Different loading combinations were considered based on the dead load, yield stress assumptions, material response and internal pressurization. The analyses considered only short term failure (quasi static) modes, long term failure modes were not considered. Short term failure modes include plastic instabilities of the structure and failure due to exceeding the failure strain. Long term failure odes would be caused by creep rupture that leads to plastic instability of the structure. Due to the sort time durations analyzed, creep was not considered in the analyses presented

  12. Time estimation in Parkinson's disease and degenerative cerebellar disease

    NARCIS (Netherlands)

    Beudel, Martijin; Galama, Sjoukje; Leenders, Klaus L.; de Jong, Bauke M.

    2008-01-01

    With functional MRI, we recently identified fronto-cerebellar activations in predicting time to reach a target and basal ganglia activation in velocity estimation, that is, small interval assessment. We now tested these functions in patients with Parkinson's disease (PD) and degenerative cerebellar

  13. On algebraic time-derivative estimation and deadbeat state reconstruction

    DEFF Research Database (Denmark)

    Reger, Johann; Jouffroy, Jerome

    2009-01-01

    This paper places into perspective the so-called algebraic time-derivative estimation method recently introduced by Fliess and co-authors with standard results from linear statespace theory for control systems. In particular, it is shown that the algebraic method can essentially be seen...

  14. Seasonal adjustment methods and real time trend-cycle estimation

    CERN Document Server

    Bee Dagum, Estela

    2016-01-01

    This book explores widely used seasonal adjustment methods and recent developments in real time trend-cycle estimation. It discusses in detail the properties and limitations of X12ARIMA, TRAMO-SEATS and STAMP - the main seasonal adjustment methods used by statistical agencies. Several real-world cases illustrate each method and real data examples can be followed throughout the text. The trend-cycle estimation is presented using nonparametric techniques based on moving averages, linear filters and reproducing kernel Hilbert spaces, taking recent advances into account. The book provides a systematical treatment of results that to date have been scattered throughout the literature. Seasonal adjustment and real time trend-cycle prediction play an essential part at all levels of activity in modern economies. They are used by governments to counteract cyclical recessions, by central banks to control inflation, by decision makers for better modeling and planning and by hospitals, manufacturers, builders, transportat...

  15. Estimation of time-varying reactivity by the H∞ optimal linear filter

    International Nuclear Information System (INIS)

    Suzuki, Katsuo; Shimazaki, Junya; Watanabe, Koiti

    1995-01-01

    The problem of estimating the time-varying net reactivity from flux measurements is solved for a point reactor kinetics model using a linear filtering technique in an H ∞ settings. In order to sue this technique, an appropriate dynamical model of the reactivity is constructed that can be embedded into the reactor model as one of its variables. A filter, which minimizes the H ∞ norm of the estimation error power spectrum, operates on neutron density measurements corrupted by noise and provides an estimate of the dynamic net reactivity. Computer simulations are performed to reveal the basic characteristics of the H ∞ optimal filter. The results of the simulation indicate that the filter can be used to determine the time-varying reactivity from neutron density measurements that have been corrupted by noise

  16. Kernel density estimation-based real-time prediction for respiratory motion

    International Nuclear Information System (INIS)

    Ruan, Dan

    2010-01-01

    Effective delivery of adaptive radiotherapy requires locating the target with high precision in real time. System latency caused by data acquisition, streaming, processing and delivery control necessitates prediction. Prediction is particularly challenging for highly mobile targets such as thoracic and abdominal tumors undergoing respiration-induced motion. The complexity of the respiratory motion makes it difficult to build and justify explicit models. In this study, we honor the intrinsic uncertainties in respiratory motion and propose a statistical treatment of the prediction problem. Instead of asking for a deterministic covariate-response map and a unique estimate value for future target position, we aim to obtain a distribution of the future target position (response variable) conditioned on the observed historical sample values (covariate variable). The key idea is to estimate the joint probability distribution (pdf) of the covariate and response variables using an efficient kernel density estimation method. Then, the problem of identifying the distribution of the future target position reduces to identifying the section in the joint pdf based on the observed covariate. Subsequently, estimators are derived based on this estimated conditional distribution. This probabilistic perspective has some distinctive advantages over existing deterministic schemes: (1) it is compatible with potentially inconsistent training samples, i.e., when close covariate variables correspond to dramatically different response values; (2) it is not restricted by any prior structural assumption on the map between the covariate and the response; (3) the two-stage setup allows much freedom in choosing statistical estimates and provides a full nonparametric description of the uncertainty for the resulting estimate. We evaluated the prediction performance on ten patient RPM traces, using the root mean squared difference between the prediction and the observed value normalized by the

  17. Pulsar timing residuals due to individual non-evolving gravitational wave sources

    International Nuclear Information System (INIS)

    Tong Ming-Lei; Zhao Cheng-Shi; Yan Bao-Rong; Yang Ting-Gao; Gao Yu-Ping

    2014-01-01

    The pulsar timing residuals induced by gravitational waves from non-evolving single binary sources are affected by many parameters related to the relative positions of the pulsar and the gravitational wave sources. We will analyze the various effects due to different parameters. The standard deviations of the timing residuals will be calculated with a variable parameter fixing a set of other parameters. The orbits of the binary sources will be generally assumed to be elliptical. The influences of different eccentricities on the pulsar timing residuals will also be studied in detail. We find that the effects of the related parameters are quite different, and some of them display certain regularities

  18. Robust time estimation reconciles views of the antiquity of placental mammals.

    Directory of Open Access Journals (Sweden)

    Yasuhiro Kitazoe

    2007-04-01

    Full Text Available Molecular studies have reported divergence times of modern placental orders long before the Cretaceous-Tertiary boundary and far older than paleontological data. However, this discrepancy may not be real, but rather appear because of the violation of implicit assumptions in the estimation procedures, such as non-gradual change of evolutionary rate and failure to correct for convergent evolution.New procedures for divergence-time estimation robust to abrupt changes in the rate of molecular evolution are described. We used a variant of the multidimensional vector space (MVS procedure to take account of possible convergent evolution. Numerical simulations of abrupt rate change and convergent evolution showed good performance of the new procedures in contrast to current methods. Application to complete mitochondrial genomes identified marked rate accelerations and decelerations, which are not obtained with current methods. The root of placental mammals is estimated to be approximately 18 million years more recent than when assuming a log Brownian motion model. Correcting the pairwise distances for convergent evolution using MVS lowers the age of the root about another 20 million years compared to using standard maximum likelihood tree branch lengths. These two procedures combined revise the root time of placental mammals from around 122 million years ago to close to 84 million years ago. As a result, the estimated distribution of molecular divergence times is broadly consistent with quantitative analysis of the North American fossil record and traditional morphological views.By including the dual effects of abrupt rate change and directly accounting for convergent evolution at the molecular level, these estimates provide congruence between the molecular results, paleontological analyses and morphological expectations. The programs developed here are provided along with sample data that reproduce the results of this study and are especially

  19. Estimation of the time-dependent radioactive source-term from the Fukushima nuclear power plant accident using atmospheric transport modelling

    Science.gov (United States)

    Schoeppner, M.; Plastino, W.; Budano, A.; De Vincenzi, M.; Ruggieri, F.

    2012-04-01

    Several nuclear reactors at the Fukushima Dai-ichi power plant have been severely damaged from the Tōhoku earthquake and the subsequent tsunami in March 2011. Due to the extremely difficult on-site situation it has been not been possible to directly determine the emissions of radioactive material. However, during the following days and weeks radionuclides of 137-Caesium and 131-Iodine (amongst others) were detected at monitoring stations throughout the world. Atmospheric transport models are able to simulate the worldwide dispersion of particles accordant to location, time and meteorological conditions following the release. The Lagrangian atmospheric transport model Flexpart is used by many authorities and has been proven to make valid predictions in this regard. The Flexpart software has first has been ported to a local cluster computer at the Grid Lab of INFN and Department of Physics of University of Roma Tre (Rome, Italy) and subsequently also to the European Mediterranean Grid (EUMEDGRID). Due to this computing power being available it has been possible to simulate the transport of particles originating from the Fukushima Dai-ichi plant site. Using the time series of the sampled concentration data and the assumption that the Fukushima accident was the only source of these radionuclides, it has been possible to estimate the time-dependent source-term for fourteen days following the accident using the atmospheric transport model. A reasonable agreement has been obtained between the modelling results and the estimated radionuclide release rates from the Fukushima accident.

  20. Using fuzzy logic to improve the project time and cost estimation based on Project Evaluation and Review Technique (PERT

    Directory of Open Access Journals (Sweden)

    Farhad Habibi

    2018-09-01

    Full Text Available Among different factors, correct scheduling is one of the vital elements for project management success. There are several ways to schedule projects including the Critical Path Method (CPM and Program Evaluation and Review Technique (PERT. Due to problems in estimating dura-tions of activities, these methods cannot accurately and completely model actual projects. The use of fuzzy theory is a basic way to improve scheduling and deal with such problems. Fuzzy theory approximates project scheduling models to reality by taking into account uncertainties in decision parameters and expert experience and mental models. This paper provides a step-by-step approach for accurate estimation of time and cost of projects using the Project Evaluation and Review Technique (PERT and expert views as fuzzy numbers. The proposed method included several steps. In the first step, the necessary information for project time and cost is estimated using the Critical Path Method (CPM and the Project Evaluation and Review Technique (PERT. The second step considers the duration and cost of the project activities as the trapezoidal fuzzy numbers, and then, the time and cost of the project are recalculated. The duration and cost of activities are estimated using the questionnaires as well as weighing the expert opinions, averaging and defuzzification based on a step-by-step algorithm. The calculating procedures for evaluating these methods are applied in a real project; and the obtained results are explained.

  1. Estimated time spent on preventive services by primary care physicians

    Directory of Open Access Journals (Sweden)

    Gradison Margaret

    2008-12-01

    Full Text Available Abstract Background Delivery of preventive health services in primary care is lacking. One of the main barriers is lack of time. We estimated the amount of time primary care physicians spend on important preventive health services. Methods We analyzed a large dataset of primary care (family and internal medicine visits using the National Ambulatory Medical Care Survey (2001–4; analyses were conducted 2007–8. Multiple linear regression was used to estimate the amount of time spent delivering each preventive service, controlling for demographic covariates. Results Preventive visits were longer than chronic care visits (M = 22.4, SD = 11.8, M = 18.9, SD = 9.2, respectively. New patients required more time from physicians. Services on which physicians spent relatively more time were prostate specific antigen (PSA, cholesterol, Papanicolaou (Pap smear, mammography, exercise counseling, and blood pressure. Physicians spent less time than recommended on two "A" rated ("good evidence" services, tobacco cessation and Pap smear (in preventive visits, and one "B" rated ("at least fair evidence" service, nutrition counseling. Physicians spent substantial time on two services that have an "I" rating ("inconclusive evidence of effectiveness", PSA and exercise counseling. Conclusion Even with limited time, physicians address many of the "A" rated services adequately. However, they may be spending less time than recommended for important services, especially smoking cessation, Pap smear, and nutrition counseling. Future research is needed to understand how physicians decide how to allocate their time to address preventive health.

  2. Mesospheric temperature estimation from meteor decay times of weak and strong meteor trails

    Science.gov (United States)

    Kim, Jeong-Han; Kim, Yong Ha; Jee, Geonhwa; Lee, Changsup

    2012-11-01

    Neutral temperatures near the mesopause region were estimated from the decay times of the meteor echoes observed by a VHF meteor radar during a period covering 2007 to 2009 at King Sejong Station (62.22°S, 58.78°W), Antarctica. While some previous studies have used all meteor echoes to determine the slope from a height profile of log inverse decay times for temperature estimation, we have divided meteor echoes into weak and strong groups of underdense meteor trails, depending on the strength of estimated relative electron line densities within meteor trails. We found that the slopes from the strong group are inappropriate for temperature estimation because the decay times of strong meteors are considerably scattered, whereas the slopes from the weak group clearly define the variation of decay times with height. We thus utilize the slopes only from the weak group in the altitude region between 86 km and 96 km to estimate mesospheric temperatures. The meteor estimated temperatures show a typical seasonal variation near the mesopause region and the monthly mean temperatures are in good agreement with SABER temperatures within a mean difference of 4.8 K throughout the year. The meteor temperatures, representing typically the region around the altitude of 91 km, are lower on average by 2.1 K than simultaneously measured SATI OH(6-2) rotational temperatures during winter (March-October).

  3. Maximum Likelihood Time-of-Arrival Estimation of Optical Pulses via Photon-Counting Photodetectors

    Science.gov (United States)

    Erkmen, Baris I.; Moision, Bruce E.

    2010-01-01

    Many optical imaging, ranging, and communications systems rely on the estimation of the arrival time of an optical pulse. Recently, such systems have been increasingly employing photon-counting photodetector technology, which changes the statistics of the observed photocurrent. This requires time-of-arrival estimators to be developed and their performances characterized. The statistics of the output of an ideal photodetector, which are well modeled as a Poisson point process, were considered. An analytical model was developed for the mean-square error of the maximum likelihood (ML) estimator, demonstrating two phenomena that cause deviations from the minimum achievable error at low signal power. An approximation was derived to the threshold at which the ML estimator essentially fails to provide better than a random guess of the pulse arrival time. Comparing the analytic model performance predictions to those obtained via simulations, it was verified that the model accurately predicts the ML performance over all regimes considered. There is little prior art that attempts to understand the fundamental limitations to time-of-arrival estimation from Poisson statistics. This work establishes both a simple mathematical description of the error behavior, and the associated physical processes that yield this behavior. Previous work on mean-square error characterization for ML estimators has predominantly focused on additive Gaussian noise. This work demonstrates that the discrete nature of the Poisson noise process leads to a distinctly different error behavior.

  4. Time of arrival based location estimation for cooperative relay networks

    KAUST Repository

    Ç elebi, Hasari Burak; Abdallah, Mohamed M.; Hussain, Syed Imtiaz; Qaraqe, Khalid A.; Alouini, Mohamed-Slim

    2010-01-01

    In this paper, we investigate the performance of a cooperative relay network performing location estimation through time of arrival (TOA). We derive Cramer-Rao lower bound (CRLB) for the location estimates using the relay network. The analysis is extended to obtain average CRLB considering the signal fluctuations in both relay and direct links. The effects of the channel fading of both relay and direct links and amplification factor and location of the relay node on average CRLB are investigated. Simulation results show that the channel fading of both relay and direct links and amplification factor and location of relay node affect the accuracy of TOA based location estimation. ©2010 IEEE.

  5. Time of arrival based location estimation for cooperative relay networks

    KAUST Repository

    Çelebi, Hasari Burak

    2010-09-01

    In this paper, we investigate the performance of a cooperative relay network performing location estimation through time of arrival (TOA). We derive Cramer-Rao lower bound (CRLB) for the location estimates using the relay network. The analysis is extended to obtain average CRLB considering the signal fluctuations in both relay and direct links. The effects of the channel fading of both relay and direct links and amplification factor and location of the relay node on average CRLB are investigated. Simulation results show that the channel fading of both relay and direct links and amplification factor and location of relay node affect the accuracy of TOA based location estimation. ©2010 IEEE.

  6. Face to phase: pitfalls in time delay estimation from coherency phase

    NARCIS (Netherlands)

    Campfens, S.F.; van der Kooij, Herman; Schouten, Alfred Christiaan

    2014-01-01

    Coherency phase is often interpreted as a time delay reflecting a transmission delay between spatially separated neural populations. However, time delays estimated from corticomuscular coherency are conflicting and often shorter than expected physiologically. Recent work suggests that

  7. Dose rates as a function of time due to postulated radionuclide releases from the U.S. Yucca Mountain high-level radioactive waste repository

    International Nuclear Information System (INIS)

    Moeller, Dade W.; Sun, Lin-Shen C.; Cherry, Robert

    2008-01-01

    The Yucca Mountain repository, which is located in a remote area in the State of Nevada, is being constructed for the long-term care and disposal of spent nuclear fuel and vitrified high-level radioactive waste. In accordance with U.S. law, the U.S. Environmental Protection Agency (USEPA) promulgated Standards that limit the dose rates to members of the public due to the consumption of ground water, alone, and the consumption of ground water plus agricultural products irrigated with the contaminated ground water, and other exposures, such as those from external sources and the inhalation of airborne radioactive materials. As part of this exercise, the USEPA identified eight specific radionuclides to which their Standards are to apply. These are: 14 C, 99 Tc, 129 I, 226 Ra, 228 Ra, 237 Np, 239 Pu, and 241 Am. For purposes of the associated dose rate estimates, a range of conservative assumptions have been applied, all of which are designed to assure that the estimated dose rates are well above what might be expected under 'real-world' conditions. As a first step, it was assumed that: (1) at 10 4 year after repository closure, a fractional release of 10 -5 of the entire repository radionuclide inventory occurred; (2) the only prior reduction in the inventory was that due to radioactive decay; and (3) the sole path of exposure to neighboring population groups was through the consumption of 2 L d -1 of contaminated ground water. The accompanying analyses revealed that, of the eight radionuclides, only 226 Ra, 237 Np, and 239 Pu, will represent a significant source of dose at that time. To provide perspective and insights, the next step was to estimate the committed effective dose rates for all eight radionuclides based on an assumed fractional release each year of 10 -5 of the inventory from the time of repository closure up through the 10 6 year. For purposes of providing perspective, it was assumed that each dose rate estimate was independent, that is, no releases

  8. A test of alternative estimators for volume at time 1 from remeasured point samples

    Science.gov (United States)

    Francis A. Roesch; Edwin J. Green; Charles T. Scott

    1993-01-01

    Two estimators for volume at time 1 for use with permanent horizontal point samples are evaluated. One estimator, used traditionally, uses only the trees sampled at time 1, while the second estimator, originally presented by Roesch and coauthors (F.A. Roesch, Jr., E.J. Green, and C.T. Scott. 1989. For. Sci. 35(2):281-293). takes advantage of additional sample...

  9. Optimal replacement time estimation for machines and equipment based on cost function

    Directory of Open Access Journals (Sweden)

    J. Šebo

    2013-01-01

    Full Text Available The article deals with a multidisciplinary issue of estimating the optimal replacement time for the machines. Considered categories of machines, for which the optimization method is usable, are of the metallurgical and engineering production. Different models of cost function are considered (both with one and two variables. Parameters of the models were calculated through the least squares method. Models testing show that all are good enough, so for estimation of optimal replacement time is sufficient to use simpler models. In addition to the testing of models we developed the method (tested on selected simple model which enable us in actual real time (with limited data set to indicate the optimal replacement time. The indicated time moment is close enough to the optimal replacement time t*.

  10. Freeway travel time estimation using existing fixed traffic sensors : phase 2.

    Science.gov (United States)

    2015-03-01

    Travel time, one of the most important freeway performance metrics, can be easily estimated using the : data collected from fixed traffic sensors, avoiding the need to install additional travel time data collectors. : This project is aimed at fully u...

  11. Optimal replacement time estimation for machines and equipment based on cost function

    OpenAIRE

    J. Šebo; J. Buša; P. Demeč; J. Svetlík

    2013-01-01

    The article deals with a multidisciplinary issue of estimating the optimal replacement time for the machines. Considered categories of machines, for which the optimization method is usable, are of the metallurgical and engineering production. Different models of cost function are considered (both with one and two variables). Parameters of the models were calculated through the least squares method. Models testing show that all are good enough, so for estimation of optimal replacement time is ...

  12. Environmental Kuznets revisited. Time-series versus panel estimation. The CO2-case

    International Nuclear Information System (INIS)

    Dijkgraaf, E.; Vollebergh, H.R.J.

    1998-01-01

    According to the Environmental Kuznets Curve (EKC) hypothesis economic growth and improving environmental quality are compatible. An inverted U-shaped relationship would exist between economic performance and environmental quality suggesting that after some threshold a growing economy would cause smaller pollution. Usually the EKC hypothesis is tested for pooled panel data of some (sub)set of environmental indicators and GDP. The essential assumption behind pooling the observations of different countries in one panel is that the outcome of the economic process would be the same for all countries with respect to emissions. That is, the curvature of the Income-Emission Relation (IER) is the same for the pooled countries as far as they have the same GDP range. In our study we show this methodology to be misleading for at least the special case of carbon dioxide (CO2) emissions and the GDP level. Using OECD-wide data from 1960 to 1990, we find that the pooled estimation results show positive auto-correlation. Other studies correct for this problem. However, as we show, it seems more likely that the estimated regression model is not correct due to the pooling of the country observations. Testing the model per country reveals that the IER is very different for countries. While some countries show growing carbon dioxide emissions per capita with increasing income, others show a stabilizing pattern or even an EKC. This indicates that estimations based on pooling techniques can bias the conclusion about the true IER leading to unjustified inferences on the existence of the EKC. Extending the basic model may result in the justification of pooling. However, our estimations including country specific variables, like population density, openness of the economy and the availability of own fuel sources (endowment effects), do not make us optimistic. The autocorrelation problem remains. If a more general model can not be found the only remaining conclusion is that testing the EKC

  13. Real-time soft tissue motion estimation for lung tumors during radiotherapy delivery

    International Nuclear Information System (INIS)

    Rottmann, Joerg; Berbeco, Ross; Keall, Paul

    2013-01-01

    Purpose: To provide real-time lung tumor motion estimation during radiotherapy treatment delivery without the need for implanted fiducial markers or additional imaging dose to the patient.Methods: 2D radiographs from the therapy beam's-eye-view (BEV) perspective are captured at a frame rate of 12.8 Hz with a frame grabber allowing direct RAM access to the image buffer. An in-house developed real-time soft tissue localization algorithm is utilized to calculate soft tissue displacement from these images in real-time. The system is tested with a Varian TX linear accelerator and an AS-1000 amorphous silicon electronic portal imaging device operating at a resolution of 512 × 384 pixels. The accuracy of the motion estimation is verified with a dynamic motion phantom. Clinical accuracy was tested on lung SBRT images acquired at 2 fps.Results: Real-time lung tumor motion estimation from BEV images without fiducial markers is successfully demonstrated. For the phantom study, a mean tracking error <1.0 mm [root mean square (rms) error of 0.3 mm] was observed. The tracking rms accuracy on BEV images from a lung SBRT patient (≈20 mm tumor motion range) is 1.0 mm.Conclusions: The authors demonstrate for the first time real-time markerless lung tumor motion estimation from BEV images alone. The described system can operate at a frame rate of 12.8 Hz and does not require prior knowledge to establish traceable landmarks for tracking on the fly. The authors show that the geometric accuracy is similar to (or better than) previously published markerless algorithms not operating in real-time

  14. Real-time soft tissue motion estimation for lung tumors during radiotherapy delivery

    Energy Technology Data Exchange (ETDEWEB)

    Rottmann, Joerg; Berbeco, Ross [Brigham and Women' s Hospital, Dana Farber-Cancer Institute and Harvard Medical School, Boston, Massachusetts 02115 (United States); Keall, Paul [Radiation Physics Laboratory, Sydney Medical School, University of Sydney, Sydney NSW 2006 (Australia)

    2013-09-15

    Purpose: To provide real-time lung tumor motion estimation during radiotherapy treatment delivery without the need for implanted fiducial markers or additional imaging dose to the patient.Methods: 2D radiographs from the therapy beam's-eye-view (BEV) perspective are captured at a frame rate of 12.8 Hz with a frame grabber allowing direct RAM access to the image buffer. An in-house developed real-time soft tissue localization algorithm is utilized to calculate soft tissue displacement from these images in real-time. The system is tested with a Varian TX linear accelerator and an AS-1000 amorphous silicon electronic portal imaging device operating at a resolution of 512 × 384 pixels. The accuracy of the motion estimation is verified with a dynamic motion phantom. Clinical accuracy was tested on lung SBRT images acquired at 2 fps.Results: Real-time lung tumor motion estimation from BEV images without fiducial markers is successfully demonstrated. For the phantom study, a mean tracking error <1.0 mm [root mean square (rms) error of 0.3 mm] was observed. The tracking rms accuracy on BEV images from a lung SBRT patient (≈20 mm tumor motion range) is 1.0 mm.Conclusions: The authors demonstrate for the first time real-time markerless lung tumor motion estimation from BEV images alone. The described system can operate at a frame rate of 12.8 Hz and does not require prior knowledge to establish traceable landmarks for tracking on the fly. The authors show that the geometric accuracy is similar to (or better than) previously published markerless algorithms not operating in real-time.

  15. A novel method for in-situ estimation of time constant for core temperature monitoring thermocouples of operating reactors

    International Nuclear Information System (INIS)

    Sylvia, J.I.; Chandar, S. Clement Ravi; Velusamy, K.

    2014-01-01

    Highlights: • Core temperature sensor was mathematically modeled. • Ramp signal generated during reactor operating condition is used. • Procedure and methodology has been demonstrated by applying it to FBTR. • Same technique will be implemented for all fast reactors. - Abstract: Core temperature monitoring system is an important component of reactor protection system in the current generation fast reactors. In this system, multiple thermocouples are housed inside a thermowell of fuel subassemblies. Response time of the thermocouple assembly forms an important input for safety analysis of fast reactor and hence frequent calibration/time constant estimation is essential. In fast reactors the central fuel subassembly is provided with bare fast response thermocouples to detect under cooling events in reactor and take proper safety action. On the other hand, thermocouples in thermowell are mainly used for blockage detection in individual fuel subassemblies. The time constant of thermocouples in thermowell can drift due to creep, vibration and thermal fatigue of the thermowell assembly. A novel method for in-situ estimation of time constant is proposed. This method uses the Safety Control Rod Accelerated Mechanism (SCRAM) or lowering of control Rod (LOR) signals of the reactor along with response of the central subassembly thermocouples as reference data. Validation of the procedure has been demonstrated by applying it to FBTR

  16. Analytical model for time to cover cracking in RC structures due to rebar corrosion

    International Nuclear Information System (INIS)

    Bhargava, Kapilesh; Ghosh, A.K.; Mori, Yasuhiro; Ramanujam, S.

    2006-01-01

    The structural degradation of concrete structures due to reinforcement corrosion is a major worldwide problem. Reinforcement corrosion causes a volume increase due to the oxidation of metallic iron, which is mainly responsible for exerting the expansive radial pressure at the steel-concrete interface and development of hoop tensile stresses in the surrounding concrete. Cracking occurs, once the maximum hoop tensile stress exceeds the tensile strength of the concrete. The cracking begins at the steel-concrete interface and propagates outwards and eventually results in the thorough cracking of the cover concrete and this would indicate the loss of service life for the corrosion affected structures. An analytical model is proposed to predict the time required for cover cracking and the weight loss of reinforcing bar in corrosion affected reinforced concrete structures. The modelling aspects of the residual strength of cracked concrete and the stiffness contribution from the combination of reinforcement and expansive corrosion products have also been incorporated in the model. The problem is modeled as a boundary value problem and the governing equations are expressed in terms of the radial displacement. The analytical solutions are presented considering a simple two-zone model for the cover concrete, viz. cracked or uncracked. Reasonable estimation of the various parameters in the model related to the composition and properties of expansive corrosion products based on the available published experimental data has also been discussed. The performance of the proposed corrosion cracking model is then investigated through its ability to reproduce available experimental trends. Reasonably good agreement between experimental results and the analytical predictions has been obtained. It has also been found that tensile strength and initial tangent modulus of cover concrete, annual mean corrosion rate and modulus of elasticity of reinforcement plus corrosion products combined

  17. Estimate of lifetime excess lung cancer risk due to indoor exposure to natural radon-222 daughters in Korea

    International Nuclear Information System (INIS)

    Si-Young Chang; Jeong-Ho Lee; Chung-Woo Ha

    1993-01-01

    Lifetime excess lung cancer risk due to indoor 222 Rn daughters exposure in Korea was quantitatively estimated by a modified relative risk projection model proposed by the U.S. National Academy of Science and the recent Korean life table data. The lifetime excess risk of lung cancer death attributable to annual constant exposure to Korean indoor radon daughters was estimated to be about 230/10 6 per WLM, which seemed to be nearly in the median of the range of 150-450/10 6 per WLM reported by the UNSCEAR in 1988. (1 fig., 2 tabs.)

  18. Real-time measurements and their effects on state estimation of distribution power system

    DEFF Research Database (Denmark)

    Han, Xue; You, Shi; Thordarson, Fannar

    2013-01-01

    between the estimated values (voltage and injected power) and the measurements are applied to evaluate the accuracy of the estimated grid states. Eventually, some suggestions are provided for the distribution grid operators on placing the real-time meters in the distribution grid.......This paper aims at analyzing the potential value of using different real-time metering and measuring instruments applied in the low voltage distribution networks for state-estimation. An algorithm is presented to evaluate different combinations of metering data using a tailored state estimator....... It is followed by a case study based on the proposed algorithm. A real distribution grid feeder with different types of meters installed either in the cabinets or at the customer side is selected for simulation and analysis. Standard load templates are used to initiate the state estimation. The deviations...

  19. Estimating WACC for Regulated Industries on Developing Financial Markets and in Times of Market Uncertainty

    Directory of Open Access Journals (Sweden)

    Igor Stubelj

    2014-03-01

    Full Text Available The paper deals with the estimation of weighted average cost of capital (WACC for regulated industries in developing financial markets from the perspective of the current financial-economic crisis. In current financial market situation some evident changes have occurred: risk-free rates in solid and developed financial markets (e. g. USA, Germany have fallen, but due to increased market volatility, the risk premiums have increased. The latter is especially evident in transition economies where the amplitude of market volatility is extremely high. In such circumstances, there is a question of how to calculate WACC properly. WACC is an important measure in financial management decisions and in our case, business regulation. We argue in the paper that the most accurate method for calculating WACC is the estimation of the long-term WACC, which takes into consideration a long-term stable yield of capital and not the current market conditions. Following this, we propose some solutions that could be used for calculating WACC for regulated industries on the developing financial markets in times of market uncertainty. As an example, we present an estimation of the capital cost for a selected Slovenian company, which operates in the regulated industry of electric distribution.

  20. Estimate of the Effective Dose Equivalent to the Cypriot Population due to Diagnostic Nuclear Medicine Procedures in the Public Sector

    Energy Technology Data Exchange (ETDEWEB)

    Christofides, S [Medical Physics Department, Nicosia General Hospital (Cyprus)

    1994-12-31

    The Effective Dose Equivalent (EDE) to the Cypriot population due to Diagnostic Nuclear Medicine procedures has been estimated from data published by the Government of Cyprus, in its Health and Hospital Statistics Series for the years 1990, 1991, and 1992. The average EDE per patient was estimated to be 3,09, 3,75 and 4,01 microSievert for 1990, 1991 and 1992 respectively, while the per caput EDE was estimated to be 11,75, 15,16 and 17,09 microSieverts for 1990, 1991 and 1992 respectively, from the procedures in the public sector. (author). 11 refs, 4 tabs.

  1. Estimate of the Effective Dose Equivalent to the Cypriot Population due to Diagnostic Nuclear Medicine Procedures in the Public Sector

    International Nuclear Information System (INIS)

    Christofides, S.

    1994-01-01

    The Effective Dose Equivalent (EDE) to the Cypriot population due to Diagnostic Nuclear Medicine procedures has been estimated from data published by the Government of Cyprus, in its Health and Hospital Statistics Series for the years 1990, 1991, and 1992. The average EDE per patient was estimated to be 3,09, 3,75 and 4,01 microSievert for 1990, 1991 and 1992 respectively, while the per caput EDE was estimated to be 11,75, 15,16 and 17,09 microSieverts for 1990, 1991 and 1992 respectively, from the procedures in the public sector. (author)

  2. Estimating time-based instantaneous total mortality rate based on the age-structured abundance index

    Science.gov (United States)

    Wang, Yingbin; Jiao, Yan

    2015-05-01

    The instantaneous total mortality rate ( Z) of a fish population is one of the important parameters in fisheries stock assessment. The estimation of Z is crucial to fish population dynamics analysis, abundance and catch forecast, and fisheries management. A catch curve-based method for estimating time-based Z and its change trend from catch per unit effort (CPUE) data of multiple cohorts is developed. Unlike the traditional catch-curve method, the method developed here does not need the assumption of constant Z throughout the time, but the Z values in n continuous years are assumed constant, and then the Z values in different n continuous years are estimated using the age-based CPUE data within these years. The results of the simulation analyses show that the trends of the estimated time-based Z are consistent with the trends of the true Z, and the estimated rates of change from this approach are close to the true change rates (the relative differences between the change rates of the estimated Z and the true Z are smaller than 10%). Variations of both Z and recruitment can affect the estimates of Z value and the trend of Z. The most appropriate value of n can be different given the effects of different factors. Therefore, the appropriate value of n for different fisheries should be determined through a simulation analysis as we demonstrated in this study. Further analyses suggested that selectivity and age estimation are also two factors that can affect the estimated Z values if there is error in either of them, but the estimated change rates of Z are still close to the true change rates. We also applied this approach to the Atlantic cod ( Gadus morhua) fishery of eastern Newfoundland and Labrador from 1983 to 1997, and obtained reasonable estimates of time-based Z.

  3. Low-sampling-rate ultra-wideband channel estimation using equivalent-time sampling

    KAUST Repository

    Ballal, Tarig

    2014-09-01

    In this paper, a low-sampling-rate scheme for ultra-wideband channel estimation is proposed. The scheme exploits multiple observations generated by transmitting multiple pulses. In the proposed scheme, P pulses are transmitted to produce channel impulse response estimates at a desired sampling rate, while the ADC samples at a rate that is P times slower. To avoid loss of fidelity, the number of sampling periods (based on the desired rate) in the inter-pulse interval is restricted to be co-prime with P. This condition is affected when clock drift is present and the transmitted pulse locations change. To handle this case, and to achieve an overall good channel estimation performance, without using prior information, we derive an improved estimator based on the bounded data uncertainty (BDU) model. It is shown that this estimator is related to the Bayesian linear minimum mean squared error (LMMSE) estimator. Channel estimation performance of the proposed sub-sampling scheme combined with the new estimator is assessed in simulation. The results show that high reduction in sampling rate can be achieved. The proposed estimator outperforms the least squares estimator in almost all cases, while in the high SNR regime it also outperforms the LMMSE estimator. In addition to channel estimation, a synchronization method is also proposed that utilizes the same pulse sequence used for channel estimation. © 2014 IEEE.

  4. Evaluating methods for estimating space-time paths of individuals in calculating long-term personal exposure to air pollution

    Science.gov (United States)

    Schmitz, Oliver; Soenario, Ivan; Vaartjes, Ilonca; Strak, Maciek; Hoek, Gerard; Brunekreef, Bert; Dijst, Martin; Karssenberg, Derek

    2016-04-01

    Air pollution is one of the major concerns for human health. Associations between air pollution and health are often calculated using long-term (i.e. years to decades) information on personal exposure for each individual in a cohort. Personal exposure is the air pollution aggregated along the space-time path visited by an individual. As air pollution may vary considerably in space and time, for instance due to motorised traffic, the estimation of the spatio-temporal location of a persons' space-time path is important to identify the personal exposure. However, long term exposure is mostly calculated using the air pollution concentration at the x, y location of someone's home which does not consider that individuals are mobile (commuting, recreation, relocation). This assumption is often made as it is a major challenge to estimate space-time paths for all individuals in large cohorts, mostly because limited information on mobility of individuals is available. We address this issue by evaluating multiple approaches for the calculation of space-time paths, thereby estimating the personal exposure along these space-time paths with hyper resolution air pollution maps at national scale. This allows us to evaluate the effect of the space-time path and resulting personal exposure. Air pollution (e.g. NO2, PM10) was mapped for the entire Netherlands at a resolution of 5×5 m2 using the land use regression models developed in the European Study of Cohorts for Air Pollution Effects (ESCAPE, http://escapeproject.eu/) and the open source software PCRaster (http://www.pcraster.eu). The models use predictor variables like population density, land use, and traffic related data sets, and are able to model spatial variation and within-city variability of annual average concentration values. We approximated space-time paths for all individuals in a cohort using various aggregations, including those representing space-time paths as the outline of a persons' home or associated parcel

  5. Estimation of Nonlinear Functions of State Vector for Linear Systems with Time-Delays and Uncertainties

    Directory of Open Access Journals (Sweden)

    Il Young Song

    2015-01-01

    Full Text Available This paper focuses on estimation of a nonlinear function of state vector (NFS in discrete-time linear systems with time-delays and model uncertainties. The NFS represents a multivariate nonlinear function of state variables, which can indicate useful information of a target system for control. The optimal nonlinear estimator of an NFS (in mean square sense represents a function of the receding horizon estimate and its error covariance. The proposed receding horizon filter represents the standard Kalman filter with time-delays and special initial horizon conditions described by the Lyapunov-like equations. In general case to calculate an optimal estimator of an NFS we propose using the unscented transformation. Important class of polynomial NFS is considered in detail. In the case of polynomial NFS an optimal estimator has a closed-form computational procedure. The subsequent application of the proposed receding horizon filter and nonlinear estimator to a linear stochastic system with time-delays and uncertainties demonstrates their effectiveness.

  6. Access to destinations : arterial data acquisition and network-wide travel time estimation (phase II).

    Science.gov (United States)

    2010-03-01

    The objectives of this project were to (a) produce historic estimates of travel times on Twin-Cities arterials : for 1995 and 2005, and (b) develop an initial architecture and database that could, in the future, produce timely : estimates of arterial...

  7. Simultaneous Robust Fault and State Estimation for Linear Discrete-Time Uncertain Systems

    Directory of Open Access Journals (Sweden)

    Feten Gannouni

    2017-01-01

    Full Text Available We consider the problem of robust simultaneous fault and state estimation for linear uncertain discrete-time systems with unknown faults which affect both the state and the observation matrices. Using transformation of the original system, a new robust proportional integral filter (RPIF having an error variance with an optimized guaranteed upper bound for any allowed uncertainty is proposed to improve robust estimation of unknown time-varying faults and to improve robustness against uncertainties. In this study, the minimization problem of the upper bound of the estimation error variance is formulated as a convex optimization problem subject to linear matrix inequalities (LMI for all admissible uncertainties. The proportional and the integral gains are optimally chosen by solving the convex optimization problem. Simulation results are given in order to illustrate the performance of the proposed filter, in particular to solve the problem of joint fault and state estimation.

  8. On the real-time estimation of the wheel-rail contact force by means of a new nonlinear estimator design model

    Science.gov (United States)

    Strano, Salvatore; Terzo, Mario

    2018-05-01

    The dynamics of the railway vehicles is strongly influenced by the interaction between the wheel and the rail. This kind of contact is affected by several conditioning factors such as vehicle speed, wear, adhesion level and, moreover, it is nonlinear. As a consequence, the modelling and the observation of this kind of phenomenon are complex tasks but, at the same time, they constitute a fundamental step for the estimation of the adhesion level or for the vehicle condition monitoring. This paper presents a novel technique for the real time estimation of the wheel-rail contact forces based on an estimator design model that takes into account the nonlinearities of the interaction by means of a fitting model functional to reproduce the contact mechanics in a wide range of slip and to be easily integrated in a complete model based estimator for railway vehicle.

  9. Parameter Estimation of a Closed Loop Coupled Tank Time Varying System using Recursive Methods

    International Nuclear Information System (INIS)

    Basir, Siti Nora; Yussof, Hanafiah; Shamsuddin, Syamimi; Selamat, Hazlina; Zahari, Nur Ismarrubie

    2013-01-01

    This project investigates the direct identification of closed loop plant using discrete-time approach. The uses of Recursive Least Squares (RLS), Recursive Instrumental Variable (RIV) and Recursive Instrumental Variable with Centre-Of-Triangle (RIV + COT) in the parameter estimation of closed loop time varying system have been considered. The algorithms were applied in a coupled tank system that employs covariance resetting technique where the time of parameter changes occur is unknown. The performances of all the parameter estimation methods, RLS, RIV and RIV + COT were compared. The estimation of the system whose output was corrupted with white and coloured noises were investigated. Covariance resetting technique successfully executed when the parameters change. RIV + COT gives better estimates than RLS and RIV in terms of convergence and maximum overshoot

  10. Fast Estimation Method of Space-Time Two-Dimensional Positioning Parameters Based on Hadamard Product

    Directory of Open Access Journals (Sweden)

    Haiwen Li

    2018-01-01

    Full Text Available The estimation speed of positioning parameters determines the effectiveness of the positioning system. The time of arrival (TOA and direction of arrival (DOA parameters can be estimated by the space-time two-dimensional multiple signal classification (2D-MUSIC algorithm for array antenna. However, this algorithm needs much time to complete the two-dimensional pseudo spectral peak search, which makes it difficult to apply in practice. Aiming at solving this problem, a fast estimation method of space-time two-dimensional positioning parameters based on Hadamard product is proposed in orthogonal frequency division multiplexing (OFDM system, and the Cramer-Rao bound (CRB is also presented. Firstly, according to the channel frequency domain response vector of each array, the channel frequency domain estimation vector is constructed using the Hadamard product form containing location information. Then, the autocorrelation matrix of the channel response vector for the extended array element in frequency domain and the noise subspace are calculated successively. Finally, by combining the closed-form solution and parameter pairing, the fast joint estimation for time delay and arrival direction is accomplished. The theoretical analysis and simulation results show that the proposed algorithm can significantly reduce the computational complexity and guarantee that the estimation accuracy is not only better than estimating signal parameters via rotational invariance techniques (ESPRIT algorithm and 2D matrix pencil (MP algorithm but also close to 2D-MUSIC algorithm. Moreover, the proposed algorithm also has certain adaptability to multipath environment and effectively improves the ability of fast acquisition of location parameters.

  11. Nonparametric Estimation of Interval Reliability for Discrete-Time Semi-Markov Systems

    DEFF Research Database (Denmark)

    Georgiadis, Stylianos; Limnios, Nikolaos

    2016-01-01

    In this article, we consider a repairable discrete-time semi-Markov system with finite state space. The measure of the interval reliability is given as the probability of the system being operational over a given finite-length time interval. A nonparametric estimator is proposed for the interval...

  12. Between-Trial Forgetting Due to Interference and Time in Motor Adaptation.

    Directory of Open Access Journals (Sweden)

    Sungshin Kim

    Full Text Available Learning a motor task with temporally spaced presentations or with other tasks intermixed between presentations reduces performance during training, but can enhance retention post training. These two effects are known as the spacing and contextual interference effect, respectively. Here, we aimed at testing a unifying hypothesis of the spacing and contextual interference effects in visuomotor adaptation, according to which forgetting between trials due to either spaced presentations or interference by another task will promote between-trial forgetting, which will depress performance during acquisition, but will promote retention. We first performed an experiment with three visuomotor adaptation conditions: a short inter-trial-interval (ITI condition (SHORT-ITI; a long ITI condition (LONG-ITI; and an alternating condition with two alternated opposite tasks (ALT, with the same single-task ITI as in LONG-ITI. In the SHORT-ITI condition, there was fastest increase in performance during training and largest immediate forgetting in the retention tests. In contrast, in the ALT condition, there was slowest increase in performance during training and little immediate forgetting in the retention tests. Compared to these two conditions, in the LONG-ITI, we found intermediate increase in performance during training and intermediate immediate forgetting. To account for these results, we fitted to the data six possible adaptation models with one or two time scales, and with interference in the fast, or in the slow, or in both time scales. Model comparison confirmed that two time scales and some degree of interferences in either time scale are needed to account for our experimental results. In summary, our results suggest that retention following adaptation is modulated by the degree of between-trial forgetting, which is due to time-based decay in single adaptation task and interferences in multiple adaptation tasks.

  13. Eye movements discriminate fatigue due to chronotypical factors and time spent on task--a double dissociation.

    Directory of Open Access Journals (Sweden)

    Dario Cazzoli

    Full Text Available Systematic differences in circadian rhythmicity are thought to be a substantial factor determining inter-individual differences in fatigue and cognitive performance. The synchronicity effect (when time of testing coincides with the respective circadian peak period seems to play an important role. Eye movements have been shown to be a reliable indicator of fatigue due to sleep deprivation or time spent on cognitive tasks. However, eye movements have not been used so far to investigate the circadian synchronicity effect and the resulting differences in fatigue. The aim of the present study was to assess how different oculomotor parameters in a free visual exploration task are influenced by: a fatigue due to chronotypical factors (being a 'morning type' or an 'evening type'; b fatigue due to the time spent on task. Eighteen healthy participants performed a free visual exploration task of naturalistic pictures while their eye movements were recorded. The task was performed twice, once at their optimal and once at their non-optimal time of the day. Moreover, participants rated their subjective fatigue. The non-optimal time of the day triggered a significant and stable increase in the mean visual fixation duration during the free visual exploration task for both chronotypes. The increase in the mean visual fixation duration correlated with the difference in subjectively perceived fatigue at optimal and non-optimal times of the day. Conversely, the mean saccadic speed significantly and progressively decreased throughout the duration of the task, but was not influenced by the optimal or non-optimal time of the day for both chronotypes. The results suggest that different oculomotor parameters are discriminative for fatigue due to different sources. A decrease in saccadic speed seems to reflect fatigue due to time spent on task, whereas an increase in mean fixation duration a lack of synchronicity between chronotype and time of the day.

  14. Risk Reduction Effects Due to the Start Time Extension of EDGs in OPR-1000

    International Nuclear Information System (INIS)

    Lim, Ho-Gon; Yang, Joon-Eon; Hwang, Mee-Jeong

    2006-01-01

    Under the condition that the ECCS rule in Korea will be revised based on the new U.S. 10 CFR 50.46, the risk impact due to the EDG start time extension is analyzed in the present study. This paper is composed of 6 sections. In the section 2, the LOCA break size that cannot be mitigable under the condition of extended EDG start time is obtained from the thermal hydraulic analysis. The section 3 discusses the frequency of the immitigable LOCA and the probability of the LOOP given a LOCA. In the section 4, the effect of the EDG start time extension on its failure probability is discussed with a qualitative manner. Finally, the whole risk change due to the EDG start time extension is calculated in the section 5 with the conclusions given in the section 6

  15. Trend Estimation and Regression Analysis in Climatological Time Series: An Application of Structural Time Series Models and the Kalman Filter.

    Science.gov (United States)

    Visser, H.; Molenaar, J.

    1995-05-01

    The detection of trends in climatological data has become central to the discussion on climate change due to the enhanced greenhouse effect. To prove detection, a method is needed (i) to make inferences on significant rises or declines in trends, (ii) to take into account natural variability in climate series, and (iii) to compare output from GCMs with the trends in observed climate data. To meet these requirements, flexible mathematical tools are needed. A structural time series model is proposed with which a stochastic trend, a deterministic trend, and regression coefficients can be estimated simultaneously. The stochastic trend component is described using the class of ARIMA models. The regression component is assumed to be linear. However, the regression coefficients corresponding with the explanatory variables may be time dependent to validate this assumption. The mathematical technique used to estimate this trend-regression model is the Kaiman filter. The main features of the filter are discussed.Examples of trend estimation are given using annual mean temperatures at a single station in the Netherlands (1706-1990) and annual mean temperatures at Northern Hemisphere land stations (1851-1990). The inclusion of explanatory variables is shown by regressing the latter temperature series on four variables: Southern Oscillation index (SOI), volcanic dust index (VDI), sunspot numbers (SSN), and a simulated temperature signal, induced by increasing greenhouse gases (GHG). In all analyses, the influence of SSN on global temperatures is found to be negligible. The correlations between temperatures and SOI and VDI appear to be negative. For SOI, this correlation is significant, but for VDI it is not, probably because of a lack of volcanic eruptions during the sample period. The relation between temperatures and GHG is positive, which is in agreement with the hypothesis of a warming climate because of increasing levels of greenhouse gases. The prediction performance of

  16. Estimating medication stopping fraction and real-time prevalence of drug use in pharmaco-epidemiologic databases. An application of the reverse waiting time distribution

    DEFF Research Database (Denmark)

    Støvring, Henrik; Pottegård, Anton; Hallas, Jesper

    2017-01-01

    Purpose: To introduce the reverse waiting time distribution (WTD) and show how it can be used to estimate stopping fractions and real-time prevalence of treatment in pharmacoepidemiological studies. Methods: The reverse WTD is the distribution of time from the last dispensed prescription of each......-hoc decision rules for automated implementations, and it yields estimates of real-time prevalence....... patient within a time window to the end of it. It is a mirrored version of the ordinary WTD, which considers the first dispensed prescription of patients within a time window. Based on renewal process theory, the reverse WTD can be analyzed as an ordinary WTD with maximum likelihood estimation. Based...

  17. Impact of time displaced precipitation estimates for on-line updated models

    DEFF Research Database (Denmark)

    Borup, Morten; Grum, Morten; Mikkelsen, Peter Steen

    2012-01-01

    When an online runoff model is updated from system measurements the requirements to the precipitation estimates change. Using rain gauge data as precipitation input there will be a displacement between the time where the rain intensity hits the gauge and the time where the rain hits the actual...

  18. DeepTravel: a Neural Network Based Travel Time Estimation Model with Auxiliary Supervision

    OpenAIRE

    Zhang, Hanyuan; Wu, Hao; Sun, Weiwei; Zheng, Baihua

    2018-01-01

    Estimating the travel time of a path is of great importance to smart urban mobility. Existing approaches are either based on estimating the time cost of each road segment which are not able to capture many cross-segment complex factors, or designed heuristically in a non-learning-based way which fail to utilize the existing abundant temporal labels of the data, i.e., the time stamp of each trajectory point. In this paper, we leverage on new development of deep neural networks and propose a no...

  19. Real-Time Earthquake Intensity Estimation Using Streaming Data Analysis of Social and Physical Sensors

    Science.gov (United States)

    Kropivnitskaya, Yelena; Tiampo, Kristy F.; Qin, Jinhui; Bauer, Michael A.

    2017-06-01

    Earthquake intensity is one of the key components of the decision-making process for disaster response and emergency services. Accurate and rapid intensity calculations can help to reduce total loss and the number of casualties after an earthquake. Modern intensity assessment procedures handle a variety of information sources, which can be divided into two main categories. The first type of data is that derived from physical sensors, such as seismographs and accelerometers, while the second type consists of data obtained from social sensors, such as witness observations of the consequences of the earthquake itself. Estimation approaches using additional data sources or that combine sources from both data types tend to increase intensity uncertainty due to human factors and inadequate procedures for temporal and spatial estimation, resulting in precision errors in both time and space. Here we present a processing approach for the real-time analysis of streams of data from both source types. The physical sensor data is acquired from the U.S. Geological Survey (USGS) seismic network in California and the social sensor data is based on Twitter user observations. First, empirical relationships between tweet rate and observed Modified Mercalli Intensity (MMI) are developed using data from the M6.0 South Napa, CAF earthquake that occurred on August 24, 2014. Second, the streams of both data types are analyzed together in simulated real-time to produce one intensity map. The second implementation is based on IBM InfoSphere Streams, a cloud platform for real-time analytics of big data. To handle large processing workloads for data from various sources, it is deployed and run on a cloud-based cluster of virtual machines. We compare the quality and evolution of intensity maps from different data sources over 10-min time intervals immediately following the earthquake. Results from the joint analysis shows that it provides more complete coverage, with better accuracy and higher

  20. Eliminating bias in rainfall estimates from microwave links due to antenna wetting

    Science.gov (United States)

    Fencl, Martin; Rieckermann, Jörg; Bareš, Vojtěch

    2014-05-01

    Commercial microwave links (MWLs) are point-to-point radio systems which are widely used in telecommunication systems. They operate at frequencies where the transmitted power is mainly disturbed by precipitation. Thus, signal attenuation from MWLs can be used to estimate path-averaged rain rates, which is conceptually very promising, since MWLs cover about 20 % of surface area. Unfortunately, MWL rainfall estimates are often positively biased due to additional attenuation caused by antenna wetting. To correct MWL observations a posteriori to reduce the wet antenna effect (WAE), both empirically and physically based models have been suggested. However, it is challenging to calibrate these models, because the wet antenna attenuation depends both on the MWL properties (frequency, type of antennas, shielding etc.) and different climatic factors (temperature, due point, wind velocity and direction, etc.). Instead, it seems straight forward to keep antennas dry by shielding them. In this investigation we compare the effectiveness of antenna shielding to model-based corrections to reduce the WAE. The experimental setup, located in Dübendorf-Switzerland, consisted of 1.85-km long commercial dual-polarization microwave link at 38 GHz and 5 optical disdrometers. The MWL was operated without shielding in the period from March to October 2011 and with shielding from October 2011 to July 2012. This unique experimental design made it possible to identify the attenuation due to antenna wetting, which can be computed as the difference between the measured and theoretical attenuation. The theoretical path-averaged attenuation was calculated from the path-averaged drop size distribution. During the unshielded periods, the total bias caused by WAE was 0.74 dB, which was reduced by shielding to 0.39 dB for the horizontal polarization (vertical: reduction from 0.96 dB to 0.44 dB). Interestingly, the model-based correction (Schleiss et al. 2013) was more effective because it reduced

  1. On the estimation of wall pressure coherence using time-resolved tomographic PIV

    Science.gov (United States)

    Pröbsting, Stefan; Scarano, Fulvio; Bernardini, Matteo; Pirozzoli, Sergio

    2013-07-01

    Three-dimensional time-resolved velocity field measurements are obtained using a high-speed tomographic Particle Image Velocimetry (PIV) system on a fully developed flat plate turbulent boundary layer for the estimation of wall pressure fluctuations. The work focuses on the applicability of tomographic PIV to compute the coherence of pressure fluctuations, with attention to the estimation of the stream and spanwise coherence length. The latter is required for estimations of aeroacoustic noise radiation by boundary layers and trailing edge flows, but is also of interest for vibro-structural problems. The pressure field is obtained by solving the Poisson equation for incompressible flows, where the source terms are provided by time-resolved velocity field measurements. Measured 3D velocity data is compared to results obtained from planar PIV, and a Direct Numerical Simulation (DNS) at similar Reynolds number. An improved method for the estimation of the material based on a least squares estimator of the velocity derivative along a particle trajectory is proposed and applied. Computed surface pressure fluctuations are further verified by means of simultaneous measurements by a pinhole microphone and compared to the DNS results and a semi-empirical model available from literature. The correlation coefficient for the reconstructed pressure time series with respect to pinhole microphone measurements attains approximately 0.5 for the band-pass filtered signal over the range of frequencies resolved by the velocity field measurements. Scaled power spectra of the pressure at a single point compare favorably to the DNS results and those available from literature. Finally, the coherence of surface pressure fluctuations and the resulting span- and streamwise coherence lengths are estimated and compared to semi-empirical models and DNS results.

  2. An ELM-Based Approach for Estimating Train Dwell Time in Urban Rail Traffic

    Directory of Open Access Journals (Sweden)

    Wen-jun Chu

    2015-01-01

    Full Text Available Dwell time estimation plays an important role in the operation of urban rail system. On this specific problem, a range of models based on either polynomial regression or microsimulation have been proposed. However, the generalization performance of polynomial regression models is limited and the accuracy of existing microsimulation models is unstable. In this paper, a new dwell time estimation model based on extreme learning machine (ELM is proposed. The underlying factors that may affect urban rail dwell time are analyzed first. Then, the relationships among different factors are extracted and modeled by ELM neural networks, on basis of which an overall estimation model is proposed. At last, a set of observed data from Beijing subway is used to illustrate the proposed method and verify its overall performance.

  3. Evidence Estimation for Bayesian Partially Observed MRFs

    NARCIS (Netherlands)

    Chen, Y.; Welling, M.

    2013-01-01

    Bayesian estimation in Markov random fields is very hard due to the intractability of the partition function. The introduction of hidden units makes the situation even worse due to the presence of potentially very many modes in the posterior distribution. For the first time we propose a

  4. Single-machine scheduling with release dates, due dates and family setup times

    NARCIS (Netherlands)

    Schutten, Johannes M.J.; van de Velde, S.L.; van de Velde, S.L.; Zijm, Willem H.M.

    1996-01-01

    We address the NP-hard problem of scheduling n independent jobs with release dates, due dates, and family setup times on a single machine to minimize the maximum lateness. This problem arises from the constant tug-of-war going on in manufacturing between efficient production and delivery

  5. The Time Is Up: Compression of Visual Time Interval Estimations of Bimodal Aperiodic Patterns

    Science.gov (United States)

    Duarte, Fabiola; Lemus, Luis

    2017-01-01

    The ability to estimate time intervals subserves many of our behaviors and perceptual experiences. However, it is not clear how aperiodic (AP) stimuli affect our perception of time intervals across sensory modalities. To address this question, we evaluated the human capacity to discriminate between two acoustic (A), visual (V) or audiovisual (AV) time intervals of trains of scattered pulses. We first measured the periodicity of those stimuli and then sought for correlations with the accuracy and reaction times (RTs) of the subjects. We found that, for all time intervals tested in our experiment, the visual system consistently perceived AP stimuli as being shorter than the periodic (P) ones. In contrast, such a compression phenomenon was not apparent during auditory trials. Our conclusions are: first, the subjects exposed to P stimuli are more likely to measure their durations accurately. Second, perceptual time compression occurs for AP visual stimuli. Lastly, AV discriminations are determined by A dominance rather than by AV enhancement. PMID:28848406

  6. On the background estimation by time slides in a network of gravitational wave detectors

    International Nuclear Information System (INIS)

    Was, Michal; Bizouard, Marie-Anne; Brisson, Violette; Cavalier, Fabien; Davier, Michel; Hello, Patrice; Leroy, Nicolas; Robinet, Florent; Vavoulidis, Miltiadis

    2010-01-01

    Time shifting the outputs of gravitational wave detectors operating in coincidence is a convenient way to estimate the background in a search for short-duration signals. However, this procedure is limited as increasing indefinitely the number of time shifts does not provide better estimates. We show that the false alarm rate estimation error saturates with the number of time shifts. In particular, for detectors with very different trigger rates, this error saturates at a large value. Explicit computations are done for two detectors, and for three detectors where the detection statistic relies on the logical 'OR' of the coincidences of the three couples in the network.

  7. On the background estimation by time slides in a network of gravitational wave detectors

    Energy Technology Data Exchange (ETDEWEB)

    Was, Michal; Bizouard, Marie-Anne; Brisson, Violette; Cavalier, Fabien; Davier, Michel; Hello, Patrice; Leroy, Nicolas; Robinet, Florent; Vavoulidis, Miltiadis, E-mail: mwas@lal.in2p3.f [LAL, Univ. Paris-Sud, CNRS/IN2P3, Orsay (France)

    2010-01-07

    Time shifting the outputs of gravitational wave detectors operating in coincidence is a convenient way to estimate the background in a search for short-duration signals. However, this procedure is limited as increasing indefinitely the number of time shifts does not provide better estimates. We show that the false alarm rate estimation error saturates with the number of time shifts. In particular, for detectors with very different trigger rates, this error saturates at a large value. Explicit computations are done for two detectors, and for three detectors where the detection statistic relies on the logical 'OR' of the coincidences of the three couples in the network.

  8. Using Indirect Turbulence Measurements for Real-Time Parameter Estimation in Turbulent Air

    Science.gov (United States)

    Martos, Borja; Morelli, Eugene A.

    2012-01-01

    The use of indirect turbulence measurements for real-time estimation of parameters in a linear longitudinal dynamics model in atmospheric turbulence was studied. It is shown that measuring the atmospheric turbulence makes it possible to treat the turbulence as a measured explanatory variable in the parameter estimation problem. Commercial off-the-shelf sensors were researched and evaluated, then compared to air data booms. Sources of colored noise in the explanatory variables resulting from typical turbulence measurement techniques were identified and studied. A major source of colored noise in the explanatory variables was identified as frequency dependent upwash and time delay. The resulting upwash and time delay corrections were analyzed and compared to previous time shift dynamic modeling research. Simulation data as well as flight test data in atmospheric turbulence were used to verify the time delay behavior. Recommendations are given for follow on flight research and instrumentation.

  9. Power System Real-Time Monitoring by Using PMU-Based Robust State Estimation Method

    DEFF Research Database (Denmark)

    Zhao, Junbo; Zhang, Gexiang; Das, Kaushik

    2016-01-01

    Accurate real-time states provided by the state estimator are critical for power system reliable operation and control. This paper proposes a novel phasor measurement unit (PMU)-based robust state estimation method (PRSEM) to real-time monitor a power system under different operation conditions...... the system real-time states with good robustness and can address several kinds of BD.......-based bad data (BD) detection method, which can handle the smearing effect and critical measurement errors, is presented. We evaluate PRSEM by using IEEE benchmark test systems and a realistic utility system. The numerical results indicate that, in short computation time, PRSEM can effectively track...

  10. The estimation of turnover time in the Japan Sea bottom water by 129I

    International Nuclear Information System (INIS)

    Suzuki, Takashi; Togawa, Orihiko; Minakawa, Masayuki

    2010-01-01

    It is well known that the Japan Sea is sensitive for the environment such a global warming. To understand the oceanic circulation in the Japan Sea, we estimated a turnover time and a potential formation rate of the Japan Sea bottom water (JSBW) using an oceanographic tracer of 129 I. The turnover time of JSBW was calculated based on the increased concentration during the nuclear era. The turnover time was estimated to be 180 - 210 years. The potential formation rate of JSBW is calculated based on the existence of the anthropogenic 129 I in the JSBW. The potential formation rate of JSBW is estimated to be (3.6-4.1) x 10 12 m 3 /y which is consistent with another estimation and is about quarter of that of the upper Japan Sea proper water. (author)

  11. State estimation for discrete-time Markovian jumping neural networks with mixed mode-dependent delays

    International Nuclear Information System (INIS)

    Liu Yurong; Wang Zidong; Liu Xiaohui

    2008-01-01

    In this Letter, we investigate the state estimation problem for a new class of discrete-time neural networks with Markovian jumping parameters as well as mode-dependent mixed time-delays. The parameters of the discrete-time neural networks are subject to the switching from one mode to another at different times according to a Markov chain, and the mixed time-delays consist of both discrete and distributed delays that are dependent on the Markovian jumping mode. New techniques are developed to deal with the mixed time-delays in the discrete-time setting, and a novel Lyapunov-Krasovskii functional is put forward to reflect the mode-dependent time-delays. Sufficient conditions are established in terms of linear matrix inequalities (LMIs) that guarantee the existence of the state estimators. We show that both the existence conditions and the explicit expression of the desired estimator can be characterized in terms of the solution to an LMI. A numerical example is exploited to show the usefulness of the derived LMI-based conditions

  12. Instrumental variables estimation of exposure effects on a time-to-event endpoint using structural cumulative survival models.

    Science.gov (United States)

    Martinussen, Torben; Vansteelandt, Stijn; Tchetgen Tchetgen, Eric J; Zucker, David M

    2017-12-01

    The use of instrumental variables for estimating the effect of an exposure on an outcome is popular in econometrics, and increasingly so in epidemiology. This increasing popularity may be attributed to the natural occurrence of instrumental variables in observational studies that incorporate elements of randomization, either by design or by nature (e.g., random inheritance of genes). Instrumental variables estimation of exposure effects is well established for continuous outcomes and to some extent for binary outcomes. It is, however, largely lacking for time-to-event outcomes because of complications due to censoring and survivorship bias. In this article, we make a novel proposal under a class of structural cumulative survival models which parameterize time-varying effects of a point exposure directly on the scale of the survival function; these models are essentially equivalent with a semi-parametric variant of the instrumental variables additive hazards model. We propose a class of recursive instrumental variable estimators for these exposure effects, and derive their large sample properties along with inferential tools. We examine the performance of the proposed method in simulation studies and illustrate it in a Mendelian randomization study to evaluate the effect of diabetes on mortality using data from the Health and Retirement Study. We further use the proposed method to investigate potential benefit from breast cancer screening on subsequent breast cancer mortality based on the HIP-study. © 2017, The International Biometric Society.

  13. Observing expertise-related actions leads to perfect time flow estimations.

    Directory of Open Access Journals (Sweden)

    Yin-Hua Chen

    Full Text Available The estimation of the time of exposure of a picture portraying an action increases as a function of the amount of movement implied in the action represented. This effect suggests that the perceiver creates an internal embodiment of the action observed as if internally simulating the entire movement sequence. Little is known however about the timing accuracy of these internal action simulations, specifically whether they are affected by the level of familiarity and experience that the observer has of the action. In this study we asked professional pianists to reproduce different durations of exposure (shorter or longer than one second of visual displays both specific (a hand in piano-playing action and non-specific to their domain of expertise (a hand in finger-thumb opposition and scrambled-pixels and compared their performance with non-pianists. Pianists outperformed non-pianists independently of the time of exposure of the stimuli; remarkably the group difference was particularly magnified by the pianists' enhanced accuracy and stability only when observing the hand in the act of playing the piano. These results for the first time provide evidence that through musical training, pianists create a selective and self-determined dynamic internal representation of an observed movement that allows them to estimate precisely its temporal duration.

  14. Single-machine scheduling with release dates, due dates, and family setup times

    NARCIS (Netherlands)

    J.M.J. Schutten (Marco); S.L. van de Velde (Steef); W.H.M. Zijm

    1996-01-01

    textabstractWe address the NP-hard problem of scheduling n independent jobs with release dates, due dates, and family setup times on a single machine to minimize the maximum lateness. This problem arises from the constant tug-of-war going on in manufacturing between efficient production and delivery

  15. Energy and round time estimation method for mobile wireless sensor networks

    International Nuclear Information System (INIS)

    Ismat, M.; Qureshi, R.; Imam, M.U.

    2018-01-01

    Clustered WSN (Wireless Sensor Networks) is a hierarchical network structure that conserves energy by distributing the task of sensing and data transfer to destination among the non-CH (Cluster-Head) and CH (Cluster Head) node in a cluster. In clustered MWSN (Mobile Wireless Sensor Network), cluster maintenance to increase at a reception at the destination during communication operation is difficult due to the movement of CHs and non-CH nodes in and out of the cluster. To conserve energy and increased data transfer to the destination, it is necessary to find the duration after which sensor node’s role should be changed from CH to non-CH and vice-versa. In this paper, we have proposed an energy independent round time scheme to identify the duration after which re-clustering procedure should be invoked for changing roles of sensor nodes as CHs and associated nodes to conserve energy and increased data delivery. This depends on the dissemination interval of the sensor nodes rather than sensor node’s energy. We have also provided a complete analytical estimate of network energy consumption with energy consumed in every phase of a around. (author)

  16. Multifractals embedded in short time series: An unbiased estimation of probability moment

    Science.gov (United States)

    Qiu, Lu; Yang, Tianguang; Yin, Yanhua; Gu, Changgui; Yang, Huijie

    2016-12-01

    An exact estimation of probability moments is the base for several essential concepts, such as the multifractals, the Tsallis entropy, and the transfer entropy. By means of approximation theory we propose a new method called factorial-moment-based estimation of probability moments. Theoretical prediction and computational results show that it can provide us an unbiased estimation of the probability moments of continuous order. Calculations on probability redistribution model verify that it can extract exactly multifractal behaviors from several hundred recordings. Its powerfulness in monitoring evolution of scaling behaviors is exemplified by two empirical cases, i.e., the gait time series for fast, normal, and slow trials of a healthy volunteer, and the closing price series for Shanghai stock market. By using short time series with several hundred lengths, a comparison with the well-established tools displays significant advantages of its performance over the other methods. The factorial-moment-based estimation can evaluate correctly the scaling behaviors in a scale range about three generations wider than the multifractal detrended fluctuation analysis and the basic estimation. The estimation of partition function given by the wavelet transform modulus maxima has unacceptable fluctuations. Besides the scaling invariance focused in the present paper, the proposed factorial moment of continuous order can find its various uses, such as finding nonextensive behaviors of a complex system and reconstructing the causality relationship network between elements of a complex system.

  17. Estimating time-varying exposure-outcome associations using case-control data: logistic and case-cohort analyses.

    Science.gov (United States)

    Keogh, Ruth H; Mangtani, Punam; Rodrigues, Laura; Nguipdop Djomo, Patrick

    2016-01-05

    Traditional analyses of standard case-control studies using logistic regression do not allow estimation of time-varying associations between exposures and the outcome. We present two approaches which allow this. The motivation is a study of vaccine efficacy as a function of time since vaccination. Our first approach is to estimate time-varying exposure-outcome associations by fitting a series of logistic regressions within successive time periods, reusing controls across periods. Our second approach treats the case-control sample as a case-cohort study, with the controls forming the subcohort. In the case-cohort analysis, controls contribute information at all times they are at risk. Extensions allow left truncation, frequency matching and, using the case-cohort analysis, time-varying exposures. Simulations are used to investigate the methods. The simulation results show that both methods give correct estimates of time-varying effects of exposures using standard case-control data. Using the logistic approach there are efficiency gains by reusing controls over time and care should be taken over the definition of controls within time periods. However, using the case-cohort analysis there is no ambiguity over the definition of controls. The performance of the two analyses is very similar when controls are used most efficiently under the logistic approach. Using our methods, case-control studies can be used to estimate time-varying exposure-outcome associations where they may not previously have been considered. The case-cohort analysis has several advantages, including that it allows estimation of time-varying associations as a continuous function of time, while the logistic regression approach is restricted to assuming a step function form for the time-varying association.

  18. GLODEP2: a computer model for estimating gamma dose due to worldwide fallout of radioactive debris

    International Nuclear Information System (INIS)

    Edwards, L.L.; Harvey, T.F.; Peterson, K.R.

    1984-03-01

    The GLODEP2 computer code provides estimates of the surface deposition of worldwide radioactivity and the gamma-ray dose to man from intermediate and long-term fallout. The code is based on empirical models derived primarily from injection-deposition experience gained from the US and USSR nuclear tests in 1958. Under the assumption that a nuclear power facility is destroyed and that its debris behaves in the same manner as the radioactive cloud produced by the nuclear weapon that attached the facility, predictions are made for the gamma does from this source of radioactivity. As a comparison study the gamma dose due to the atmospheric nuclear tests from the period of 1951 to 1962 has been computed. The computed and measured values from Grove, UK and Chiba, Japan agree to within a few percent. The global deposition of radioactivity and resultant gamma dose from a hypothetical strategic nuclear exchange between the US and the USSR is reported. Of the assumed 5300 Mton in the exchange, 2031 Mton of radioactive debris is injected in the atmosphere. The highest estimated average whole body total integrated dose over 50 years (assuming no reduction by sheltering or weathering) is 23 rem in the 30 to 50 degree latitude band. If the attack included a 100 GW(e) nuclear power industry as targets in the US, this dose is increased to 84.6 rem. Hotspots due to rainfall could increase these values by factors of 10 to 50

  19. Improvement of radiation dose estimation due to nuclear accidents using deep neural network and GPU

    Energy Technology Data Exchange (ETDEWEB)

    Desterro, Filipe S.M.; Almeida, Adino A.H.; Pereira, Claudio M.N.A., E-mail: filipesantana18@gmail.com, E-mail: adino@ien.gov.br, E-mail: cmcoelho@ien.gov.br [Instituto de Engenharia Nuclear (IEN/CNEN-RJ), Rio de Janeiro, RJ (Brazil)

    2017-07-01

    Recently, the use of mobile devices has been proposed for dose assessment during nuclear accidents. The idea is to support field teams, providing an approximated estimation of the dose distribution map in the vicinity of the nuclear power plant (NPP), without needing to be connected to the NPP systems. In order to provide such stand-alone execution, the use of artificial neural networks (ANN) has been proposed in substitution of the complex and time consuming physical models executed by the atmospheric dispersion radionuclide (ADR) system. One limitation observed on such approach is the very time-consuming training of the ANNs. Moreover, if the number of input parameters increases the performance of standard ANNs, like Multilayer-Perceptron (MLP) with backpropagation training, is affected leading to unreasonable training time. To improve learning, allowing better dose estimations, more complex ANN architectures are required. ANNs with many layers (much more than a typical number of layers), referred to as Deep Neural Networks (DNN), for example, have demonstrating to achieve better results. On the other hand, the training of such ANNs is very much slow. In order to allow the use of such DNNs in a reasonable training time, a parallel programming solution, using Graphic Processing Units (GPU) and Computing Unified Device Architecture (CUDA) is proposed. This work focuses on the study of computational technologies for improvement of the ANNs to be used in the mobile application, as well as their training algorithms. (author)

  20. Improvement of radiation dose estimation due to nuclear accidents using deep neural network and GPU

    International Nuclear Information System (INIS)

    Desterro, Filipe S.M.; Almeida, Adino A.H.; Pereira, Claudio M.N.A.

    2017-01-01

    Recently, the use of mobile devices has been proposed for dose assessment during nuclear accidents. The idea is to support field teams, providing an approximated estimation of the dose distribution map in the vicinity of the nuclear power plant (NPP), without needing to be connected to the NPP systems. In order to provide such stand-alone execution, the use of artificial neural networks (ANN) has been proposed in substitution of the complex and time consuming physical models executed by the atmospheric dispersion radionuclide (ADR) system. One limitation observed on such approach is the very time-consuming training of the ANNs. Moreover, if the number of input parameters increases the performance of standard ANNs, like Multilayer-Perceptron (MLP) with backpropagation training, is affected leading to unreasonable training time. To improve learning, allowing better dose estimations, more complex ANN architectures are required. ANNs with many layers (much more than a typical number of layers), referred to as Deep Neural Networks (DNN), for example, have demonstrating to achieve better results. On the other hand, the training of such ANNs is very much slow. In order to allow the use of such DNNs in a reasonable training time, a parallel programming solution, using Graphic Processing Units (GPU) and Computing Unified Device Architecture (CUDA) is proposed. This work focuses on the study of computational technologies for improvement of the ANNs to be used in the mobile application, as well as their training algorithms. (author)

  1. Estimation of resource savings due to fly ash utilization in road construction

    Energy Technology Data Exchange (ETDEWEB)

    Kumar, Subodh; Patil, C.B. [Centre for Energy Studies, Indian Institute of Technology, New Delhi 110016 (India)

    2006-08-15

    A methodology for estimation of natural resource savings due to fly ash utilization in road construction in India is presented. Analytical expressions for the savings of various resources namely soil, stone aggregate, stone chips, sand and cement in the embankment, granular sub-base (GSB), water bound macadam (WBM) and pavement quality concrete (PQC) layers of fly ash based road formation with flexible and rigid pavements of a given geometry have been developed. The quantity of fly ash utilized in these layers of different pavements has also been quantified. In the present study, the maximum amount of resource savings is found in GSB followed by WBM and other layers of pavement. The soil quantity saved increases asymptotically with the rise in the embankment height. The results of financial analysis based on Indian fly ash based road construction cost data indicate that the savings in construction cost decrease with the lead and the investment on this alternative is found to be financially attractive only for a lead less than 60 and 90km for flexible and rigid pavements, respectively. (author)

  2. Statistical methods of parameter estimation for deterministically chaotic time series

    Science.gov (United States)

    Pisarenko, V. F.; Sornette, D.

    2004-03-01

    We discuss the possibility of applying some standard statistical methods (the least-square method, the maximum likelihood method, and the method of statistical moments for estimation of parameters) to deterministically chaotic low-dimensional dynamic system (the logistic map) containing an observational noise. A “segmentation fitting” maximum likelihood (ML) method is suggested to estimate the structural parameter of the logistic map along with the initial value x1 considered as an additional unknown parameter. The segmentation fitting method, called “piece-wise” ML, is similar in spirit but simpler and has smaller bias than the “multiple shooting” previously proposed. Comparisons with different previously proposed techniques on simulated numerical examples give favorable results (at least, for the investigated combinations of sample size N and noise level). Besides, unlike some suggested techniques, our method does not require the a priori knowledge of the noise variance. We also clarify the nature of the inherent difficulties in the statistical analysis of deterministically chaotic time series and the status of previously proposed Bayesian approaches. We note the trade off between the need of using a large number of data points in the ML analysis to decrease the bias (to guarantee consistency of the estimation) and the unstable nature of dynamical trajectories with exponentially fast loss of memory of the initial condition. The method of statistical moments for the estimation of the parameter of the logistic map is discussed. This method seems to be the unique method whose consistency for deterministically chaotic time series is proved so far theoretically (not only numerically).

  3. H infinity Integrated Fault Estimation and Fault Tolerant Control of Discrete-time Piecewise Linear Systems

    DEFF Research Database (Denmark)

    Tabatabaeipour, Seyed Mojtaba; Bak, Thomas

    2012-01-01

    In this paper we consider the problem of fault estimation and accommodation for discrete time piecewise linear systems. A robust fault estimator is designed to estimate the fault such that the estimation error converges to zero and H∞ performance of the fault estimation is minimized. Then, the es...

  4. Secondary task for full flight simulation incorporating tasks that commonly cause pilot error: Time estimation

    Science.gov (United States)

    Rosch, E.

    1975-01-01

    The task of time estimation, an activity occasionally performed by pilots during actual flight, was investigated with the objective of providing human factors investigators with an unobtrusive and minimally loading additional task that is sensitive to differences in flying conditions and flight instrumentation associated with the main task of piloting an aircraft simulator. Previous research indicated that the duration and consistency of time estimates is associated with the cognitive, perceptual, and motor loads imposed by concurrent simple tasks. The relationships between the length and variability of time estimates and concurrent task variables under a more complex situation involving simulated flight were clarified. The wrap-around effect with respect to baseline duration, a consequence of mode switching at intermediate levels of concurrent task distraction, should contribute substantially to estimate variability and have a complex effect on the shape of the resulting distribution of estimates.

  5. Time-to-impact estimation in passive missile warning systems

    Science.gov (United States)

    Şahıngıl, Mehmet Cihan

    2017-05-01

    A missile warning system can detect the incoming missile threat(s) and automatically cue the other Electronic Attack (EA) systems in the suit, such as Directed Infrared Counter Measure (DIRCM) system and/or Counter Measure Dispensing System (CMDS). Most missile warning systems are currently based on passive sensor technology operating in either Solar Blind Ultraviolet (SBUV) or Midwave Infrared (MWIR) bands on which there is an intensive emission from the exhaust plume of the threatening missile. Although passive missile warning systems have some clear advantages over pulse-Doppler radar (PDR) based active missile warning systems, they show poorer performance in terms of time-to-impact (TTI) estimation which is critical for optimizing the countermeasures and also "passive kill assessment". In this paper, we consider this problem, namely, TTI estimation from passive measurements and present a TTI estimation scheme which can be used in passive missile warning systems. Our problem formulation is based on Extended Kalman Filter (EKF). The algorithm uses the area parameter of the threat plume which is derived from the used image frame.

  6. Fast and unbiased estimator of the time-dependent Hurst exponent

    Science.gov (United States)

    Pianese, Augusto; Bianchi, Sergio; Palazzo, Anna Maria

    2018-03-01

    We combine two existing estimators of the local Hurst exponent to improve both the goodness of fit and the computational speed of the algorithm. An application with simulated time series is implemented, and a Monte Carlo simulation is performed to provide evidence of the improvement.

  7. A multi-camera system for real-time pose estimation

    Science.gov (United States)

    Savakis, Andreas; Erhard, Matthew; Schimmel, James; Hnatow, Justin

    2007-04-01

    This paper presents a multi-camera system that performs face detection and pose estimation in real-time and may be used for intelligent computing within a visual sensor network for surveillance or human-computer interaction. The system consists of a Scene View Camera (SVC), which operates at a fixed zoom level, and an Object View Camera (OVC), which continuously adjusts its zoom level to match objects of interest. The SVC is set to survey the whole filed of view. Once a region has been identified by the SVC as a potential object of interest, e.g. a face, the OVC zooms in to locate specific features. In this system, face candidate regions are selected based on skin color and face detection is accomplished using a Support Vector Machine classifier. The locations of the eyes and mouth are detected inside the face region using neural network feature detectors. Pose estimation is performed based on a geometrical model, where the head is modeled as a spherical object that rotates upon the vertical axis. The triangle formed by the mouth and eyes defines a vertical plane that intersects the head sphere. By projecting the eyes-mouth triangle onto a two dimensional viewing plane, equations were obtained that describe the change in its angles as the yaw pose angle increases. These equations are then combined and used for efficient pose estimation. The system achieves real-time performance for live video input. Testing results assessing system performance are presented for both still images and video.

  8. Accurate estimation of camera shot noise in the real-time

    Science.gov (United States)

    Cheremkhin, Pavel A.; Evtikhiev, Nikolay N.; Krasnov, Vitaly V.; Rodin, Vladislav G.; Starikov, Rostislav S.

    2017-10-01

    Nowadays digital cameras are essential parts of various technological processes and daily tasks. They are widely used in optics and photonics, astronomy, biology and other various fields of science and technology such as control systems and video-surveillance monitoring. One of the main information limitations of photo- and videocameras are noises of photosensor pixels. Camera's photosensor noise can be divided into random and pattern components. Temporal noise includes random noise component while spatial noise includes pattern noise component. Temporal noise can be divided into signal-dependent shot noise and signal-nondependent dark temporal noise. For measurement of camera noise characteristics, the most widely used methods are standards (for example, EMVA Standard 1288). It allows precise shot and dark temporal noise measurement but difficult in implementation and time-consuming. Earlier we proposed method for measurement of temporal noise of photo- and videocameras. It is based on the automatic segmentation of nonuniform targets (ASNT). Only two frames are sufficient for noise measurement with the modified method. In this paper, we registered frames and estimated shot and dark temporal noises of cameras consistently in the real-time. The modified ASNT method is used. Estimation was performed for the cameras: consumer photocamera Canon EOS 400D (CMOS, 10.1 MP, 12 bit ADC), scientific camera MegaPlus II ES11000 (CCD, 10.7 MP, 12 bit ADC), industrial camera PixeLink PL-B781F (CMOS, 6.6 MP, 10 bit ADC) and video-surveillance camera Watec LCL-902C (CCD, 0.47 MP, external 8 bit ADC). Experimental dependencies of temporal noise on signal value are in good agreement with fitted curves based on a Poisson distribution excluding areas near saturation. Time of registering and processing of frames used for temporal noise estimation was measured. Using standard computer, frames were registered and processed during a fraction of second to several seconds only. Also the

  9. A Timing Estimation Method Based-on Skewness Analysis in Vehicular Wireless Networks.

    Science.gov (United States)

    Cui, Xuerong; Li, Juan; Wu, Chunlei; Liu, Jian-Hang

    2015-11-13

    Vehicle positioning technology has drawn more and more attention in vehicular wireless networks to reduce transportation time and traffic accidents. Nowadays, global navigation satellite systems (GNSS) are widely used in land vehicle positioning, but most of them are lack precision and reliability in situations where their signals are blocked. Positioning systems base-on short range wireless communication are another effective way that can be used in vehicle positioning or vehicle ranging. IEEE 802.11p is a new real-time short range wireless communication standard for vehicles, so a new method is proposed to estimate the time delay or ranges between vehicles based on the IEEE 802.11p standard which includes three main steps: cross-correlation between the received signal and the short preamble, summing up the correlated results in groups, and finding the maximum peak using a dynamic threshold based on the skewness analysis. With the range between each vehicle or road-side infrastructure, the position of neighboring vehicles can be estimated correctly. Simulation results were presented in the International Telecommunications Union (ITU) vehicular multipath channel, which show that the proposed method provides better precision than some well-known timing estimation techniques, especially in low signal to noise ratio (SNR) environments.

  10. A Timing Estimation Method Based-on Skewness Analysis in Vehicular Wireless Networks

    Directory of Open Access Journals (Sweden)

    Xuerong Cui

    2015-11-01

    Full Text Available Vehicle positioning technology has drawn more and more attention in vehicular wireless networks to reduce transportation time and traffic accidents. Nowadays, global navigation satellite systems (GNSS are widely used in land vehicle positioning, but most of them are lack precision and reliability in situations where their signals are blocked. Positioning systems base-on short range wireless communication are another effective way that can be used in vehicle positioning or vehicle ranging. IEEE 802.11p is a new real-time short range wireless communication standard for vehicles, so a new method is proposed to estimate the time delay or ranges between vehicles based on the IEEE 802.11p standard which includes three main steps: cross-correlation between the received signal and the short preamble, summing up the correlated results in groups, and finding the maximum peak using a dynamic threshold based on the skewness analysis. With the range between each vehicle or road-side infrastructure, the position of neighboring vehicles can be estimated correctly. Simulation results were presented in the International Telecommunications Union (ITU vehicular multipath channel, which show that the proposed method provides better precision than some well-known timing estimation techniques, especially in low signal to noise ratio (SNR environments.

  11. Estimating time-varying exposure-outcome associations using case-control data: logistic and case-cohort analyses

    Directory of Open Access Journals (Sweden)

    Ruth H. Keogh

    2016-01-01

    Full Text Available Abstract Background Traditional analyses of standard case-control studies using logistic regression do not allow estimation of time-varying associations between exposures and the outcome. We present two approaches which allow this. The motivation is a study of vaccine efficacy as a function of time since vaccination. Methods Our first approach is to estimate time-varying exposure-outcome associations by fitting a series of logistic regressions within successive time periods, reusing controls across periods. Our second approach treats the case-control sample as a case-cohort study, with the controls forming the subcohort. In the case-cohort analysis, controls contribute information at all times they are at risk. Extensions allow left truncation, frequency matching and, using the case-cohort analysis, time-varying exposures. Simulations are used to investigate the methods. Results The simulation results show that both methods give correct estimates of time-varying effects of exposures using standard case-control data. Using the logistic approach there are efficiency gains by reusing controls over time and care should be taken over the definition of controls within time periods. However, using the case-cohort analysis there is no ambiguity over the definition of controls. The performance of the two analyses is very similar when controls are used most efficiently under the logistic approach. Conclusions Using our methods, case-control studies can be used to estimate time-varying exposure-outcome associations where they may not previously have been considered. The case-cohort analysis has several advantages, including that it allows estimation of time-varying associations as a continuous function of time, while the logistic regression approach is restricted to assuming a step function form for the time-varying association.

  12. Bayesian switching factor analysis for estimating time-varying functional connectivity in fMRI.

    Science.gov (United States)

    Taghia, Jalil; Ryali, Srikanth; Chen, Tianwen; Supekar, Kaustubh; Cai, Weidong; Menon, Vinod

    2017-07-15

    There is growing interest in understanding the dynamical properties of functional interactions between distributed brain regions. However, robust estimation of temporal dynamics from functional magnetic resonance imaging (fMRI) data remains challenging due to limitations in extant multivariate methods for modeling time-varying functional interactions between multiple brain areas. Here, we develop a Bayesian generative model for fMRI time-series within the framework of hidden Markov models (HMMs). The model is a dynamic variant of the static factor analysis model (Ghahramani and Beal, 2000). We refer to this model as Bayesian switching factor analysis (BSFA) as it integrates factor analysis into a generative HMM in a unified Bayesian framework. In BSFA, brain dynamic functional networks are represented by latent states which are learnt from the data. Crucially, BSFA is a generative model which estimates the temporal evolution of brain states and transition probabilities between states as a function of time. An attractive feature of BSFA is the automatic determination of the number of latent states via Bayesian model selection arising from penalization of excessively complex models. Key features of BSFA are validated using extensive simulations on carefully designed synthetic data. We further validate BSFA using fingerprint analysis of multisession resting-state fMRI data from the Human Connectome Project (HCP). Our results show that modeling temporal dependencies in the generative model of BSFA results in improved fingerprinting of individual participants. Finally, we apply BSFA to elucidate the dynamic functional organization of the salience, central-executive, and default mode networks-three core neurocognitive systems with central role in cognitive and affective information processing (Menon, 2011). Across two HCP sessions, we demonstrate a high level of dynamic interactions between these networks and determine that the salience network has the highest temporal

  13. A state-space Bayesian framework for estimating biogeochemical transformations using time-lapse geophysical data

    Energy Technology Data Exchange (ETDEWEB)

    Chen, J.; Hubbard, S.; Williams, K.; Pride, S.; Li, L.; Steefel, C.; Slater, L.

    2009-04-15

    We develop a state-space Bayesian framework to combine time-lapse geophysical data with other types of information for quantitative estimation of biogeochemical parameters during bioremediation. We consider characteristics of end-products of biogeochemical transformations as state vectors, which evolve under constraints of local environments through evolution equations, and consider time-lapse geophysical data as available observations, which could be linked to the state vectors through petrophysical models. We estimate the state vectors and their associated unknown parameters over time using Markov chain Monte Carlo sampling methods. To demonstrate the use of the state-space approach, we apply it to complex resistivity data collected during laboratory column biostimulation experiments that were poised to precipitate iron and zinc sulfides during sulfate reduction. We develop a petrophysical model based on sphere-shaped cells to link the sulfide precipitate properties to the time-lapse geophysical attributes and estimate volume fraction of the sulfide precipitates, fraction of the dispersed, sulfide-encrusted cells, mean radius of the aggregated clusters, and permeability over the course of the experiments. Results of the case study suggest that the developed state-space approach permits the use of geophysical datasets for providing quantitative estimates of end-product characteristics and hydrological feedbacks associated with biogeochemical transformations. Although tested here on laboratory column experiment datasets, the developed framework provides the foundation needed for quantitative field-scale estimation of biogeochemical parameters over space and time using direct, but often sparse wellbore data with indirect, but more spatially extensive geophysical datasets.

  14. Estimating the time evolution of NMR systems via a quantum-speed-limit-like expression

    Science.gov (United States)

    Villamizar, D. V.; Duzzioni, E. I.; Leal, A. C. S.; Auccaise, R.

    2018-05-01

    Finding the solutions of the equations that describe the dynamics of a given physical system is crucial in order to obtain important information about its evolution. However, by using estimation theory, it is possible to obtain, under certain limitations, some information on its dynamics. The quantum-speed-limit (QSL) theory was originally used to estimate the shortest time in which a Hamiltonian drives an initial state to a final one for a given fidelity. Using the QSL theory in a slightly different way, we are able to estimate the running time of a given quantum process. For that purpose, we impose the saturation of the Anandan-Aharonov bound in a rotating frame of reference where the state of the system travels slower than in the original frame (laboratory frame). Through this procedure it is possible to estimate the actual evolution time in the laboratory frame of reference with good accuracy when compared to previous methods. Our method is tested successfully to predict the time spent in the evolution of nuclear spins 1/2 and 3/2 in NMR systems. We find that the estimated time according to our method is better than previous approaches by up to four orders of magnitude. One disadvantage of our method is that we need to solve a number of transcendental equations, which increases with the system dimension and parameter discretization used to solve such equations numerically.

  15. Collective Odor Source Estimation and Search in Time-Variant Airflow Environments Using Mobile Robots

    Science.gov (United States)

    Meng, Qing-Hao; Yang, Wei-Xing; Wang, Yang; Zeng, Ming

    2011-01-01

    This paper addresses the collective odor source localization (OSL) problem in a time-varying airflow environment using mobile robots. A novel OSL methodology which combines odor-source probability estimation and multiple robots’ search is proposed. The estimation phase consists of two steps: firstly, the separate probability-distribution map of odor source is estimated via Bayesian rules and fuzzy inference based on a single robot’s detection events; secondly, the separate maps estimated by different robots at different times are fused into a combined map by way of distance based superposition. The multi-robot search behaviors are coordinated via a particle swarm optimization algorithm, where the estimated odor-source probability distribution is used to express the fitness functions. In the process of OSL, the estimation phase provides the prior knowledge for the searching while the searching verifies the estimation results, and both phases are implemented iteratively. The results of simulations for large-scale advection–diffusion plume environments and experiments using real robots in an indoor airflow environment validate the feasibility and robustness of the proposed OSL method. PMID:22346650

  16. A cellular automata approach to estimate incident-related travel time on Interstate 66 in near real time : final contract report.

    Science.gov (United States)

    2010-03-01

    Incidents account for a large portion of all congestion and a need clearly exists for tools to predict and estimate incident effects. This study examined (1) congestion back propagation to estimate the length of the queue and travel time from upstrea...

  17. A Bayesian Combined Model for Time-Dependent Turning Movement Proportions Estimation at Intersections

    Directory of Open Access Journals (Sweden)

    Pengpeng Jiao

    2014-01-01

    Full Text Available Time-dependent turning movement flows are very important input data for intelligent transportation systems but are impossible to be detected directly through current traffic surveillance systems. Existing estimation models have proved to be not accurate and reliable enough during all intervals. An improved way to address this problem is to develop a combined model framework that can integrate multiple submodels running simultaneously. This paper first presents a back propagation neural network model to estimate dynamic turning movements, as well as the self-adaptive learning rate approach and the gradient descent with momentum method for solving. Second, this paper develops an efficient Kalman filtering model and designs a revised sequential Kalman filtering algorithm. Based on the Bayesian method using both historical data and currently estimated results for error calibration, this paper further integrates above two submodels into a Bayesian combined model framework and proposes a corresponding algorithm. A field survey is implemented at an intersection in Beijing city to collect both time series of link counts and actual time-dependent turning movement flows, including historical and present data. The reported estimation results show that the Bayesian combined model is much more accurate and stable than other models.

  18. Online cell SOC estimation of Li-ion battery packs using a dual time-scale Kalman filtering for EV applications

    International Nuclear Information System (INIS)

    Dai, Haifeng; Wei, Xuezhe; Sun, Zechang; Wang, Jiayuan; Gu, Weijun

    2012-01-01

    Highlights: ► We use an equivalent circuit model to describe the characteristics of battery. ► A dual time-scale estimator is used to calculate pack average SOC and cell SOC. ► The estimator is based on the dynamic descriptions and extended Kalman filter. ► Three different test cases are designed to validate the proposed method. ► Test results indicate a good performance of the method for EV applications. -- Abstract: For the vehicular operation, due to the voltage and power/energy requirements, the battery systems are usually composed of up to hundreds of cells connected in series or parallel. To accommodate the operation conditions, the battery management system (BMS) should estimate State of Charge (SOC) to facilitate safe and efficient utilization of the battery. The performance difference among the cells makes a pure pack SOC estimation hardly provide sufficient information, which at last affects the computation of available energy and power and the safety of the battery system. So for a reliable and accurate management, the BMS should “know” the SOC of each individual cell. Several possible solutions on this issue have been reported in the recent years. This paper studies a method to determine online all individual cell SOCs of a series-connected battery pack. This method, with an equivalent circuit based “averaged cell” model, estimates the battery pack’s average SOC first, and then incorporates the performance divergences between the “averaged cell” and each individual cell to generate the SOC estimations for all cells. This method is developed based on extended Kalman filter (EKF), and to reduce the computation cost, a dual time-scale implementation is designed. The method is validated using results obtained from the measurements of a Li-ion battery pack under three different tests, and analysis indicates the good performance of the algorithm.

  19. "A space-time ensemble Kalman filter for state and parameter estimation of groundwater transport models"

    Science.gov (United States)

    Briseño, Jessica; Herrera, Graciela S.

    2010-05-01

    Herrera (1998) proposed a method for the optimal design of groundwater quality monitoring networks that involves space and time in a combined form. The method was applied later by Herrera et al (2001) and by Herrera and Pinder (2005). To get the estimates of the contaminant concentration being analyzed, this method uses a space-time ensemble Kalman filter, based on a stochastic flow and transport model. When the method is applied, it is important that the characteristics of the stochastic model be congruent with field data, but, in general, it is laborious to manually achieve a good match between them. For this reason, the main objective of this work is to extend the space-time ensemble Kalman filter proposed by Herrera, to estimate the hydraulic conductivity, together with hydraulic head and contaminant concentration, and its application in a synthetic example. The method has three steps: 1) Given the mean and the semivariogram of the natural logarithm of hydraulic conductivity (ln K), random realizations of this parameter are obtained through two alternatives: Gaussian simulation (SGSim) and Latin Hypercube Sampling method (LHC). 2) The stochastic model is used to produce hydraulic head (h) and contaminant (C) realizations, for each one of the conductivity realizations. With these realization the mean of ln K, h and C are obtained, for h and C, the mean is calculated in space and time, and also the cross covariance matrix h-ln K-C in space and time. The covariance matrix is obtained averaging products of the ln K, h and C realizations on the estimation points and times, and the positions and times with data of the analyzed variables. The estimation points are the positions at which estimates of ln K, h or C are gathered. In an analogous way, the estimation times are those at which estimates of any of the three variables are gathered. 3) Finally the ln K, h and C estimate are obtained using the space-time ensemble Kalman filter. The realization mean for each one

  20. H∞ state estimation of stochastic memristor-based neural networks with time-varying delays.

    Science.gov (United States)

    Bao, Haibo; Cao, Jinde; Kurths, Jürgen; Alsaedi, Ahmed; Ahmad, Bashir

    2018-03-01

    This paper addresses the problem of H ∞ state estimation for a class of stochastic memristor-based neural networks with time-varying delays. Under the framework of Filippov solution, the stochastic memristor-based neural networks are transformed into systems with interval parameters. The present paper is the first to investigate the H ∞ state estimation problem for continuous-time Itô-type stochastic memristor-based neural networks. By means of Lyapunov functionals and some stochastic technique, sufficient conditions are derived to ensure that the estimation error system is asymptotically stable in the mean square with a prescribed H ∞ performance. An explicit expression of the state estimator gain is given in terms of linear matrix inequalities (LMIs). Compared with other results, our results reduce control gain and control cost effectively. Finally, numerical simulations are provided to demonstrate the efficiency of the theoretical results. Copyright © 2018 Elsevier Ltd. All rights reserved.

  1. Estimation of vegetation cover resilience from satellite time series

    Directory of Open Access Journals (Sweden)

    T. Simoniello

    2008-07-01

    Full Text Available Resilience is a fundamental concept for understanding vegetation as a dynamic component of the climate system. It expresses the ability of ecosystems to tolerate disturbances and to recover their initial state. Recovery times are basic parameters of the vegetation's response to forcing and, therefore, are essential for describing realistic vegetation within dynamical models. Healthy vegetation tends to rapidly recover from shock and to persist in growth and expansion. On the contrary, climatic and anthropic stress can reduce resilience thus favouring persistent decrease in vegetation activity.

    In order to characterize resilience, we analyzed the time series 1982–2003 of 8 km GIMMS AVHRR-NDVI maps of the Italian territory. Persistence probability of negative and positive trends was estimated according to the vegetation cover class, altitude, and climate. Generally, mean recovery times from negative trends were shorter than those estimated for positive trends, as expected for vegetation of healthy status. Some signatures of inefficient resilience were found in high-level mountainous areas and in the Mediterranean sub-tropical ones. This analysis was refined by aggregating pixels according to phenology. This multitemporal clustering synthesized information on vegetation cover, climate, and orography rather well. The consequent persistence estimations confirmed and detailed hints obtained from the previous analyses. Under the same climatic regime, different vegetation resilience levels were found. In particular, within the Mediterranean sub-tropical climate, clustering was able to identify features with different persistence levels in areas that are liable to different levels of anthropic pressure. Moreover, it was capable of enhancing reduced vegetation resilience also in the southern areas under Warm Temperate sub-continental climate. The general consistency of the obtained results showed that, with the help of suited analysis

  2. Real-time estimation of differential piston at the LBT

    Science.gov (United States)

    Böhm, Michael; Pott, Jörg-Uwe; Sawodny, Oliver; Herbst, Tom; Kürster, Martin

    2014-07-01

    In this paper, we present and compare different strategies to minimize the effects of telescope vibrations to the differential piston (OPD) for LINC/NIRVANA at the LBT using an accelerometer feedforward compensation approach. We summarize why this technology is of importance for LINC/NIRVANA, but also for future telescopes and instruments. We outline the estimation problem in general and its specifics at the LBT. Model based estimation and broadband filtering techniques can be used to solve the estimation task, each having its own advantages and disadvantages, which will be discussed. Simulation results and measurements at the LBT are shown to motivate and support our choice of the estimation algorithm for the instrument LINC/NIRVANA. We explain our laboratory setup aimed at imitating the vibration behaviour at the LBT in general, and the M2 as main contributor in particular, and we demonstrate the controller's ability to suppress vibrations in the frequency range of 8 Hz to 60 Hz. In this range, telescope vibrations are the most dominant disturbance to the optical path. For our measurements, we introduce a disturbance time series which has a frequency spectrum comparable to what can be measured at the LBT on a typical night. We show promising experimental results, indicating the ability to suppress differential piston induced by telescope vibrations by a factor of about 5 (RMS), which is significantly better than any currently commissioned system.

  3. Religious affiliation at time of death - Global estimates and projections.

    Science.gov (United States)

    Skirbekk, Vegard; Todd, Megan; Stonawski, Marcin

    2018-03-01

    Religious affiliation influences societal practices regarding death and dying, including palliative care, religiously acceptable health service procedures, funeral rites and beliefs about an afterlife. We aimed to estimate and project religious affiliation at the time of death globally, as this information has been lacking. We compiled data on demographic information and religious affiliation from more than 2500 surveys, registers and censuses covering 198 nations/territories. We present estimates of religious affiliation at the time of death as of 2010, projections up to and including 2060, taking into account trends in mortality, religious conversion, intergenerational transmission of religion, differential fertility, and gross migration flows, by age and sex. We find that Christianity continues to be the most common religion at death, although its share will fall from 37% to 31% of global deaths between 2010 and 2060. The share of individuals identifying as Muslim at the time of death increases from 21% to 24%. The share of religiously unaffiliated will peak at 17% in 2035 followed by a slight decline thereafter. In specific regions, such as Europe, the unaffiliated share will continue to rises from 14% to 21% throughout the period. Religious affiliation at the time of death is changing globally, with distinct regional patterns. This could affect spatial variation in healthcare and social customs relating to death and dying.

  4. Discrete-time state estimation for stochastic polynomial systems over polynomial observations

    Science.gov (United States)

    Hernandez-Gonzalez, M.; Basin, M.; Stepanov, O.

    2018-07-01

    This paper presents a solution to the mean-square state estimation problem for stochastic nonlinear polynomial systems over polynomial observations confused with additive white Gaussian noises. The solution is given in two steps: (a) computing the time-update equations and (b) computing the measurement-update equations for the state estimate and error covariance matrix. A closed form of this filter is obtained by expressing conditional expectations of polynomial terms as functions of the state estimate and error covariance. As a particular case, the mean-square filtering equations are derived for a third-degree polynomial system with second-degree polynomial measurements. Numerical simulations show effectiveness of the proposed filter compared to the extended Kalman filter.

  5. Robust state estimation for uncertain fuzzy bidirectional associative memory networks with time-varying delays

    Science.gov (United States)

    Vadivel, P.; Sakthivel, R.; Mathiyalagan, K.; Arunkumar, A.

    2013-09-01

    This paper addresses the issue of robust state estimation for a class of fuzzy bidirectional associative memory (BAM) neural networks with time-varying delays and parameter uncertainties. By constructing the Lyapunov-Krasovskii functional, which contains the triple-integral term and using the free-weighting matrix technique, a set of sufficient conditions are derived in terms of linear matrix inequalities (LMIs) to estimate the neuron states through available output measurements such that the dynamics of the estimation error system is robustly asymptotically stable. In particular, we consider a generalized activation function in which the traditional assumptions on the boundedness, monotony and differentiability of the activation functions are removed. More precisely, the design of the state estimator for such BAM neural networks can be obtained by solving some LMIs, which are dependent on the size of the time derivative of the time-varying delays. Finally, a numerical example with simulation result is given to illustrate the obtained theoretical results.

  6. Robust state estimation for uncertain fuzzy bidirectional associative memory networks with time-varying delays

    International Nuclear Information System (INIS)

    Vadivel, P; Sakthivel, R; Mathiyalagan, K; Arunkumar, A

    2013-01-01

    This paper addresses the issue of robust state estimation for a class of fuzzy bidirectional associative memory (BAM) neural networks with time-varying delays and parameter uncertainties. By constructing the Lyapunov–Krasovskii functional, which contains the triple-integral term and using the free-weighting matrix technique, a set of sufficient conditions are derived in terms of linear matrix inequalities (LMIs) to estimate the neuron states through available output measurements such that the dynamics of the estimation error system is robustly asymptotically stable. In particular, we consider a generalized activation function in which the traditional assumptions on the boundedness, monotony and differentiability of the activation functions are removed. More precisely, the design of the state estimator for such BAM neural networks can be obtained by solving some LMIs, which are dependent on the size of the time derivative of the time-varying delays. Finally, a numerical example with simulation result is given to illustrate the obtained theoretical results. (paper)

  7. Online Estimation of Time-Varying Volatility Using a Continuous-Discrete LMS Algorithm

    Directory of Open Access Journals (Sweden)

    Jacques Oksman

    2008-09-01

    Full Text Available The following paper addresses a problem of inference in financial engineering, namely, online time-varying volatility estimation. The proposed method is based on an adaptive predictor for the stock price, built from an implicit integration formula. An estimate for the current volatility value which minimizes the mean square prediction error is calculated recursively using an LMS algorithm. The method is then validated on several synthetic examples as well as on real data. Throughout the illustration, the proposed method is compared with both UKF and offline volatility estimation.

  8. Spectral Estimation of UV-Vis Absorbance Time Series for Water Quality Monitoring

    Directory of Open Access Journals (Sweden)

    Leonardo Plazas-Nossa

    2017-05-01

    Full Text Available Context: Signals recorded as multivariate time series by UV-Vis absorbance captors installed in urban sewer systems, can be non-stationary, yielding complications in the analysis of water quality monitoring. This work proposes to perform spectral estimation using the Box-Cox transformation and differentiation in order to obtain stationary multivariate time series in a wide sense. Additionally, Principal Component Analysis (PCA is applied to reduce their dimensionality. Method: Three different UV-Vis absorbance time series for different Colombian locations were studied: (i El-Salitre Wastewater Treatment Plant (WWTP in Bogotá; (ii Gibraltar Pumping Station (GPS in Bogotá; and (iii San-Fernando WWTP in Itagüí. Each UV-Vis absorbance time series had equal sample number (5705. The esti-mation of the spectral power density is obtained using the average of modified periodograms with rectangular window and an overlap of 50%, with the 20 most important harmonics from the Discrete Fourier Transform (DFT and Inverse Fast Fourier Transform (IFFT. Results: Absorbance time series dimensionality reduction using PCA, resulted in 6, 8 and 7 principal components for each study site respectively, altogether explaining more than 97% of their variability. Values of differences below 30% for the UV range were obtained for the three study sites, while for the visible range the maximum differences obtained were: (i 35% for El-Salitre WWTP; (ii 61% for GPS; and (iii 75% for San-Fernando WWTP. Conclusions: The Box-Cox transformation and the differentiation process applied to the UV-Vis absorbance time series for the study sites (El-Salitre, GPS and San-Fernando, allowed to reduce variance and to eliminate ten-dency of the time series. A pre-processing of UV-Vis absorbance time series is recommended to detect and remove outliers and then apply the proposed process for spectral estimation. Language: Spanish.

  9. A Generalized Estimating Equations Approach to Model Heterogeneity and Time Dependence in Capture-Recapture Studies

    Directory of Open Access Journals (Sweden)

    Akanda Md. Abdus Salam

    2017-03-01

    Full Text Available Individual heterogeneity in capture probabilities and time dependence are fundamentally important for estimating the closed animal population parameters in capture-recapture studies. A generalized estimating equations (GEE approach accounts for linear correlation among capture-recapture occasions, and individual heterogeneity in capture probabilities in a closed population capture-recapture individual heterogeneity and time variation model. The estimated capture probabilities are used to estimate animal population parameters. Two real data sets are used for illustrative purposes. A simulation study is carried out to assess the performance of the GEE estimator. A Quasi-Likelihood Information Criterion (QIC is applied for the selection of the best fitting model. This approach performs well when the estimated population parameters depend on the individual heterogeneity and the nature of linear correlation among capture-recapture occasions.

  10. Multiple linear regression to estimate time-frequency electrophysiological responses in single trials.

    Science.gov (United States)

    Hu, L; Zhang, Z G; Mouraux, A; Iannetti, G D

    2015-05-01

    Transient sensory, motor or cognitive event elicit not only phase-locked event-related potentials (ERPs) in the ongoing electroencephalogram (EEG), but also induce non-phase-locked modulations of ongoing EEG oscillations. These modulations can be detected when single-trial waveforms are analysed in the time-frequency domain, and consist in stimulus-induced decreases (event-related desynchronization, ERD) or increases (event-related synchronization, ERS) of synchrony in the activity of the underlying neuronal populations. ERD and ERS reflect changes in the parameters that control oscillations in neuronal networks and, depending on the frequency at which they occur, represent neuronal mechanisms involved in cortical activation, inhibition and binding. ERD and ERS are commonly estimated by averaging the time-frequency decomposition of single trials. However, their trial-to-trial variability that can reflect physiologically-important information is lost by across-trial averaging. Here, we aim to (1) develop novel approaches to explore single-trial parameters (including latency, frequency and magnitude) of ERP/ERD/ERS; (2) disclose the relationship between estimated single-trial parameters and other experimental factors (e.g., perceived intensity). We found that (1) stimulus-elicited ERP/ERD/ERS can be correctly separated using principal component analysis (PCA) decomposition with Varimax rotation on the single-trial time-frequency distributions; (2) time-frequency multiple linear regression with dispersion term (TF-MLRd) enhances the signal-to-noise ratio of ERP/ERD/ERS in single trials, and provides an unbiased estimation of their latency, frequency, and magnitude at single-trial level; (3) these estimates can be meaningfully correlated with each other and with other experimental factors at single-trial level (e.g., perceived stimulus intensity and ERP magnitude). The methods described in this article allow exploring fully non-phase-locked stimulus-induced cortical

  11. A Fuzzy Logic-Based Approach for Estimation of Dwelling Times of Panama Metro Stations

    Directory of Open Access Journals (Sweden)

    Aranzazu Berbey Alvarez

    2015-04-01

    Full Text Available Passenger flow modeling and station dwelling time estimation are significant elements for railway mass transit planning, but system operators usually have limited information to model the passenger flow. In this paper, an artificial-intelligence technique known as fuzzy logic is applied for the estimation of the elements of the origin-destination matrix and the dwelling time of stations in a railway transport system. The fuzzy inference engine used in the algorithm is based in the principle of maximum entropy. The approach considers passengers’ preferences to assign a level of congestion in each car of the train in function of the properties of the station platforms. This approach is implemented to estimate the passenger flow and dwelling times of the recently opened Line 1 of the Panama Metro. The dwelling times obtained from the simulation are compared to real measurements to validate the approach.

  12. Sensitivity of Hurst parameter estimation to periodic signals in time series and filtering approaches

    Science.gov (United States)

    Marković, D.; Koch, M.

    2005-09-01

    The influence of the periodic signals in time series on the Hurst parameter estimate is investigated with temporal, spectral and time-scale methods. The Hurst parameter estimates of the simulated periodic time series with a white noise background show a high sensitivity on the signal to noise ratio and for some methods, also on the data length used. The analysis is then carried on to the investigation of extreme monthly river flows of the Elbe River (Dresden) and of the Rhine River (Kaub). Effects of removing the periodic components employing different filtering approaches are discussed and it is shown that such procedures are a prerequisite for an unbiased estimation of H. In summary, our results imply that the first step in a time series long-correlation study should be the separation of the deterministic components from the stochastic ones. Otherwise wrong conclusions concerning possible memory effects may be drawn.

  13. H∞ state estimation of generalised neural networks with interval time-varying delays

    Science.gov (United States)

    Saravanakumar, R.; Syed Ali, M.; Cao, Jinde; Huang, He

    2016-12-01

    This paper focuses on studying the H∞ state estimation of generalised neural networks with interval time-varying delays. The integral terms in the time derivative of the Lyapunov-Krasovskii functional are handled by the Jensen's inequality, reciprocally convex combination approach and a new Wirtinger-based double integral inequality. A delay-dependent criterion is derived under which the estimation error system is globally asymptotically stable with H∞ performance. The proposed conditions are represented by linear matrix inequalities. Optimal H∞ norm bounds are obtained easily by solving convex problems in terms of linear matrix inequalities. The advantage of employing the proposed inequalities is illustrated by numerical examples.

  14. Ultra-small time-delay estimation via a weak measurement technique with post-selection

    International Nuclear Information System (INIS)

    Fang, Chen; Huang, Jing-Zheng; Yu, Yang; Li, Qinzheng; Zeng, Guihua

    2016-01-01

    Weak measurement is a novel technique for parameter estimation with higher precision. In this paper we develop a general theory for the parameter estimation based on a weak measurement technique with arbitrary post-selection. The weak-value amplification model and the joint weak measurement model are two special cases in our theory. Applying the developed theory, time-delay estimation is investigated in both theory and experiments. The experimental results show that when the time delay is ultra-small, the joint weak measurement scheme outperforms the weak-value amplification scheme, and is robust against not only misalignment errors but also the wavelength dependence of the optical components. These results are consistent with theoretical predictions that have not been previously verified by any experiment. (paper)

  15. Uniform Estimate of the Finite-Time Ruin Probability for All Times in a Generalized Compound Renewal Risk Model

    Directory of Open Access Journals (Sweden)

    Qingwu Gao

    2012-01-01

    Full Text Available We discuss the uniformly asymptotic estimate of the finite-time ruin probability for all times in a generalized compound renewal risk model, where the interarrival times of successive accidents and all the claim sizes caused by an accident are two sequences of random variables following a wide dependence structure. This wide dependence structure allows random variables to be either negatively dependent or positively dependent.

  16. Modelling the Magnetic Disturbances Due to Road-Traffic

    Directory of Open Access Journals (Sweden)

    J -J Schott

    2011-07-01

    Full Text Available Magnetic disturbances due to the traffic are tentatively modelled assuming that the sources are moving dipoles. The influencing section of the road ("useful" portion should be modelled in 3D. The parameters of the model (time of closest position to the magnetometer, velocity, including its sign, dipole moment are fairly accurately estimated. The fit is improved with the incorporation of a small induction effect.

  17. Reconstruction of temporal variations of evapotranspiration using instantaneous estimates at the time of satellite overpass

    Directory of Open Access Journals (Sweden)

    E. Delogu

    2012-08-01

    Full Text Available Evapotranspiration estimates can be derived from remote sensing data and ancillary, mostly meterorological, information. For this purpose, two types of methods are classically used: the first type estimates a potential evapotranspiration rate from vegetation indices, and adjusts this rate according to water availability derived from either a surface temperature index or a first guess obtained from a rough estimate of the water budget, while the second family of methods relies on the link between the surface temperature and the latent heat flux through the surface energy budget. The latter provides an instantaneous estimate at the time of satellite overpass. In order to compute daily evapotranspiration, one needs an extrapolation algorithm. Since no image is acquired during cloudy conditions, these methods can only be applied during clear sky days. In order to derive seasonal evapotranspiration, one needs an interpolation method. Two combined interpolation/extrapolation methods based on the self preservation of evaporative fraction and the stress factor are compared to reconstruct seasonal evapotranspiration from instantaneous measurements acquired in clear sky conditions. Those measurements are taken from instantaneous latent heat flux from 11 datasets in Southern France and Morocco. Results show that both methods have comparable performances with a clear advantage for the evaporative fraction for datasets with several water stress events. Both interpolation algorithms tend to underestimate evapotranspiration due to the energy limiting conditions that prevail during cloudy days. Taking into account the diurnal variations of the evaporative fraction according to an empirical relationship derived from a previous study improved the performance of the extrapolation algorithm and therefore the retrieval of the seasonal evapotranspiration for all but one datasets.

  18. Estimated Incident Cost Savings in Shipping Due to Inspections

    NARCIS (Netherlands)

    S. Knapp (Sabine); G.E. Bijwaard (Govert); C. Heij (Christiaan)

    2010-01-01

    textabstractThe effectiveness of safety inspections has been analysed from various angles, but until now, relatively little attention has been given to translate risk reduction into incident cost savings. This paper quantifies estimated cost savings based on port state control inspections and

  19. Quantifying uncertainty in NDSHA estimates due to earthquake catalogue

    Science.gov (United States)

    Magrin, Andrea; Peresan, Antonella; Vaccari, Franco; Panza, Giuliano

    2014-05-01

    The procedure for the neo-deterministic seismic zoning, NDSHA, is based on the calculation of synthetic seismograms by the modal summation technique. This approach makes use of information about the space distribution of large magnitude earthquakes, which can be defined based on seismic history and seismotectonics, as well as incorporating information from a wide set of geological and geophysical data (e.g., morphostructural features and ongoing deformation processes identified by earth observations). Hence the method does not make use of attenuation models (GMPE), which may be unable to account for the complexity of the product between seismic source tensor and medium Green function and are often poorly constrained by the available observations. NDSHA defines the hazard from the envelope of the values of ground motion parameters determined considering a wide set of scenario earthquakes; accordingly, the simplest outcome of this method is a map where the maximum of a given seismic parameter is associated to each site. In NDSHA uncertainties are not statistically treated as in PSHA, where aleatory uncertainty is traditionally handled with probability density functions (e.g., for magnitude and distance random variables) and epistemic uncertainty is considered by applying logic trees that allow the use of alternative models and alternative parameter values of each model, but the treatment of uncertainties is performed by sensitivity analyses for key modelling parameters. To fix the uncertainty related to a particular input parameter is an important component of the procedure. The input parameters must account for the uncertainty in the prediction of fault radiation and in the use of Green functions for a given medium. A key parameter is the magnitude of sources used in the simulation that is based on catalogue informations, seismogenic zones and seismogenic nodes. Because the largest part of the existing catalogues is based on macroseismic intensity, a rough estimate

  20. Timing-based location estimation for OFDM signals with applications in LTE, WLAN and Wimax Timing-based location estimation for OFDM signals with applications in LTE, WLAN and Wimax Timing-based location estimation for OFDM signals with applications in LTE, WLAN and Wimax

    OpenAIRE

    Juárez Leria, Víctor

    2012-01-01

    Projecte realitzat en el marc d'un programa de mobilitat amb la Tampere University of Technology (Tampereen Teknillinen Yliopisto) [ANGLÈS] Literature study of various preamble-based and non-preamble-based timing synchronization algorithms in OFDM - Implementation (in Matlab) of the following preamble-based algorithms: Schmidl, Minn, Park, Kim, Ren and Kang. Implementation (in Matlab) of Correlation-Based Timing Synchronization and MUltiple SIgnal Classification estimators. Comparison of v...

  1. Two-pass imputation algorithm for missing value estimation in gene expression time series.

    Science.gov (United States)

    Tsiporkova, Elena; Boeva, Veselka

    2007-10-01

    Gene expression microarray experiments frequently generate datasets with multiple values missing. However, most of the analysis, mining, and classification methods for gene expression data require a complete matrix of gene array values. Therefore, the accurate estimation of missing values in such datasets has been recognized as an important issue, and several imputation algorithms have already been proposed to the biological community. Most of these approaches, however, are not particularly suitable for time series expression profiles. In view of this, we propose a novel imputation algorithm, which is specially suited for the estimation of missing values in gene expression time series data. The algorithm utilizes Dynamic Time Warping (DTW) distance in order to measure the similarity between time expression profiles, and subsequently selects for each gene expression profile with missing values a dedicated set of candidate profiles for estimation. Three different DTW-based imputation (DTWimpute) algorithms have been considered: position-wise, neighborhood-wise, and two-pass imputation. These have initially been prototyped in Perl, and their accuracy has been evaluated on yeast expression time series data using several different parameter settings. The experiments have shown that the two-pass algorithm consistently outperforms, in particular for datasets with a higher level of missing entries, the neighborhood-wise and the position-wise algorithms. The performance of the two-pass DTWimpute algorithm has further been benchmarked against the weighted K-Nearest Neighbors algorithm, which is widely used in the biological community; the former algorithm has appeared superior to the latter one. Motivated by these findings, indicating clearly the added value of the DTW techniques for missing value estimation in time series data, we have built an optimized C++ implementation of the two-pass DTWimpute algorithm. The software also provides for a choice between three different

  2. Uncertainties in surface mass and energy flux estimates due to different eddy covariance sensors and technical set-up

    Science.gov (United States)

    Arriga, Nicola; Fratini, Gerardo; Forgione, Antonio; Tomassucci, Michele; Papale, Dario

    2010-05-01

    Eddy covariance is a well established and widely used methodology for the measurement of turbulent fluxes of mass and energy in the atmospheric boundary layer, in particular to estimate CO2/H2O and heat exchange above ecologically relevant surfaces (Aubinet 2000, Baldocchi 2003). Despite its long term application and theoretical studies, many issues are still open about the effect of different experimental set-up on final flux estimates. Open issues are the evaluation of the performances of different kind of sensors (e.g. open path vs closed path infra-red gas analysers, vertical vs horizontal mounting ultrasonic anemometers), the quantification of the impact of corresponding physical corrections to be applied to get robust flux estimates taking in account all processes concurring to the measurement (e.g. the so-called WPL term, signal attenuation due to air sampling system for closed path analyser, relative position of analyser and anemometer) and the differences between several data transmission protocols used (analogue, digital RS-232, SDM). A field experiment was designed to study these issues using several instruments among those most used within the Fluxnet community and to compare their performances under conditions supposed to be critical: rainy and cold weather conditions for open-path analysers (Burba 2008), water transport and absorption at high air relative humidity conditions for closed-path systems (Ibrom, 2007), frequency sampling limits and recorded data robustness due to different transmission protocols (RS232, SDM, USB, Ethernet) and finally the effect of the displacement between anemometer and analyser using at least two identical analysers placed at different horizontal and vertical distances from the anemometer. Aim of this experiment is to quantify the effect of several technical solutions on the final estimates of fluxes measured at a point in the space and if they represent a significant source of uncertainty for mass and energy cycle

  3. Time-dependent excitation and ionization modelling of absorption-line variability due to GRB080310

    DEFF Research Database (Denmark)

    Vreeswijk, P.M.; De Cia, A.; Jakobsson, P.

    2013-01-01

    .42743. To estimate the rest-frame afterglow brightness as a function of time, we use a combination of the optical VRI photometry obtained by the RAPTOR-T telescope array, which is presented in this paper, and Swift's X-Ray Telescope (XRT) observations. Excitation alone, which has been successfully applied...

  4. Estimating Reaction Rate Coefficients Within a Travel-Time Modeling Framework

    Energy Technology Data Exchange (ETDEWEB)

    Gong, R [Georgia Institute of Technology; Lu, C [Georgia Institute of Technology; Luo, Jian [Georgia Institute of Technology; Wu, Wei-min [Stanford University; Cheng, H. [Stanford University; Criddle, Craig [Stanford University; Kitanidis, Peter K. [Stanford University; Gu, Baohua [ORNL; Watson, David B [ORNL; Jardine, Philip M [ORNL; Brooks, Scott C [ORNL

    2011-03-01

    A generalized, efficient, and practical approach based on the travel-time modeling framework is developed to estimate in situ reaction rate coefficients for groundwater remediation in heterogeneous aquifers. The required information for this approach can be obtained by conducting tracer tests with injection of a mixture of conservative and reactive tracers and measurements of both breakthrough curves (BTCs). The conservative BTC is used to infer the travel-time distribution from the injection point to the observation point. For advection-dominant reactive transport with well-mixed reactive species and a constant travel-time distribution, the reactive BTC is obtained by integrating the solutions to advective-reactive transport over the entire travel-time distribution, and then is used in optimization to determine the in situ reaction rate coefficients. By directly working on the conservative and reactive BTCs, this approach avoids costly aquifer characterization and improves the estimation for transport in heterogeneous aquifers which may not be sufficiently described by traditional mechanistic transport models with constant transport parameters. Simplified schemes are proposed for reactive transport with zero-, first-, nth-order, and Michaelis-Menten reactions. The proposed approach is validated by a reactive transport case in a two-dimensional synthetic heterogeneous aquifer and a field-scale bioremediation experiment conducted at Oak Ridge, Tennessee. The field application indicates that ethanol degradation for U(VI)-bioremediation is better approximated by zero-order reaction kinetics than first-order reaction kinetics.

  5. Investigation of Bicycle Travel Time Estimation Using Bluetooth Sensors for Low Sampling Rates

    Directory of Open Access Journals (Sweden)

    Zhenyu Mei

    2014-10-01

    Full Text Available Filtering the data for bicycle travel time using Bluetooth sensors is crucial to the estimation of link travel times on a corridor. The current paper describes an adaptive filtering algorithm for estimating bicycle travel times using Bluetooth data, with consideration of low sampling rates. The data for bicycle travel time using Bluetooth sensors has two characteristics. First, the bicycle flow contains stable and unstable conditions. Second, the collected data have low sampling rates (less than 1%. To avoid erroneous inference, filters are introduced to “purify” multiple time series. The valid data are identified within a dynamically varying validity window with the use of a robust data-filtering procedure. The size of the validity window varies based on the number of preceding sampling intervals without a Bluetooth record. Applications of the proposed algorithm to the dataset from Genshan East Road and Moganshan Road in Hangzhou demonstrate its ability to track typical variations in bicycle travel time efficiently, while suppressing high frequency noise signals.

  6. RealCalc : a real time Java calculation tool. Application to HVSR estimation

    Science.gov (United States)

    Hloupis, G.; Vallianatos, F.

    2009-04-01

    Java computation platform is not a newcomer in the seismology field. It is mainly used for applications regarding collecting, requesting, spreading and visualizing seismological data because it is productive, safe and has low maintenance costs. Although it has very attractive characteristics for the engineers, Java didn't used frequently in real time applications where prediction and reliability required as a reaction to real world events. The main reasons for this are the absence of priority support (such as priority ceiling or priority inversion) and the use of an automated memory management (called garbage collector). To overcome these problems a number of extensions have been proposed with the Real Time Specification for Java (RTSJ) being the most promising and used one. In the current study we used the RTSJ to build an application that receives data continuously and provides estimations in real time. The application consists of four main modules: incoming data, preprocessing, estimation and publication. As an application example we present real time HVSR estimation. Microtremors recordings are collected continuously from the incoming data module. The preprocessing module consists of a window selector tool based on wavelets which is applied on the incoming data stream in order derive the most stationary parts. The estimation module provides all the necessary calculations according to user specifications. Finally the publication module except the results presentation it also calculates attributes and relevant statistics for each site (temporal variations, HVSR stability). Acknowledgements This work is partially supported by the Greek General Secretariat of Research and Technology in the frame of Crete Regional Project 2000- 2006 (M1.2): "TALOS: An integrated system of seismic hazard monitoring and management in the front of the Hellenic Arc", CRETE PEP7 (KP_7).

  7. Estimation of Continuous Time Models in Economics: an Overview

    OpenAIRE

    Clifford R. Wymer

    2009-01-01

    The dynamics of economic behaviour is often developed in theory as a continuous time system. Rigorous estimation and testing of such systems, and the analysis of some aspects of their properties, is of particular importance in distinguishing between competing hypotheses and the resulting models. The consequences for the international economy during the past eighteen months of failures in the financial sector, and particularly the banking sector, make it essential that the dynamics of financia...

  8. MODELING TIME DISPERSION DUE TO OPTICAL PATH LENGTH DIFFERENCES IN SCINTILLATION DETECTORS*

    Science.gov (United States)

    Moses, W.W.; Choong, W.-S.; Derenzo, S.E.

    2015-01-01

    We characterize the nature of the time dispersion in scintillation detectors caused by path length differences of the scintillation photons as they travel from their generation point to the photodetector. Using Monte Carlo simulation, we find that the initial portion of the distribution (which is the only portion that affects the timing resolution) can usually be modeled by an exponential decay. The peak amplitude and decay time depend both on the geometry of the crystal, the position within the crystal that the scintillation light originates from, and the surface finish. In a rectangular parallelpiped LSO crystal with 3 mm × 3 mm cross section and polished surfaces, the decay time ranges from 10 ps (for interactions 1 mm from the photodetector) up to 80 ps (for interactions 50 mm from the photodetector). Over that same range of distances, the peak amplitude ranges from 100% (defined as the peak amplitude for interactions 1 mm from the photodetector) down to 4% for interactions 50 mm from the photodetector. Higher values for the decay time are obtained for rough surfaces, but the exact value depends on the simulation details. Estimates for the decay time and peak amplitude can be made for different cross section sizes via simple scaling arguments. PMID:25729464

  9. [Estimation of the population attributable fraction due to obesity in hospital admissions for flu valued according to Body Mass Index (BMI) and CUN-BAE].

    Science.gov (United States)

    Dávila-Batista, V; Carriedo, D; Díez, F; Pueyo Bastida, A; Martínez Durán, B; Martin, V

    2018-03-01

    The obesity pandemic together with the influenza pandemic could lead to a significant burden of disease. The body mass index (BMI) does not discriminate obesity appropriately. The CUN-BAE has recently been used as an estimate of body fatness for Caucasians, including BMI, gender, and age. The aim of this study is to assess the population attributable fraction of hospital admissions due to influenza, due to the body fatness measured with the BMI, and the CUN-BAE. A multicentre study was conducted using matched case-controls. Cases were hospital admissions with the influenza confirmed by the RT-PCR method between 2009 and 2011. The risk of hospital admission and the population attribuible fraction were calculated using the BMI or the CUN-BAE for each adiposity category in a conditional logical regression analysis adjusted for confounding variables. The analyzes were estimated in the total sample, in unvaccinated people, and those less than 65 years-old. A total of 472 hospitalised cases and 493 controls were included in the study. Compared to normal weight, the aOR of influenza hospital admissions increases with each level of BMI (aOR=1.26; 2.06 and 11.64) and CUN-BAE (aOR=2.78; 4.29; 5.43 and 15.18). The population attributable fraction of influenza admissions using CUN-BAE is 3 times higher than that estimated with BMI (0,72 vs. 0,27), with the differences found being similar the non-vaccinated and under 65 year-olds. The BMI could be underestimating the burden of disease attributable to obesity in individuals hospitalised with influenza. There needs to be an appropriate assessment of the impact of obesity and vaccine recommendation criteria. Copyright © 2017 Sociedad Española de Médicos de Atención Primaria (SEMERGEN). Publicado por Elsevier España, S.L.U. All rights reserved.

  10. A model for beta skin dose estimation due to the use of a necklace with uranium depleted bullets

    International Nuclear Information System (INIS)

    Lavalle Heibron, P.H.; Pérez Guerrero, J.S.; Oliveira, J.F. de

    2015-01-01

    Depleted uranium bullets were use as munitions during the Kuwait – Iraq war and the International Atomic Energy Agency sampling expert’s team found fragments in the environment when the war was over. Consequently, there is a possibility that members of the public, especially children, collects DU fragments and use it, for example, to make a necklace. This paper estimates the beta skin dose to a child that uses a necklace made with a depleted uranium bullet. The theoretical model for dose estimation is based on Loevinguer’s equation with a correction factor adjusted for the maximum beta energy in the range between 0.1 and 2.5 MeV calculated taking into account the International Atomic Energy Agency expected doses rates in air at one meter distance of a point source of 37 GBq, function of the maximum beta energy. The dose rate estimated by this work due to the child use of a necklace with one depleted uranium bullet of 300 g was in good agreement with other results founded in literature. (authors)

  11. An estimate of energy dissipation due to soil-moisture hysteresis

    KAUST Repository

    McNamara, H.

    2014-01-01

    Processes of infiltration, transport, and outflow in unsaturated soil necessarily involve the dissipation of energy through various processes. Accounting for these energetic processes can contribute to modeling hydrological and ecological systems. The well-documented hysteretic relationship between matric potential and moisture content in soil suggests that one such mechanism of energy dissipation is associated with the cycling between wetting and drying processes, but it is challenging to estimate the magnitude of the effect in situ. The Preisach model, a generalization of the Independent Domain model, allows hysteresis effects to be incorporated into dynamical systems of differential equations. Building on earlier work using such systems with field data from the south-west of Ireland, this work estimates the average rate of hysteretic energy dissipation. Through some straightforward assumptions, the magnitude of this rate is found to be of O(10-5) W m-3. Key Points Hysteresis in soil-water dissipates energy The rate of dissipation can be estimated directly from saturation data The rate of heating caused is significant ©2013. American Geophysical Union. All Rights Reserved.

  12. Improving value of travel time savings estimation for more effective transportation project evaluation.

    Science.gov (United States)

    2012-12-01

    Estimates of value of time (VOT) and value of travel time savings (VTTS) are critical elements in benefitcost : analyses of transportation projects and in developing congestion pricing policies. In addition, : differences in VTTS among various modes ...

  13. A methodological framework of travel time distribution estimation for urban signalized arterial roads

    NARCIS (Netherlands)

    Zheng, Fangfang; van Zuylen, H.J.; Liu, Xiaobo

    2017-01-01

    Urban travel times are rather variable as a result of a lot of stochastic factors both in traffic flows, signals, and other conditions on the infrastructure. However, the most common way both in literature and practice is to estimate or predict only expected travel times, not travel time

  14. Joint Symbol Timing and CFO Estimation for OFDM/OQAM Systems in Multipath Channels

    Directory of Open Access Journals (Sweden)

    Petrella Angelo

    2010-01-01

    Full Text Available The problem of data-aided synchronization for orthogonal frequency division multiplexing (OFDM systems based on offset quadrature amplitude modulation (OQAM in multipath channels is considered. In particular, the joint maximum-likelihood (ML estimator for carrier-frequency offset (CFO, amplitudes, phases, and delays, exploiting a short known preamble, is derived. The ML estimators for phases and amplitudes are in closed form. Moreover, under the assumption that the CFO is sufficiently small, a closed form approximate ML (AML CFO estimator is obtained. By exploiting the obtained closed form solutions a cost function whose peaks provide an estimate of the delays is derived. In particular, the symbol timing (i.e., the delay of the first multipath component is obtained by considering the smallest estimated delay. The performance of the proposed joint AML estimator is assessed via computer simulations and compared with that achieved by the joint AML estimator designed for AWGN channel and that achieved by a previously derived joint estimator for OFDM systems.

  15. Estimation of Peaking Factor Uncertainty due to Manufacturing Tolerance using Statistical Sampling Method

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Kyung Hoon; Park, Ho Jin; Lee, Chung Chan; Cho, Jin Young [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2015-10-15

    The purpose of this paper is to study the effect on output parameters in the lattice physics calculation due to the last input uncertainty such as manufacturing deviations from nominal value for material composition and geometric dimensions. In a nuclear design and analysis, the lattice physics calculations are usually employed to generate lattice parameters for the nodal core simulation and pin power reconstruction. These lattice parameters which consist of homogenized few-group cross-sections, assembly discontinuity factors, and form-functions can be affected by input uncertainties which arise from three different sources: 1) multi-group cross-section uncertainties, 2) the uncertainties associated with methods and modeling approximations utilized in lattice physics codes, and 3) fuel/assembly manufacturing uncertainties. In this paper, data provided by the light water reactor (LWR) uncertainty analysis in modeling (UAM) benchmark has been used as the manufacturing uncertainties. First, the effect of each input parameter has been investigated through sensitivity calculations at the fuel assembly level. Then, uncertainty in prediction of peaking factor due to the most sensitive input parameter has been estimated using the statistical sampling method, often called the brute force method. For our analysis, the two-dimensional transport lattice code DeCART2D and its ENDF/B-VII.1 based 47-group library were used to perform the lattice physics calculation. Sensitivity calculations have been performed in order to study the influence of manufacturing tolerances on the lattice parameters. The manufacturing tolerance that has the largest influence on the k-inf is the fuel density. The second most sensitive parameter is the outer clad diameter.

  16. Time-varying coefficient estimation in SURE models. Application to portfolio management

    DEFF Research Database (Denmark)

    Casas, Isabel; Ferreira, Eva; Orbe, Susan

    This paper provides a detailed analysis of the asymptotic properties of a kernel estimator for a Seemingly Unrelated Regression Equations model with time-varying coefficients (tv-SURE) under very general conditions. Theoretical results together with a simulation study differentiates the cases...

  17. A multi-mode real-time terrain parameter estimation method for wheeled motion control of mobile robots

    Science.gov (United States)

    Li, Yuankai; Ding, Liang; Zheng, Zhizhong; Yang, Qizhi; Zhao, Xingang; Liu, Guangjun

    2018-05-01

    For motion control of wheeled planetary rovers traversing on deformable terrain, real-time terrain parameter estimation is critical in modeling the wheel-terrain interaction and compensating the effect of wheel slipping. A multi-mode real-time estimation method is proposed in this paper to achieve accurate terrain parameter estimation. The proposed method is composed of an inner layer for real-time filtering and an outer layer for online update. In the inner layer, sinkage exponent and internal frictional angle, which have higher sensitivity than that of the other terrain parameters to wheel-terrain interaction forces, are estimated in real time by using an adaptive robust extended Kalman filter (AREKF), whereas the other parameters are fixed with nominal values. The inner layer result can help synthesize the current wheel-terrain contact forces with adequate precision, but has limited prediction capability for time-variable wheel slipping. To improve estimation accuracy of the result from the inner layer, an outer layer based on recursive Gauss-Newton (RGN) algorithm is introduced to refine the result of real-time filtering according to the innovation contained in the history data. With the two-layer structure, the proposed method can work in three fundamental estimation modes: EKF, REKF and RGN, making the method applicable for flat, rough and non-uniform terrains. Simulations have demonstrated the effectiveness of the proposed method under three terrain types, showing the advantages of introducing the two-layer structure.

  18. Alcohol Prices and Mortality Due to Liver Cirrhosis

    Directory of Open Access Journals (Sweden)

    Jon P. Nelson

    2015-06-01

    Full Text Available This study estimates a reduced-form regression model for mortality rates due to alcoholic liver diseases, with alcohol prices and income as explanatory variables. Panel data cover the years 2000-2010 for 21 member countries of the European Union. In the reduced form, prices affect mortality rates indirectly through the demand for alcohol, while income has potential direct and indirect effects. Country and time fixed effects are used to control for other factors that influence alcohol consumption and mortality. Special attention is paid to outliers in the data, and final results are based on the MS-estimator for robust regressions. Regression results for alcohol prices and income are sensitive to adjustments for stationary data and down-weighting of outliers and other influential data points. Final results indicate that alcohol prices do not affect mortality rates due to chronic liver diseases. Empirical results in the study do not lend support to broad price-based approaches to alcohol policy.

  19. Estimating physical activity in children: impact of pedometer wear time and metric.

    Science.gov (United States)

    Laurson, Kelly R; Welk, Gregory J; Eisenmann, Joey C

    2015-01-01

    The purpose of this study was to provide a practical demonstration of the impact of monitoring frame and metric when assessing pedometer-determined physical activity (PA) in youth. Children (N = 1111) were asked to wear pedometers over a 7-day period during which time worn and steps were recorded each day. Varying data-exclusion criteria were used to demonstrate changes in estimates of PA. Steps were expressed using several metrics and criteria, and construct validity was demonstrated via correlations with adiposity. Meaningful fluctuations in average steps per day and percentage meeting PA recommendations were apparent when different criteria were used. Children who wore the pedometer longer appeared more active, with each minute the pedometer was worn each day accounting for an approximate increase of 11 and 8 steps for boys and girls, respectively (P < .05). Using more restrictive exclusion criteria led to stronger correlations between indices of steps per day, steps per minute, steps per leg length, steps per minute per leg length, and obesity. Wear time has a meaningful impact on estimates of PA. This should be considered when determining exclusion criteria and making comparisons between studies. Results also suggest that incorporating wear time per day and leg length into the metric may increase validity of PA estimates.

  20. Real-Time Algebraic Derivative Estimations Using a Novel Low-Cost Architecture Based on Reconfigurable Logic

    Science.gov (United States)

    Morales, Rafael; Rincón, Fernando; Gazzano, Julio Dondo; López, Juan Carlos

    2014-01-01

    Time derivative estimation of signals plays a very important role in several fields, such as signal processing and control engineering, just to name a few of them. For that purpose, a non-asymptotic algebraic procedure for the approximate estimation of the system states is used in this work. The method is based on results from differential algebra and furnishes some general formulae for the time derivatives of a measurable signal in which two algebraic derivative estimators run simultaneously, but in an overlapping fashion. The algebraic derivative algorithm presented in this paper is computed online and in real-time, offering high robustness properties with regard to corrupting noises, versatility and ease of implementation. Besides, in this work, we introduce a novel architecture to accelerate this algebraic derivative estimator using reconfigurable logic. The core of the algorithm is implemented in an FPGA, improving the speed of the system and achieving real-time performance. Finally, this work proposes a low-cost platform for the integration of hardware in the loop in MATLAB. PMID:24859033

  1. Numerical estimation of structural integrity of salt cavern wells.

    NARCIS (Netherlands)

    Orlic, B.; Thienen-Visser, K. van; Schreppers, G.J.

    2016-01-01

    Finite element analyses were performed to estimate axial deformation of cavern wells due to gas storage operations in solution-mined salt caverns. Caverns shrink over time due to salt creep and the cavern roof subsides potentially threatening well integrity. Cavern deformation, deformation of salt

  2. State of the art in evacuation time estimate studies for nuclear power plants

    International Nuclear Information System (INIS)

    Urbanik, T.E.; Jamison, J.D.

    1992-03-01

    In the event of a major accident at a commercial nuclear power station, exposure of the public to airborne radioactive materials can be prevented or greatly reduced by evacuating the area immediately surrounding the reactor site. Reactor licensees are required to conduct studies to estimate the time needed to evacuate the public from the area surrounding each nuclear power station. The results of such studies are used by regulatory personnel and emergency planners to assess the potential effectiveness of protective responses for the public. The time required to evacuate the public from a 10-mile emergency planning radius is estimated by analyzing the available transportation facilities and other relevant conditions within this radius. To support the analysis, data must be collected and assumptions must be made regarding the transportation facilities, the size and characteristics of the population and other conditions in the planning zone. This report describes standard approaches and provides recommendations regarding the relevant information, assumptions and methods to be used in performing evacuation time estimate studies

  3. Estimating the waiting time of multi-priority emergency patients with downstream blocking.

    Science.gov (United States)

    Lin, Di; Patrick, Jonathan; Labeau, Fabrice

    2014-03-01

    To characterize the coupling effect between patient flow to access the emergency department (ED) and that to access the inpatient unit (IU), we develop a model with two connected queues: one upstream queue for the patient flow to access the ED and one downstream queue for the patient flow to access the IU. Building on this patient flow model, we employ queueing theory to estimate the average waiting time across patients. Using priority specific wait time targets, we further estimate the necessary number of ED and IU resources. Finally, we investigate how an alternative way of accessing ED (Fast Track) impacts the average waiting time of patients as well as the necessary number of ED/IU resources. This model as well as the analysis on patient flow can help the designer or manager of a hospital make decisions on the allocation of ED/IU resources in a hospital.

  4. Dynamic temperature estimation and real time emergency rating of transmission cables

    DEFF Research Database (Denmark)

    Olsen, R. S.; Holboll, J.; Gudmundsdottir, Unnur Stella

    2012-01-01

    enables real time emergency ratings, such that the transmission system operator can make well-founded decisions during faults. Hereunder is included the capability of producing high resolution loadability vs. time schedules within few minutes, such that the TSO can safely control the system.......). It is found that the calculated temperature estimations are fairly accurate — within 1.5oC of the finite element method (FEM) simulation to which it is compared — both when looking at the temperature profile (time dependent) and the temperature distribution (geometric dependent). The methodology moreover...

  5. Estimation of system parameters in discrete dynamical systems from time series

    International Nuclear Information System (INIS)

    Palaniyandi, P.; Lakshmanan, M.

    2005-01-01

    We propose a simple method to estimate the parameters involved in discrete dynamical systems from time series. The method is based on the concept of controlling chaos by constant feedback. The major advantages of the method are that it needs a minimal number of time series data (either vector or scalar) and is applicable to dynamical systems of any dimension. The method also works extremely well even in the presence of noise in the time series. The method is specifically illustrated by means of logistic and Henon maps

  6. DOTD support for UTC project : travel time estimation using bluetooth, [research project capsule].

    Science.gov (United States)

    2013-10-01

    Travel time estimates are useful tools for measuring congestion in an urban area. Current : practice involves using probe vehicles or video cameras to measure travel time, but this is a laborintensive and expensive means of obtaining the information....

  7. Recent salmon declines: a result of lost feeding opportunities due to bad timing?

    Directory of Open Access Journals (Sweden)

    Cedar M Chittenden

    Full Text Available As the timing of spring productivity blooms in near-shore areas advances due to warming trends in global climate, the selection pressures on out-migrating salmon smolts are shifting. Species and stocks that leave natal streams earlier may be favoured over later-migrating fish. The low post-release survival of hatchery fish during recent years may be in part due to static release times that do not take the timing of plankton blooms into account. This study examined the effects of release time on the migratory behaviour and survival of wild and hatchery-reared coho salmon (Oncorhynchus kisutch using acoustic and coded-wire telemetry. Plankton monitoring and near-shore seining were also conducted to determine which habitat and food sources were favoured. Acoustic tags (n = 140 and coded-wire tags (n = 266,692 were implanted into coho salmon smolts at the Seymour and Quinsam Rivers, in British Columbia, Canada. Differences between wild and hatchery fish, and early and late releases were examined during the entire lifecycle. Physiological sampling was also carried out on 30 fish from each release group. The smolt-to-adult survival of coho salmon released during periods of high marine productivity was 1.5- to 3-fold greater than those released both before and after, and the fish's degree of smoltification affected their downstream migration time and duration of stay in the estuary. Therefore, hatchery managers should consider having smolts fully developed and ready for release during the peak of the near-shore plankton blooms. Monitoring chlorophyll a levels and water temperature early in the spring could provide a forecast of the timing of these blooms, giving hatcheries time to adjust their release schedule.

  8. The effects of resonances on time delay estimation for water leak detection in plastic pipes

    Science.gov (United States)

    Almeida, Fabrício C. L.; Brennan, Michael J.; Joseph, Phillip F.; Gao, Yan; Paschoalini, Amarildo T.

    2018-04-01

    In the use of acoustic correlation methods for water leak detection, sensors are placed at pipe access points either side of a suspected leak, and the peak in the cross-correlation function of the measured signals gives the time difference (delay) between the arrival times of the leak noise at the sensors. Combining this information with the speed at which the leak noise propagates along the pipe, gives an estimate for the location of the leak with respect to one of the measurement positions. It is possible for the structural dynamics of the pipe system to corrupt the time delay estimate, which results in the leak being incorrectly located. In this paper, data from test-rigs in the United Kingdom and Canada are used to demonstrate this phenomenon, and analytical models of resonators are coupled with a pipe model to replicate the experimental results. The model is then used to investigate which of the two commonly used correlation algorithms, the Basic Cross-Correlation (BCC) function or the Phase Transform (PHAT), is more robust to the undesirable structural dynamics of the pipe system. It is found that time delay estimation is highly sensitive to the frequency bandwidth over which the analysis is conducted. Moreover, it is found that the PHAT is particularly sensitive to the presence of resonances and can give an incorrect time delay estimate, whereas the BCC function is found to be much more robust, giving a consistently accurate time delay estimate for a range of dynamic conditions.

  9. Quantitative analysis of the improvement in high zoom maritime tracking due to real-time image enhancement

    CSIR Research Space (South Africa)

    Bachoo, AK

    2011-04-01

    Full Text Available This work aims to evaluate the improvement in the performance of tracking small maritime targets due to real-time enhancement of the video streams from high zoom cameras on pan-tilt pedestal. Due to atmospheric conditions these images can frequently...

  10. Time estimation in mild Alzheimer's disease patients

    Directory of Open Access Journals (Sweden)

    Nichelli Paolo

    2009-07-01

    Full Text Available Abstract Background Time information processing relies on memory, which greatly supports the operations of hypothetical internal timekeepers. Scalar Expectancy Theory (SET postulates the existence of a memory component that is functionally separated from an internal clock and other processing stages. SET has devised several experimental procedures to map these cognitive stages onto cerebral regions and neurotransmitter systems. One of these, the time bisection procedure, has provided support for a dissociation between the clock stage, controlled by dopaminergic systems, and the memory stage, mainly supported by cholinergic neuronal networks. This study aimed at linking the specific memory processes predicted by SET to brain mechanisms, by submitting time bisection tasks to patients with probable Alzheimer's disease (AD, that are known to present substantial degeneration of the fronto-temporal regions underpinning memory. Methods Twelve mild AD patients were required to make temporal judgments about intervals either ranging from 100 to 600 ms (short time bisection task or from 1000 to 3000 ms (long time bisection task. Their performance was compared with that of a group of aged-matched control participants and a group of young control subjects. Results Long time bisection scores of AD patients were not significantly different from those of the two control groups. In contrast, AD patients showed increased variability (as indexed by increased WR values in timing millisecond durations and a generalized inconsistency of responses over the same interval in both the short and long bisection tasks. A similar, though milder, decreased millisecond interval sensitivity was found for elderly subjects. Conclusion The present results, that are consistent with those of previous timing studies in AD, are interpreted within the SET framework as not selectively dependent on working or reference memory disruptions but as possibly due to distortions in different

  11. Model parameter estimations from residual gravity anomalies due to simple-shaped sources using Differential Evolution Algorithm

    Science.gov (United States)

    Ekinci, Yunus Levent; Balkaya, Çağlayan; Göktürkler, Gökhan; Turan, Seçil

    2016-06-01

    An efficient approach to estimate model parameters from residual gravity data based on differential evolution (DE), a stochastic vector-based metaheuristic algorithm, has been presented. We have showed the applicability and effectiveness of this algorithm on both synthetic and field anomalies. According to our knowledge, this is a first attempt of applying DE for the parameter estimations of residual gravity anomalies due to isolated causative sources embedded in the subsurface. The model parameters dealt with here are the amplitude coefficient (A), the depth and exact origin of causative source (zo and xo, respectively) and the shape factors (q and ƞ). The error energy maps generated for some parameter pairs have successfully revealed the nature of the parameter estimation problem under consideration. Noise-free and noisy synthetic single gravity anomalies have been evaluated with success via DE/best/1/bin, which is a widely used strategy in DE. Additionally some complicated gravity anomalies caused by multiple source bodies have been considered, and the results obtained have showed the efficiency of the algorithm. Then using the strategy applied in synthetic examples some field anomalies observed for various mineral explorations such as a chromite deposit (Camaguey district, Cuba), a manganese deposit (Nagpur, India) and a base metal sulphide deposit (Quebec, Canada) have been considered to estimate the model parameters of the ore bodies. Applications have exhibited that the obtained results such as the depths and shapes of the ore bodies are quite consistent with those published in the literature. Uncertainty in the solutions obtained from DE algorithm has been also investigated by Metropolis-Hastings (M-H) sampling algorithm based on simulated annealing without cooling schedule. Based on the resulting histogram reconstructions of both synthetic and field data examples the algorithm has provided reliable parameter estimations being within the sampling limits of

  12. Optimal State Estimation for Discrete-Time Markov Jump Systems with Missing Observations

    Directory of Open Access Journals (Sweden)

    Qing Sun

    2014-01-01

    Full Text Available This paper is concerned with the optimal linear estimation for a class of direct-time Markov jump systems with missing observations. An observer-based approach of fault detection and isolation (FDI is investigated as a detection mechanic of fault case. For systems with known information, a conditional prediction of observations is applied and fault observations are replaced and isolated; then, an FDI linear minimum mean square error estimation (LMMSE can be developed by comprehensive utilizing of the correct information offered by systems. A recursive equation of filtering based on the geometric arguments can be obtained. Meanwhile, a stability of the state estimator will be guaranteed under appropriate assumption.

  13. Transfer Entropy Estimation and Directional Coupling Change Detection in Biomedical Time Series

    Directory of Open Access Journals (Sweden)

    Lee Joon

    2012-04-01

    Full Text Available Abstract Background The detection of change in magnitude of directional coupling between two non-linear time series is a common subject of interest in the biomedical domain, including studies involving the respiratory chemoreflex system. Although transfer entropy is a useful tool in this avenue, no study to date has investigated how different transfer entropy estimation methods perform in typical biomedical applications featuring small sample size and presence of outliers. Methods With respect to detection of increased coupling strength, we compared three transfer entropy estimation techniques using both simulated time series and respiratory recordings from lambs. The following estimation methods were analyzed: fixed-binning with ranking, kernel density estimation (KDE, and the Darbellay-Vajda (D-V adaptive partitioning algorithm extended to three dimensions. In the simulated experiment, sample size was varied from 50 to 200, while coupling strength was increased. In order to introduce outliers, the heavy-tailed Laplace distribution was utilized. In the lamb experiment, the objective was to detect increased respiratory-related chemosensitivity to O2 and CO2 induced by a drug, domperidone. Specifically, the separate influence of end-tidal PO2 and PCO2 on minute ventilation (V˙E before and after administration of domperidone was analyzed. Results In the simulation, KDE detected increased coupling strength at the lowest SNR among the three methods. In the lamb experiment, D-V partitioning resulted in the statistically strongest increase in transfer entropy post-domperidone for PO2→V˙E. In addition, D-V partitioning was the only method that could detect an increase in transfer entropy for PCO2→V˙E, in agreement with experimental findings. Conclusions Transfer entropy is capable of detecting directional coupling changes in non-linear biomedical time series analysis featuring a small number of observations and presence of outliers. The results

  14. Estimating the Probability of a Rare Event Over a Finite Time Horizon

    NARCIS (Netherlands)

    de Boer, Pieter-Tjerk; L'Ecuyer, Pierre; Rubino, Gerardo; Tuffin, Bruno

    2007-01-01

    We study an approximation for the zero-variance change of measure to estimate the probability of a rare event in a continuous-time Markov chain. The rare event occurs when the chain reaches a given set of states before some fixed time limit. The jump rates of the chain are expressed as functions of

  15. Uncertainty in urban flood damage assessment due to urban drainage modelling and depth-damage curve estimation.

    Science.gov (United States)

    Freni, G; La Loggia, G; Notaro, V

    2010-01-01

    Due to the increased occurrence of flooding events in urban areas, many procedures for flood damage quantification have been defined in recent decades. The lack of large databases in most cases is overcome by combining the output of urban drainage models and damage curves linking flooding to expected damage. The application of advanced hydraulic models as diagnostic, design and decision-making support tools has become a standard practice in hydraulic research and application. Flooding damage functions are usually evaluated by a priori estimation of potential damage (based on the value of exposed goods) or by interpolating real damage data (recorded during historical flooding events). Hydraulic models have undergone continuous advancements, pushed forward by increasing computer capacity. The details of the flooding propagation process on the surface and the details of the interconnections between underground and surface drainage systems have been studied extensively in recent years, resulting in progressively more reliable models. The same level of was advancement has not been reached with regard to damage curves, for which improvements are highly connected to data availability; this remains the main bottleneck in the expected flooding damage estimation. Such functions are usually affected by significant uncertainty intrinsically related to the collected data and to the simplified structure of the adopted functional relationships. The present paper aimed to evaluate this uncertainty by comparing the intrinsic uncertainty connected to the construction of the damage-depth function to the hydraulic model uncertainty. In this way, the paper sought to evaluate the role of hydraulic model detail level in the wider context of flood damage estimation. This paper demonstrated that the use of detailed hydraulic models might not be justified because of the higher computational cost and the significant uncertainty in damage estimation curves. This uncertainty occurs mainly

  16. Uncertainty in techno-economic estimates of cellulosic ethanol production due to experimental measurement uncertainty

    Directory of Open Access Journals (Sweden)

    Vicari Kristin J

    2012-04-01

    Full Text Available Abstract Background Cost-effective production of lignocellulosic biofuels remains a major financial and technical challenge at the industrial scale. A critical tool in biofuels process development is the techno-economic (TE model, which calculates biofuel production costs using a process model and an economic model. The process model solves mass and energy balances for each unit, and the economic model estimates capital and operating costs from the process model based on economic assumptions. The process model inputs include experimental data on the feedstock composition and intermediate product yields for each unit. These experimental yield data are calculated from primary measurements. Uncertainty in these primary measurements is propagated to the calculated yields, to the process model, and ultimately to the economic model. Thus, outputs of the TE model have a minimum uncertainty associated with the uncertainty in the primary measurements. Results We calculate the uncertainty in the Minimum Ethanol Selling Price (MESP estimate for lignocellulosic ethanol production via a biochemical conversion process: dilute sulfuric acid pretreatment of corn stover followed by enzymatic hydrolysis and co-fermentation of the resulting sugars to ethanol. We perform a sensitivity analysis on the TE model and identify the feedstock composition and conversion yields from three unit operations (xylose from pretreatment, glucose from enzymatic hydrolysis, and ethanol from fermentation as the most important variables. The uncertainty in the pretreatment xylose yield arises from multiple measurements, whereas the glucose and ethanol yields from enzymatic hydrolysis and fermentation, respectively, are dominated by a single measurement: the fraction of insoluble solids (fIS in the biomass slurries. Conclusions We calculate a $0.15/gal uncertainty in MESP from the TE model due to uncertainties in primary measurements. This result sets a lower bound on the error bars of

  17. Time delay estimation in a reverberant environment by low rate sampling of impulsive acoustic sources

    KAUST Repository

    Omer, Muhammad

    2012-07-01

    This paper presents a new method of time delay estimation (TDE) using low sample rates of an impulsive acoustic source in a room environment. The proposed method finds the time delay from the room impulse response (RIR) which makes it robust against room reverberations. The RIR is considered a sparse phenomenon and a recently proposed sparse signal reconstruction technique called orthogonal clustering (OC) is utilized for its estimation from the low rate sampled received signal. The arrival time of the direct path signal at a pair of microphones is identified from the estimated RIR and their difference yields the desired time delay. Low sampling rates reduce the hardware and computational complexity and decrease the communication between the microphones and the centralized location. The performance of the proposed technique is demonstrated by numerical simulations and experimental results. © 2012 IEEE.

  18. Statistical properties of the anomalous scaling exponent estimator based on time-averaged mean-square displacement

    Science.gov (United States)

    Sikora, Grzegorz; Teuerle, Marek; Wyłomańska, Agnieszka; Grebenkov, Denis

    2017-08-01

    The most common way of estimating the anomalous scaling exponent from single-particle trajectories consists of a linear fit of the dependence of the time-averaged mean-square displacement on the lag time at the log-log scale. We investigate the statistical properties of this estimator in the case of fractional Brownian motion (FBM). We determine the mean value, the variance, and the distribution of the estimator. Our theoretical results are confirmed by Monte Carlo simulations. In the limit of long trajectories, the estimator is shown to be asymptotically unbiased, consistent, and with vanishing variance. These properties ensure an accurate estimation of the scaling exponent even from a single (long enough) trajectory. As a consequence, we prove that the usual way to estimate the diffusion exponent of FBM is correct from the statistical point of view. Moreover, the knowledge of the estimator distribution is the first step toward new statistical tests of FBM and toward a more reliable interpretation of the experimental histograms of scaling exponents in microbiology.

  19. Simple methodologies to estimate the energy amount stored in a tree due to an explosive seed dispersal mechanism

    Science.gov (United States)

    do Carmo, Eduardo; Goncalves Hönnicke, Marcelo

    2018-05-01

    There are different forms to introduce/illustrate the energy concepts for the basic physics students. The explosive seed dispersal mechanism found in a variety of trees could be one of them. Sibipiruna trees carry out fruits (pods) who show such an explosive mechanism. During the explosion, the pods throw out seeds several meters away. In this manuscript we show simple methodologies to estimate the energy amount stored in the Sibipiruna tree due to such a process. Two different physics approaches were used to carry out this study: by monitoring indoor and in situ the explosive seed dispersal mechanism and by measuring the elastic constant of the pod shell. An energy of the order of kJ was found to be stored in a single tree due to such an explosive mechanism.

  20. Assessment of Evolving TRMM-Based Real-Time Precipitation Estimation Methods and Their Impacts on Hydrologic Prediction in a High-Latitude Basin

    Science.gov (United States)

    Yong, Bin; Hong, Yang; Ren, Li-Liang; Gourley, Jonathan; Huffman, George J.; Chen, Xi; Wang, Wen; Khan, Sadiq I.

    2013-01-01

    The real-time availability of satellite-derived precipitation estimates provides hydrologists an opportunity to improve current hydrologic prediction capability for medium to large river basins. Due to the availability of new satellite data and upgrades to the precipitation algorithms, the Tropical Rainfall Measuring Mission (TRMM) Multi-satellite Precipitation Analysis real-time estimates (TMPA-RT) have been undergoing several important revisions over the past ten years. In this study, the changes of the relative accuracy and hydrologic potential of TMPA-RT estimates over its three major evolving periods were evaluated and inter-compared at daily, monthly and seasonal scales in the high-latitude Laohahe basin in China. Assessment results show that the performance of TMPA-RT in terms of precipitation estimation and streamflow simulation was significantly improved after 3 February 2005. Overestimation during winter months was noteworthy and consistent, which is suggested to be a consequence from interference of snow cover to the passive microwave retrievals. Rainfall estimated by the new version 6 of TMPA-RT starting from 1 October 2008 to present has higher correlations with independent gauge observations and tends to perform better in detecting rain compared to the prior periods, although it suffers larger mean error and relative bias. After a simple bias correction, this latest dataset of TMPA-RT exhibited the best capability in capturing hydrologic response among the three tested periods. In summary, this study demonstrated that there is an increasing potential in the use of TMPA-RT in hydrologic streamflow simulations over its three algorithm upgrade periods, but still with significant challenges during the winter snowing events.

  1. Real-time yield estimation based on deep learning

    Science.gov (United States)

    Rahnemoonfar, Maryam; Sheppard, Clay

    2017-05-01

    Crop yield estimation is an important task in product management and marketing. Accurate yield prediction helps farmers to make better decision on cultivation practices, plant disease prevention, and the size of harvest labor force. The current practice of yield estimation based on the manual counting of fruits is very time consuming and expensive process and it is not practical for big fields. Robotic systems including Unmanned Aerial Vehicles (UAV) and Unmanned Ground Vehicles (UGV), provide an efficient, cost-effective, flexible, and scalable solution for product management and yield prediction. Recently huge data has been gathered from agricultural field, however efficient analysis of those data is still a challenging task. Computer vision approaches currently face diffident challenges in automatic counting of fruits or flowers including occlusion caused by leaves, branches or other fruits, variance in natural illumination, and scale. In this paper a novel deep convolutional network algorithm was developed to facilitate the accurate yield prediction and automatic counting of fruits and vegetables on the images. Our method is robust to occlusion, shadow, uneven illumination and scale. Experimental results in comparison to the state-of-the art show the effectiveness of our algorithm.

  2. Parametric estimation of time varying baselines in airborne interferometric SAR

    DEFF Research Database (Denmark)

    Mohr, Johan Jacob; Madsen, Søren Nørvang

    1996-01-01

    A method for estimation of time varying spatial baselines in airborne interferometric synthetic aperture radar (SAR) is described. The range and azimuth distortions between two images acquired with a non-linear baseline are derived. A parametric model of the baseline is then, in a least square...... sense, estimated from image shifts obtained by cross correlation of numerous small patches throughout the image. The method has been applied to airborne EMISAR imagery from the 1995 campaign over the Storstrommen Glacier in North East Greenland conducted by the Danish Center for Remote Sensing. This has...... reduced the baseline uncertainties from several meters to the centimeter level in a 36 km scene. Though developed for airborne SAR the method can easily be adopted to satellite data...

  3. Estimation of changes in saturation and pressure from 4D seismic AVO and time-shift analysis

    NARCIS (Netherlands)

    Trani, M.; Arts, R.; Leeuwenburgh, O.; Brouwer, J.

    2011-01-01

    A reliable estimate of reservoir pressure and fluid saturation changes from time-lapse seismic data is difficult to obtain. Existing methods generally suffer from leakage between the estimated parameters. We propose a new method using different combinations of time-lapse seismic attributes based on

  4. Estimation and Properties of a Time-Varying GQARCH(1,1-M Model

    Directory of Open Access Journals (Sweden)

    Sofia Anyfantaki

    2011-01-01

    analysis of these models computationally infeasible. This paper outlines the issues and suggests to employ a Markov chain Monte Carlo algorithm which allows the calculation of a classical estimator via the simulated EM algorithm or a simulated Bayesian solution in only ( computational operations, where is the sample size. Furthermore, the theoretical dynamic properties of a time-varying GQARCH(1,1-M are derived. We discuss them and apply the suggested Bayesian estimation to three major stock markets.

  5. On Assessment and Estimation of Potential Losses due to Land Subsidence in Urban Areas of Indonesia

    Science.gov (United States)

    Abidin, Hasanuddin Z.; Andreas, Heri; Gumilar, Irwan; Sidiq, Teguh P.

    2016-04-01

    subsidence have also relation among each other, the accurate quantification of the potential losses caused by land subsidence in urban areas is not an easy task to accomplish. The direct losses can be easier to estimate than the indirect losses. For example, the direct losses due to land subsidence in Bandung was estimated to be at least 180 Million USD; but the indirect losses is still unknown.

  6. Flow modelling to estimate suspended sediment travel times for two Canadian Deltas

    Directory of Open Access Journals (Sweden)

    S. R. Fassnacht

    2000-01-01

    Full Text Available The approximate travel times for suspended sediment transport through two multi-channel networks are estimated using flow modelling. The focus is on the movement of high sediment concentrations that travel rapidly downstream. Since suspended sediment transport through river confluences and bifurcation movement is poorly understood, it is assumed that the sediment moves at approximately the average channel velocity during periods of high sediment load movement. Calibration of the flow model is discussed, with an emphasis on the incorporation of cross-section data, that are not referenced to a datum, using a continuous water surface profile. Various flow regimes are examined for the Mackenzie and the Slave River Deltas in the Northwest Territories, Canada, and a significant variation in travel times is illustrated. One set of continuous daily sediment measurements throughout the Mackenzie Delta is used to demonstrate that the travel time estimates are reasonable. Keywords: suspended sediment; multi-channel river systems; flow modelling; sediment transport

  7. Performance estimation of control rod position indicator due to aging of magnet

    International Nuclear Information System (INIS)

    Yu, Je Yong; Kim, Ji Ho; Huh, Hyung; Choi, Myoung Hwan; Sohn, Dong Seong

    2009-01-01

    The Control Element Drive Mechanism (CEDM) for the integral reactor is designed to raise and lower the control rod in steps of 2mm in order to satisfy the design features of the integral reactor which are the soluble boron free operation and the use of a nuclear heating for the reactor start-up. The actual position of the control rod could be achieved to sense the magnet connected to the control rod by the position indicator around the upper pressure housing of CEDM. It is sufficient that the actual position information of control rod at 20mm interval from the position indicator is used for the core safety analysis. As the magnet moves upward along the position indicator assembly from the bottom to the top in the upper pressure housing, the output voltage increases linearly step-wise at 0.2VDC increments. Between every step there are transient areas which occur by a contact closing of three reed switches which is the 2-3-2 contact closing sequence. In this paper the output voltage signal corresponding to the position of control rod was estimated on the 2-1-2 contact closing sequence due to the aging of the magnet.

  8. Avian mortalities due to transmission line collisions: a review of current estimates and field methods with an emphasis on applications to the Canadian electric network

    Directory of Open Access Journals (Sweden)

    Sébastien Rioux

    2013-12-01

    Full Text Available Birds are vulnerable to collisions with human-made fixed structures. Despite ongoing development and increases in infrastructure, we have few estimates of the magnitude of collision mortality. We reviewed the existing literature on avian mortality associated with transmission lines and derived an initial estimate for Canada. Estimating mortality from collisions with power lines is challenging due to the lack of studies, especially from sites within Canada, and due to uncertainty about the magnitude of detection biases. Detection of bird collisions with transmission lines varies due to habitat type, species size, and scavenging rates. In addition, birds can be crippled by the impact and subsequently die, although crippling rates are poorly known and rarely incorporated into estimates. We used existing data to derive a range of estimates of avian mortality associated with collisions with transmission lines in Canada by incorporating detection, scavenging, and crippling biases. There are 231,966 km of transmission lines across Canada, mostly in the boreal forest. Mortality estimates ranged from 1 million to 229.5 million birds per year, depending on the bias corrections applied. We consider our most realistic estimate, taking into account variation in risk across Canada, to range from 2.5 million to 25.6 million birds killed per year. Data from multiple studies across Canada and the northern U.S. indicate that the most vulnerable bird groups are (1 waterfowl, (2 grebes, (3 shorebirds, and (4 cranes, which is consistent with other studies. Populations of several groups that are vulnerable to collisions are increasing across Canada (e.g., waterfowl, raptors, which suggests that collision mortality, at current levels, is not limiting population growth. However, there may be impacts on other declining species, such as shorebirds and some species at risk, including Alberta's Trumpeter Swans (Cygnus buccinator and western Canada's endangered Whooping

  9. Estimating the Value of Life, Injury, and Travel Time Saved Using a Stated Preference Framework.

    Science.gov (United States)

    Niroomand, Naghmeh; Jenkins, Glenn P

    2016-06-01

    The incidence of fatality over the period 2010-2014 from automobile accidents in North Cyprus is 2.75 times greater than the average for the EU. With the prospect of North Cyprus entering the EU, many investments will need to be undertaken to improve road safety in order to reach EU benchmarks. The objective of this study is to provide local estimates of the value of a statistical life and injury along with the value of time savings. These are among the parameter values needed for the evaluation of the change in the expected incidence of automotive accidents and time savings brought about by such projects. In this study we conducted a stated choice experiment to identify the preferences and tradeoffs of automobile drivers in North Cyprus for improved travel times, travel costs, and safety. The choice of route was examined using mixed logit models to obtain the marginal utilities associated with each attribute of the routes that consumers choose. These estimates were used to assess the individuals' willingness to pay (WTP) to avoid fatalities and injuries and to save travel time. We then used the results to obtain community-wide estimates of the value of a statistical life (VSL) saved, the value of injury (VI) prevented, and the value per hour of travel time saved. The estimates for the VSL range from €315,293 to €1,117,856 and the estimates of VI from € 5,603 to € 28,186. These values are consistent, after adjusting for differences in incomes, with the median results of similar studies done for EU countries. Copyright © 2016 Elsevier Ltd. All rights reserved.

  10. Improved target detection and bearing estimation utilizing fast orthogonal search for real-time spectral analysis

    International Nuclear Information System (INIS)

    Osman, Abdalla; El-Sheimy, Naser; Nourledin, Aboelamgd; Theriault, Jim; Campbell, Scott

    2009-01-01

    The problem of target detection and tracking in the ocean environment has attracted considerable attention due to its importance in military and civilian applications. Sonobuoys are one of the capable passive sonar systems used in underwater target detection. Target detection and bearing estimation are mainly obtained through spectral analysis of received signals. The frequency resolution introduced by current techniques is limited which affects the accuracy of target detection and bearing estimation at a relatively low signal-to-noise ratio (SNR). This research investigates the development of a bearing estimation method using fast orthogonal search (FOS) for enhanced spectral estimation. FOS is employed in this research in order to improve both target detection and bearing estimation in the case of low SNR inputs. The proposed methods were tested using simulated data developed for two different scenarios under different underwater environmental conditions. The results show that the proposed method is capable of enhancing the accuracy for target detection as well as bearing estimation especially in cases of a very low SNR

  11. Estimates of grassland biomass and turnover time on the Tibetan Plateau

    Science.gov (United States)

    Xia, Jiangzhou; Ma, Minna; Liang, Tiangang; Wu, Chaoyang; Yang, Yuanhe; Zhang, Li; Zhang, Yangjian; Yuan, Wenping

    2018-01-01

    The grassland of the Tibetan Plateau forms a globally significant biome, which represents 6% of the world’s grasslands and 44% of China’s grasslands. However, large uncertainties remain concerning the vegetation carbon storage and turnover time in this biome. In this study, we quantified the pool size of both the aboveground and belowground biomass and turnover time of belowground biomass across the Tibetan Plateau by combining systematic measurements taken from a substantial number of surveys (i.e. 1689 sites for aboveground biomass, 174 sites for belowground biomass) with a machine learning technique (i.e. random forest, RF). Our study demonstrated that the RF model is effective tool for upscaling local biomass observations to the regional scale, and for producing continuous biomass estimates of the Tibetan Plateau. On average, the models estimated 46.57 Tg (1 Tg = 1012g) C of aboveground biomass and 363.71 Tg C of belowground biomass in the Tibetan grasslands covering an area of 1.32 × 106 km2. The turnover time of belowground biomass demonstrated large spatial heterogeneity, with a median turnover time of 4.25 years. Our results also demonstrated large differences in the biomass simulations among the major ecosystem models used for the Tibetan Plateau, largely because of inadequate model parameterization and validation. This study provides a spatially continuous measure of vegetation carbon storage and turnover time, and provides useful information for advancing ecosystem models and improving their performance.

  12. Dynamic state estimation and prediction for real-time control and operation

    NARCIS (Netherlands)

    Nguyen, P.H.; Venayagamoorthy, G.K.; Kling, W.L.; Ribeiro, P.F.

    2013-01-01

    Real-time control and operation are crucial to deal with increasing complexity of modern power systems. To effectively enable those functions, it is required a Dynamic State Estimation (DSE) function to provide accurate network state variables at the right moment and predict their trends ahead. This

  13. A simulation study to quantify the impacts of exposure measurement error on air pollution health risk estimates in copollutant time-series models.

    Science.gov (United States)

    Dionisio, Kathie L; Chang, Howard H; Baxter, Lisa K

    2016-11-25

    Exposure measurement error in copollutant epidemiologic models has the potential to introduce bias in relative risk (RR) estimates. A simulation study was conducted using empirical data to quantify the impact of correlated measurement errors in time-series analyses of air pollution and health. ZIP-code level estimates of exposure for six pollutants (CO, NO x , EC, PM 2.5 , SO 4 , O 3 ) from 1999 to 2002 in the Atlanta metropolitan area were used to calculate spatial, population (i.e. ambient versus personal), and total exposure measurement error. Empirically determined covariance of pollutant concentration pairs and the associated measurement errors were used to simulate true exposure (exposure without error) from observed exposure. Daily emergency department visits for respiratory diseases were simulated using a Poisson time-series model with a main pollutant RR = 1.05 per interquartile range, and a null association for the copollutant (RR = 1). Monte Carlo experiments were used to evaluate the impacts of correlated exposure errors of different copollutant pairs. Substantial attenuation of RRs due to exposure error was evident in nearly all copollutant pairs studied, ranging from 10 to 40% attenuation for spatial error, 3-85% for population error, and 31-85% for total error. When CO, NO x or EC is the main pollutant, we demonstrated the possibility of false positives, specifically identifying significant, positive associations for copollutants based on the estimated type I error rate. The impact of exposure error must be considered when interpreting results of copollutant epidemiologic models, due to the possibility of attenuation of main pollutant RRs and the increased probability of false positives when measurement error is present.

  14. Estimation of time-dependent Hurst exponents with variational smoothing and application to forecasting foreign exchange rates

    Science.gov (United States)

    Garcin, Matthieu

    2017-10-01

    Hurst exponents depict the long memory of a time series. For human-dependent phenomena, as in finance, this feature may vary in the time. It justifies modelling dynamics by multifractional Brownian motions, which are consistent with time-dependent Hurst exponents. We improve the existing literature on estimating time-dependent Hurst exponents by proposing a smooth estimate obtained by variational calculus. This method is very general and not restricted to the sole Hurst framework. It is globally more accurate and easier than other existing non-parametric estimation techniques. Besides, in the field of Hurst exponents, it makes it possible to make forecasts based on the estimated multifractional Brownian motion. The application to high-frequency foreign exchange markets (GBP, CHF, SEK, USD, CAD, AUD, JPY, CNY and SGD, all against EUR) shows significantly good forecasts. When the Hurst exponent is higher than 0.5, what depicts a long-memory feature, the accuracy is higher.

  15. A neural computational model for animal's time-to-collision estimation.

    Science.gov (United States)

    Wang, Ling; Yao, Dezhong

    2013-04-17

    The time-to-collision (TTC) is the time elapsed before a looming object hits the subject. An accurate estimation of TTC plays a critical role in the survival of animals in nature and acts as an important factor in artificial intelligence systems that depend on judging and avoiding potential dangers. The theoretic formula for TTC is 1/τ≈θ'/sin θ, where θ and θ' are the visual angle and its variation, respectively, and the widely used approximation computational model is θ'/θ. However, both of these measures are too complex to be implemented by a biological neuronal model. We propose a new simple computational model: 1/τ≈Mθ-P/(θ+Q)+N, where M, P, Q, and N are constants that depend on a predefined visual angle. This model, weighted summation of visual angle model (WSVAM), can achieve perfect implementation through a widely accepted biological neuronal model. WSVAM has additional merits, including a natural minimum consumption and simplicity. Thus, it yields a precise and neuronal-implemented estimation for TTC, which provides a simple and convenient implementation for artificial vision, and represents a potential visual brain mechanism.

  16. Heterogeneous Data Fusion Method to Estimate Travel Time Distributions in Congested Road Networks

    OpenAIRE

    Chaoyang Shi; Bi Yu Chen; William H. K. Lam; Qingquan Li

    2017-01-01

    Travel times in congested urban road networks are highly stochastic. Provision of travel time distribution information, including both mean and variance, can be very useful for travelers to make reliable path choice decisions to ensure higher probability of on-time arrival. To this end, a heterogeneous data fusion method is proposed to estimate travel time distributions by fusing heterogeneous data from point and interval detectors. In the proposed method, link travel time distributions are f...

  17. Analytical model for real time, noninvasive estimation of blood glucose level.

    Science.gov (United States)

    Adhyapak, Anoop; Sidley, Matthew; Venkataraman, Jayanti

    2014-01-01

    The paper presents an analytical model to estimate blood glucose level from measurements made non-invasively and in real time by an antenna strapped to a patient's wrist. Some promising success has been shown by the RIT ETA Lab research group that an antenna's resonant frequency can track, in real time, changes in glucose concentration. Based on an in-vitro study of blood samples of diabetic patients, the paper presents a modified Cole-Cole model that incorporates a factor to represent the change in glucose level. A calibration technique using the input impedance technique is discussed and the results show a good estimation as compared to the glucose meter readings. An alternate calibration methodology has been developed that is based on the shift in the antenna resonant frequency using an equivalent circuit model containing a shunt capacitor to represent the shift in resonant frequency with changing glucose levels. Work under progress is the optimization of the technique with a larger sample of patients.

  18. A novel velocity estimator using multiple frequency carriers

    DEFF Research Database (Denmark)

    Zhang, Zhuo; Jakobsson, Andreas; Nikolov, Svetoslav

    2004-01-01

    . In this paper, we propose a nonlinear least squares (NLS) estimator. Typically, NLS estimators are computationally cumbersome, in general requiring the minimization of a multidimensional and often multimodal cost function. Here, by noting that the unknown velocity will result in a common known frequency......Most modern ultrasound scanners use the so-called pulsed-wave Doppler technique to estimate the blood velocities. Among the narrowband-based methods, the autocorrelation estimator and the Fourier-based method are the most commonly used approaches. Due to the low level of the blood echo, the signal......-to-noise ratio is low, and some averaging in depth is applied to improve the estimate. Further, due to velocity gradients in space and time, the spectrum may get smeared. An alternative approach is to use a pulse with multiple frequency carriers, and do some form of averaging in the frequency domain. However...

  19. New perspectives on the damage estimation for buried pipeline systems due to seismic wave propagation

    Energy Technology Data Exchange (ETDEWEB)

    Pineda Porras, Omar Andrey [Los Alamos National Laboratory

    2009-01-01

    Over the past three decades, seismic fragility fonnulations for buried pipeline systems have been developed following two tendencies: the use of earthquake damage scenarios from several pipeline systems to create general pipeline fragility functions; and, the use of damage scenarios from one pipeline system to create specific-system fragility functions. In this paper, the advantages and disadvantages of both tendencies are analyzed and discussed; in addition, a summary of what can be considered the new challenges for developing better pipeline seismic fragility formulations is discussed. The most important conclusion of this paper states that more efforts are needed to improve the estimation of transient ground strain -the main cause of pipeline damage due to seismic wave propagation; with relevant advances in that research field, new and better fragility formulations could be developed.

  20. An estimation model of population in China using time series DMSP night-time satellite imagery from 2002-2010

    Science.gov (United States)

    Zhang, Xiaoyong; Zhang, Zhijie; Chang, Yuguang; Chen, Zhengchao

    2015-12-01

    Accurate data on the spatial distribution and potential growth estimation of human population are playing pivotal role in addressing and mitigating heavy lose caused by earthquake. Traditional demographic data is limited in its spatial resolution and is extremely hard to update. With the accessibility of massive DMSP/OLS night time imagery, it is possible to model population distribution at the county level across China. In order to compare and improve the continuity and consistency of time-series DMSP night-time satellite imagery obtained by different satellites in same year or different years by the same satellite from 2002-2010, normalized method was deployed for the inter-correction among imageries. And we referred to the reference F162007 Jixi city, whose social-economic has been relatively stable. Through binomial model, with average R2 0.90, then derived the correction factor of each year. The normalization obviously improved consistency comparing to previous data, which enhanced the correspondent accuracy of model. Then conducted the model of population density between average night-time light intensity in eight-economic districts. According to the two parameters variation law of consecutive years, established the prediction model of next following years with R2of slope and constant typically 0.85 to 0.95 in different regions. To validate the model, taking the year of 2005 as example, retrieved quantitatively population distribution in per square kilometer based on the model, then compared the results to the statistical data based on census, the difference of the result is acceptable. In summary, the estimation model facilitates the quick estimation and prediction in relieving the damage to people, which is significant in decision-making.

  1. Schroedinger operators - geometric estimates in terms of the occupation time

    International Nuclear Information System (INIS)

    Demuth, M.; Kirsch, W.; McGillivray, I.

    1995-01-01

    The difference of Schroedinger and Dirichlet semigroups is expressed in terms of the Laplace transform of the Brownian motion occupation time. This implies quantitative upper and lower bounds for the operator norms of the corresponding resolvent differences. One spectral theoretical consequence is an estimate for the eigenfunction for a Schroedinger operator in a ball where the potential is given as a cone indicator function. 12 refs

  2. A mathematical model for estimating the vibration due to basting in Sarcheshmeh Copper Mine

    International Nuclear Information System (INIS)

    Hossaini, M. F.; Javaherian, A.; Pourghasemi Sagand, M.

    2002-01-01

    Ground vibration due to blasting is the research subject of many investigations. A lot of works have been done in order to estimate the quality and quantity of this blasting outcome. Mathematical models proposed by various investigators are a noticeable result of these investigations. In this paper, the origin of these mathematical models are studied and the short comes are pointed out. With aid of real data a new empirical model has been proposed. This model is some modification of the previous ones with some different parameters. This investigation is based on analyzing the data obtained by operating 14 blasts in Sarcheshmeh Copper Mine whose data have been collected by seismographs installed in the area. In the proposed model, instead of the amount of charge exploded in each delay the amount of charge the vibration due to which is designed to facilitate application of a radial confining pressure to a grouted bolt while pulling it axially. During the test, both axial displacement of the bolt as well as the radial dilation of the grout was monitored. Few deformed bolts were designed and manufactured to study the effect of the shape of the ribs. While pull out, the cement that is captured between lugs will shear which in turn emphasizes the importance of shear strength of the grout annulus. In this report, details of the laboratory test results are presented and conclusions are given based on the obtained results

  3. Towards real-time body pose estimation for presenters in meeting environments

    NARCIS (Netherlands)

    Poppe, Ronald Walter; Heylen, Dirk K.J.; Nijholt, Antinus; Poel, Mannes

    2005-01-01

    This paper describes a computer vision-based approach to body pose estimation. The algorithm can be executed in real-time and processes low resolution, monocular image sequences. A silhouette is extracted and matched against a projection of a 16 DOF human body model. In addition, skin color is used

  4. A model to estimate volume change due to radiolytic gas bubbles and thermal expansion in solution reactors

    International Nuclear Information System (INIS)

    Souto, F.J.; Heger, A.S.

    2001-01-01

    To investigate the effects of radiolytic gas bubbles and thermal expansion on the steady-state operation of solution reactors at the power level required for the production of medical isotopes, a calculational model has been developed. To validate this model, including its principal hypotheses, specific experiments at the Los Alamos National Laboratory SHEBA uranyl fluoride solution reactor were conducted. The following sections describe radiolytic gas generation in solution reactors, the equations to estimate the fuel solution volume change due to radiolytic gas bubbles and thermal expansion, the experiments conducted at SHEBA, and the comparison of experimental results and model calculations. (author)

  5. Estimating the cost of production stoppage

    Science.gov (United States)

    Delionback, L. M.

    1979-01-01

    Estimation model considers learning curve quantities, and time of break to forecast losses due to break in production schedule. Major parameters capable of predicting costs are number of units made prior to production sequence, length of production break, and slope of learning curve produced prior to break.

  6. Combinations of Epoch Durations and Cut-Points to Estimate Sedentary Time and Physical Activity among Adolescents

    Science.gov (United States)

    Fröberg, Andreas; Berg, Christina; Larsson, Christel; Boldemann, Cecilia; Raustorp, Anders

    2017-01-01

    The purpose of the current study was to investigate how combinations of different epoch durations and cut-points affect the estimations of sedentary time and physical activity in adolescents. Accelerometer data from 101 adolescents were derived and 30 combinations were used to estimate sedentary time, light, moderate, vigorous, and combined…

  7. The Value Estimation of an HFGW Frequency Time Standard for Telecommunications Network Optimization

    Science.gov (United States)

    Harper, Colby; Stephenson, Gary

    2007-01-01

    The emerging technology of gravitational wave control is used to augment a communication system using a development roadmap suggested in Stephenson (2003) for applications emphasized in Baker (2005). In the present paper consideration is given to the value of a High Frequency Gravitational Wave (HFGW) channel purely as providing a method of frequency and time reference distribution for use within conventional Radio Frequency (RF) telecommunications networks. Specifically, the native value of conventional telecommunications networks may be optimized by using an unperturbed frequency time standard (FTS) to (1) improve terminal navigation and Doppler estimation performance via improved time difference of arrival (TDOA) from a universal time reference, and (2) improve acquisition speed, coding efficiency, and dynamic bandwidth efficiency through the use of a universal frequency reference. A model utilizing a discounted cash flow technique provides an estimation of the additional value using HFGW FTS technology could bring to a mixed technology HFGW/RF network. By applying a simple net present value analysis with supporting reference valuations to such a network, it is demonstrated that an HFGW FTS could create a sizable improvement within an otherwise conventional RF telecommunications network. Our conservative model establishes a low-side value estimate of approximately 50B USD Net Present Value for an HFGW FTS service, with reasonable potential high-side values to significant multiples of this low-side value floor.

  8. Errors in 'BED'-derived estimates of HIV incidence will vary by place, time and age.

    Directory of Open Access Journals (Sweden)

    Timothy B Hallett

    2009-05-01

    Full Text Available The BED Capture Enzyme Immunoassay, believed to distinguish recent HIV infections, is being used to estimate HIV incidence, although an important property of the test--how specificity changes with time since infection--has not been not measured.We construct hypothetical scenarios for the performance of BED test, consistent with current knowledge, and explore how this could influence errors in BED estimates of incidence using a mathematical model of six African countries. The model is also used to determine the conditions and the sample sizes required for the BED test to reliably detect trends in HIV incidence.If the chance of misclassification by BED increases with time since infection, the overall proportion of individuals misclassified could vary widely between countries, over time, and across age-groups, in a manner determined by the historic course of the epidemic and the age-pattern of incidence. Under some circumstances, changes in BED estimates over time can approximately track actual changes in incidence, but large sample sizes (50,000+ will be required for recorded changes to be statistically significant.The relationship between BED test specificity and time since infection has not been fully measured, but, if it decreases, errors in estimates of incidence could vary by place, time and age-group. This means that post-assay adjustment procedures using parameters from different populations or at different times may not be valid. Further research is urgently needed into the properties of the BED test, and the rate of misclassification in a wide range of populations.

  9. A time-frequency analysis method to obtain stable estimates of magnetotelluric response function based on Hilbert-Huang transform

    Science.gov (United States)

    Cai, Jianhua

    2017-05-01

    The time-frequency analysis method represents signal as a function of time and frequency, and it is considered a powerful tool for handling arbitrary non-stationary time series by using instantaneous frequency and instantaneous amplitude. It also provides a possible alternative to the analysis of the non-stationary magnetotelluric (MT) signal. Based on the Hilbert-Huang transform (HHT), a time-frequency analysis method is proposed to obtain stable estimates of the magnetotelluric response function. In contrast to conventional methods, the response function estimation is performed in the time-frequency domain using instantaneous spectra rather than in the frequency domain, which allows for imaging the response parameter content as a function of time and frequency. The theory of the method is presented and the mathematical model and calculation procedure, which are used to estimate response function based on HHT time-frequency spectrum, are discussed. To evaluate the results, response function estimates are compared with estimates from a standard MT data processing method based on the Fourier transform. All results show that apparent resistivities and phases, which are calculated from the HHT time-frequency method, are generally more stable and reliable than those determined from the simple Fourier analysis. The proposed method overcomes the drawbacks of the traditional Fourier methods, and the resulting parameter minimises the estimation bias caused by the non-stationary characteristics of the MT data.

  10. Offline estimation of decay time for an optical cavity with a low pass filter cavity model.

    Science.gov (United States)

    Kallapur, Abhijit G; Boyson, Toby K; Petersen, Ian R; Harb, Charles C

    2012-08-01

    This Letter presents offline estimation results for the decay-time constant for an experimental Fabry-Perot optical cavity for cavity ring-down spectroscopy (CRDS). The cavity dynamics are modeled in terms of a low pass filter (LPF) with unity DC gain. This model is used by an extended Kalman filter (EKF) along with the recorded light intensity at the output of the cavity in order to estimate the decay-time constant. The estimation results using the LPF cavity model are compared to those obtained using the quadrature model for the cavity presented in previous work by Kallapur et al. The estimation process derived using the LPF model comprises two states as opposed to three states in the quadrature model. When considering the EKF, this means propagating two states and a (2×2) covariance matrix using the LPF model, as opposed to propagating three states and a (3×3) covariance matrix using the quadrature model. This gives the former model a computational advantage over the latter and leads to faster execution times for the corresponding EKF. It is shown in this Letter that the LPF model for the cavity with two filter states is computationally more efficient, converges faster, and is hence a more suitable method than the three-state quadrature model presented in previous work for real-time estimation of the decay-time constant for the cavity.

  11. Estimates of the first Dirichlet eigenvalue from exit time moment spectra

    DEFF Research Database (Denmark)

    Hurtado, Ana; Markvorsen, Steen; Palmer, Vicente

    2013-01-01

    We compute the first Dirichlet eigenvalue of a geodesic ball in a rotationally symmetric model space in terms of the moment spectrum for the Brownian motion exit times from the ball. This expression implies an estimate as exact as you want for the first Dirichlet eigenvalue of a geodesic ball...

  12. Characterizing subcritical assemblies with time of flight fixed by energy estimation distributions

    Science.gov (United States)

    Monterial, Mateusz; Marleau, Peter; Pozzi, Sara

    2018-04-01

    We present the Time of Flight Fixed by Energy Estimation (TOFFEE) as a measure of the fission chain dynamics in subcritical assemblies. TOFFEE is the time between correlated gamma rays and neutrons, subtracted by the estimated travel time of the incident neutron from its proton recoil. The measured subcritical assembly was the BeRP ball, a 4.482 kg sphere of α-phase weapons grade plutonium metal, which came in five configurations: bare, 0.5, 1, and 1.5 in iron, and 1 in nickel closed fitting shell reflectors. We extend the measurement with MCNPX-PoliMi simulations of shells ranging up to 6 inches in thickness, and two new reflector materials: aluminum and tungsten. We also simulated the BeRP ball with different masses ranging from 1 to 8 kg. A two-region and single-region point kinetics models were used to model the behavior of the positive side of the TOFFEE distribution from 0 to 100 ns. The single region model of the bare cases gave positive linear correlations between estimated and expected neutron decay constants and leakage multiplications. The two-region model provided a way to estimate neutron multiplication for the reflected cases, which correlated positively with expected multiplication, but the nature of the correlation (sub or superlinear) changed between material types. Finally, we found that the areal density of the reflector shells had a linear correlation with the integral of the two-region model fit. Therefore, we expect that with knowledge of reflector composition, one could determine the shell thickness, or vice versa. Furthermore, up to a certain amount and thickness of the reflector, the two-region model provides a way of distinguishing bare and reflected plutonium assemblies.

  13. A Bayesian framework to estimate diversification rates and their variation through time and space

    Directory of Open Access Journals (Sweden)

    Silvestro Daniele

    2011-10-01

    Full Text Available Abstract Background Patterns of species diversity are the result of speciation and extinction processes, and molecular phylogenetic data can provide valuable information to derive their variability through time and across clades. Bayesian Markov chain Monte Carlo methods offer a promising framework to incorporate phylogenetic uncertainty when estimating rates of diversification. Results We introduce a new approach to estimate diversification rates in a Bayesian framework over a distribution of trees under various constant and variable rate birth-death and pure-birth models, and test it on simulated phylogenies. Furthermore, speciation and extinction rates and their posterior credibility intervals can be estimated while accounting for non-random taxon sampling. The framework is particularly suitable for hypothesis testing using Bayes factors, as we demonstrate analyzing dated phylogenies of Chondrostoma (Cyprinidae and Lupinus (Fabaceae. In addition, we develop a model that extends the rate estimation to a meta-analysis framework in which different data sets are combined in a single analysis to detect general temporal and spatial trends in diversification. Conclusions Our approach provides a flexible framework for the estimation of diversification parameters and hypothesis testing while simultaneously accounting for uncertainties in the divergence times and incomplete taxon sampling.

  14. Time-Varying Transition Probability Matrix Estimation and Its Application to Brand Share Analysis.

    Directory of Open Access Journals (Sweden)

    Tomoaki Chiba

    Full Text Available In a product market or stock market, different products or stocks compete for the same consumers or purchasers. We propose a method to estimate the time-varying transition matrix of the product share using a multivariate time series of the product share. The method is based on the assumption that each of the observed time series of shares is a stationary distribution of the underlying Markov processes characterized by transition probability matrices. We estimate transition probability matrices for every observation under natural assumptions. We demonstrate, on a real-world dataset of the share of automobiles, that the proposed method can find intrinsic transition of shares. The resulting transition matrices reveal interesting phenomena, for example, the change in flows between TOYOTA group and GM group for the fiscal year where TOYOTA group's sales beat GM's sales, which is a reasonable scenario.

  15. Time-Varying Transition Probability Matrix Estimation and Its Application to Brand Share Analysis.

    Science.gov (United States)

    Chiba, Tomoaki; Hino, Hideitsu; Akaho, Shotaro; Murata, Noboru

    2017-01-01

    In a product market or stock market, different products or stocks compete for the same consumers or purchasers. We propose a method to estimate the time-varying transition matrix of the product share using a multivariate time series of the product share. The method is based on the assumption that each of the observed time series of shares is a stationary distribution of the underlying Markov processes characterized by transition probability matrices. We estimate transition probability matrices for every observation under natural assumptions. We demonstrate, on a real-world dataset of the share of automobiles, that the proposed method can find intrinsic transition of shares. The resulting transition matrices reveal interesting phenomena, for example, the change in flows between TOYOTA group and GM group for the fiscal year where TOYOTA group's sales beat GM's sales, which is a reasonable scenario.

  16. Comparisons of Crosswind Velocity Profile Estimates Used in Fast-Time Wake Vortex Prediction Models

    Science.gov (United States)

    Pruis, Mathew J.; Delisi, Donald P.; Ahmad, Nashat N.

    2011-01-01

    Five methods for estimating crosswind profiles used in fast-time wake vortex prediction models are compared in this study. Previous investigations have shown that temporal and spatial variations in the crosswind vertical profile have a large impact on the transport and time evolution of the trailing vortex pair. The most important crosswind parameters are the magnitude of the crosswind and the gradient in the crosswind shear. It is known that pulsed and continuous wave lidar measurements can provide good estimates of the wind profile in the vicinity of airports. In this study comparisons are made between estimates of the crosswind profiles from a priori information on the trajectory of the vortex pair as well as crosswind profiles derived from different sensors and a regional numerical weather prediction model.

  17. Uranium mass and neutron multiplication factor estimates from time-correlation coincidence counts

    Energy Technology Data Exchange (ETDEWEB)

    Xie, Wenxiong [China Academy of Engineering Physics, Center for Strategic Studies, Beijing 100088 (China); Li, Jiansheng [China Academy of Engineering Physics, Institute of Nuclear Physics and Chemistry, Mianyang 621900 (China); Zhu, Jianyu [China Academy of Engineering Physics, Center for Strategic Studies, Beijing 100088 (China)

    2015-10-11

    Time-correlation coincidence counts of neutrons are an important means to measure attributes of nuclear material. The main deficiency in the analysis is that an attribute of an unknown component can only be assessed by comparing it with similar known components. There is a lack of a universal method of measurement suitable for the different attributes of the components. This paper presents a new method that uses universal relations to estimate the mass and neutron multiplication factor of any uranium component with known enrichment. Based on numerical simulations and analyses of 64 highly enriched uranium components with different thicknesses and average radii, the relations between mass, multiplication and coincidence spectral features have been obtained by linear regression analysis. To examine the validity of the method in estimating the mass of uranium components with different sizes, shapes, enrichment, and shielding, the features of time-correlation coincidence-count spectra for other objects with similar attributes are simulated. Most of the masses and multiplications for these objects could also be derived by the formulation. Experimental measurements of highly enriched uranium castings have also been used to verify the formulation. The results show that for a well-designed time-dependent coincidence-count measuring system of a uranium attribute, there are a set of relations dependent on the uranium enrichment by which the mass and multiplication of the measured uranium components of any shape and size can be estimated from the features of the source-detector coincidence-count spectrum.

  18. Simple robust technique using time delay estimation for the control and synchronization of Lorenz systems

    International Nuclear Information System (INIS)

    Jin, Maolin; Chang, Pyung Hun

    2009-01-01

    This work presents two simple and robust techniques based on time delay estimation for the respective control and synchronization of chaos systems. First, one of these techniques is applied to the control of a chaotic Lorenz system with both matched and mismatched uncertainties. The nonlinearities in the Lorenz system is cancelled by time delay estimation and desired error dynamics is inserted. Second, the other technique is applied to the synchronization of the Lue system and the Lorenz system with uncertainties. The synchronization input consists of three elements that have transparent and clear meanings. Since time delay estimation enables a very effective and efficient cancellation of disturbances and nonlinearities, the techniques turn out to be simple and robust. Numerical simulation results show fast, accurate and robust performance of the proposed techniques, thereby demonstrating their effectiveness for the control and synchronization of Lorenz systems.

  19. Quantitative estimation of time-variable earthquake hazard by using fuzzy set theory

    Science.gov (United States)

    Deyi, Feng; Ichikawa, M.

    1989-11-01

    In this paper, the various methods of fuzzy set theory, called fuzzy mathematics, have been applied to the quantitative estimation of the time-variable earthquake hazard. The results obtained consist of the following. (1) Quantitative estimation of the earthquake hazard on the basis of seismicity data. By using some methods of fuzzy mathematics, seismicity patterns before large earthquakes can be studied more clearly and more quantitatively, highly active periods in a given region and quiet periods of seismic activity before large earthquakes can be recognized, similarities in temporal variation of seismic activity and seismic gaps can be examined and, on the other hand, the time-variable earthquake hazard can be assessed directly on the basis of a series of statistical indices of seismicity. Two methods of fuzzy clustering analysis, the method of fuzzy similarity, and the direct method of fuzzy pattern recognition, have been studied is particular. One method of fuzzy clustering analysis is based on fuzzy netting, and another is based on the fuzzy equivalent relation. (2) Quantitative estimation of the earthquake hazard on the basis of observational data for different precursors. The direct method of fuzzy pattern recognition has been applied to research on earthquake precursors of different kinds. On the basis of the temporal and spatial characteristics of recognized precursors, earthquake hazards in different terms can be estimated. This paper mainly deals with medium-short-term precursors observed in Japan and China.

  20. A Sandwich-Type Standard Error Estimator of SEM Models with Multivariate Time Series

    Science.gov (United States)

    Zhang, Guangjian; Chow, Sy-Miin; Ong, Anthony D.

    2011-01-01

    Structural equation models are increasingly used as a modeling tool for multivariate time series data in the social and behavioral sciences. Standard error estimators of SEM models, originally developed for independent data, require modifications to accommodate the fact that time series data are inherently dependent. In this article, we extend a…

  1. A coherent method for the detection and parameter estimation of continuous gravitational wave signals using a pulsar timing array

    International Nuclear Information System (INIS)

    Wang, Yan; Mohanty, Soumya D.; Jenet, Fredrick A.

    2014-01-01

    The use of a high precision pulsar timing array is a promising approach to detecting gravitational waves in the very low frequency regime (10 –6 -10 –9 Hz) that is complementary to ground-based efforts (e.g., LIGO, Virgo) at high frequencies (∼10-10 3 Hz) and space-based ones (e.g., LISA) at low frequencies (10 –4 -10 –1 Hz). One of the target sources for pulsar timing arrays is individual supermassive black hole binaries which are expected to form in galactic mergers. In this paper, a likelihood-based method for detection and parameter estimation is presented for a monochromatic continuous gravitational wave signal emitted by such a source. The so-called pulsar terms in the signal that arise due to the breakdown of the long-wavelength approximation are explicitly taken into account in this method. In addition, the method accounts for equality and inequality constraints involved in the semi-analytical maximization of the likelihood over a subset of the parameters. The remaining parameters are maximized over numerically using Particle Swarm Optimization. Thus, the method presented here solves the monochromatic continuous wave detection and parameter estimation problem without invoking some of the approximations that have been used in earlier studies.

  2. A coherent method for the detection and parameter estimation of continuous gravitational wave signals using a pulsar timing array

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Yan; Mohanty, Soumya D.; Jenet, Fredrick A. [Department of Physics and Astronomy, University of Texas at Brownsville, 1 West University Boulevard, Brownsville, TX 78520 (United States)

    2014-11-01

    The use of a high precision pulsar timing array is a promising approach to detecting gravitational waves in the very low frequency regime (10{sup –6}-10{sup –9} Hz) that is complementary to ground-based efforts (e.g., LIGO, Virgo) at high frequencies (∼10-10{sup 3} Hz) and space-based ones (e.g., LISA) at low frequencies (10{sup –4}-10{sup –1} Hz). One of the target sources for pulsar timing arrays is individual supermassive black hole binaries which are expected to form in galactic mergers. In this paper, a likelihood-based method for detection and parameter estimation is presented for a monochromatic continuous gravitational wave signal emitted by such a source. The so-called pulsar terms in the signal that arise due to the breakdown of the long-wavelength approximation are explicitly taken into account in this method. In addition, the method accounts for equality and inequality constraints involved in the semi-analytical maximization of the likelihood over a subset of the parameters. The remaining parameters are maximized over numerically using Particle Swarm Optimization. Thus, the method presented here solves the monochromatic continuous wave detection and parameter estimation problem without invoking some of the approximations that have been used in earlier studies.

  3. Estimating the implicit component of visuomotor rotation learning by constraining movement preparation time.

    Science.gov (United States)

    Leow, Li-Ann; Gunn, Reece; Marinovic, Welber; Carroll, Timothy J

    2017-08-01

    When sensory feedback is perturbed, accurate movement is restored by a combination of implicit processes and deliberate reaiming to strategically compensate for errors. Here, we directly compare two methods used previously to dissociate implicit from explicit learning on a trial-by-trial basis: 1 ) asking participants to report the direction that they aim their movements, and contrasting this with the directions of the target and the movement that they actually produce, and 2 ) manipulating movement preparation time. By instructing participants to reaim without a sensory perturbation, we show that reaiming is possible even with the shortest possible preparation times, particularly when targets are narrowly distributed. Nonetheless, reaiming is effortful and comes at the cost of increased variability, so we tested whether constraining preparation time is sufficient to suppress strategic reaiming during adaptation to visuomotor rotation with a broad target distribution. The rate and extent of error reduction under preparation time constraints were similar to estimates of implicit learning obtained from self-report without time pressure, suggesting that participants chose not to apply a reaiming strategy to correct visual errors under time pressure. Surprisingly, participants who reported aiming directions showed less implicit learning according to an alternative measure, obtained during trials performed without visual feedback. This suggests that the process of reporting can affect the extent or persistence of implicit learning. The data extend existing evidence that restricting preparation time can suppress explicit reaiming and provide an estimate of implicit visuomotor rotation learning that does not require participants to report their aiming directions. NEW & NOTEWORTHY During sensorimotor adaptation, implicit error-driven learning can be isolated from explicit strategy-driven reaiming by subtracting self-reported aiming directions from movement directions, or

  4. Estimated incident cost savings in shipping due to inspections.

    Science.gov (United States)

    Knapp, Sabine; Bijwaard, Govert; Heij, Christiaan

    2011-07-01

    The effectiveness of safety inspections of ships has been analysed from various angles, but until now, relatively little attention has been given to translate risk reduction into incident cost savings. This paper provides a monetary quantification of the cost savings that can be attributed to port state control inspections and industry vetting inspections. The dataset consists of more than half a million ship arrivals between 2002 and 2007 and contains inspections of port state authorities in the USA and Australia and of three industry vetting regimes. The effect of inspections in reducing the risk of total loss accidents is estimated by means of duration models, in terms of the gained probability of survival. The monetary benefit of port state control inspections is estimated to range, on average, from about 70 to 190 thousand dollars, with median values ranging from about 20 to 45 thousand dollars. Industry inspections have even higher benefits, especially for tankers. The savings are in general higher for older and larger vessels, and also for vessels with undefined flag and unknown classification society. As inspection costs are relatively low in comparison to potential cost savings, the results underline the importance of determining ships with relatively high risk of total loss. Copyright © 2011 Elsevier Ltd. All rights reserved.

  5. Estimation and Application of Ecological Memory Functions in Time and Space

    Science.gov (United States)

    Itter, M.; Finley, A. O.; Dawson, A.

    2017-12-01

    A common goal in quantitative ecology is the estimation or prediction of ecological processes as a function of explanatory variables (or covariates). Frequently, the ecological process of interest and associated covariates vary in time, space, or both. Theory indicates many ecological processes exhibit memory to local, past conditions. Despite such theoretical understanding, few methods exist to integrate observations from the recent past or within a local neighborhood as drivers of these processes. We build upon recent methodological advances in ecology and spatial statistics to develop a Bayesian hierarchical framework to estimate so-called ecological memory functions; that is, weight-generating functions that specify the relative importance of local, past covariate observations to ecological processes. Memory functions are estimated using a set of basis functions in time and/or space, allowing for flexible ecological memory based on a reduced set of parameters. Ecological memory functions are entirely data driven under the Bayesian hierarchical framework—no a priori assumptions are made regarding functional forms. Memory function uncertainty follows directly from posterior distributions for model parameters allowing for tractable propagation of error to predictions of ecological processes. We apply the model framework to simulated spatio-temporal datasets generated using memory functions of varying complexity. The framework is also applied to estimate the ecological memory of annual boreal forest growth to local, past water availability. Consistent with ecological understanding of boreal forest growth dynamics, memory to past water availability peaks in the year previous to growth and slowly decays to zero in five to eight years. The Bayesian hierarchical framework has applicability to a broad range of ecosystems and processes allowing for increased understanding of ecosystem responses to local and past conditions and improved prediction of ecological

  6. Testing and estimating time-varying elasticities of Swiss gasoline demand

    International Nuclear Information System (INIS)

    Neto, David

    2012-01-01

    This paper is intended to test and estimate time-varying elasticities for gasoline demand in Switzerland. For this purpose, a smooth time-varying cointegrating parameters model is investigated in order to describe smooth mutations of the Swiss gasoline demand. The methodology, based on Chebyshev polynomials, is rigorously outlined. Our empirical finding states that the time-invariance assumption does not hold for long-run price and income elasticities. Furthermore they highlight that gasoline demand passed through some periods of sensitivity and non sensitivity with respect to the price. Our empirical statements are of great importance to assess the performance of a gasoline tax as an instrument for CO 2 reduction policy. Indeed, such an instrument can contribute to reduce emissions of greenhouse gases only if the demand is not fully inelastic with respect to the price. Our results suggest that such a carbon-tax would not be always suitable since the price elasticity is found not stable over time and not always significant.

  7. Distributed Extended Kalman Filter for Position, Velocity, Time, Estimation in Satellite Navigation Receivers

    Directory of Open Access Journals (Sweden)

    O. Jakubov

    2013-09-01

    Full Text Available Common techniques for position-velocity-time estimation in satellite navigation, iterative least squares and the extended Kalman filter, involve matrix operations. The matrix inversion and inclusion of a matrix library pose requirements on a computational power and operating platform of the navigation processor. In this paper, we introduce a novel distributed algorithm suitable for implementation in simple parallel processing units each for a tracked satellite. Such a unit performs only scalar sum, subtraction, multiplication, and division. The algorithm can be efficiently implemented in hardware logic. Given the fast position-velocity-time estimator, frequent estimates can foster dynamic performance of a vector tracking receiver. The algorithm has been designed from a factor graph representing the extended Kalman filter by splitting vector nodes into scalar ones resulting in a cyclic graph with few iterations needed. Monte Carlo simulations have been conducted to investigate convergence and accuracy. Simulation case studies for a vector tracking architecture and experimental measurements with a real-time software receiver developed at CTU in Prague were conducted. The algorithm offers compromises in stability, accuracy, and complexity depending on the number of iterations. In scenarios with a large number of tracked satellites, it can outperform the traditional methods at low complexity.

  8. Towards real-time diffuse optical tomography for imaging brain functions cooperated with Kalman estimator

    Science.gov (United States)

    Wang, Bingyuan; Zhang, Yao; Liu, Dongyuan; Ding, Xuemei; Dan, Mai; Pan, Tiantian; Wang, Yihan; Li, Jiao; Zhou, Zhongxing; Zhang, Limin; Zhao, Huijuan; Gao, Feng

    2018-02-01

    Functional near-infrared spectroscopy (fNIRS) is a non-invasive neuroimaging method to monitor the cerebral hemodynamic through the optical changes measured at the scalp surface. It has played a more and more important role in psychology and medical imaging communities. Real-time imaging of brain function using NIRS makes it possible to explore some sophisticated human brain functions unexplored before. Kalman estimator has been frequently used in combination with modified Beer-Lamber Law (MBLL) based optical topology (OT), for real-time brain function imaging. However, the spatial resolution of the OT is low, hampering the application of OT in exploring some complicated brain functions. In this paper, we develop a real-time imaging method combining diffuse optical tomography (DOT) and Kalman estimator, much improving the spatial resolution. Instead of only presenting one spatially distributed image indicating the changes of the absorption coefficients at each time point during the recording process, one real-time updated image using the Kalman estimator is provided. Its each voxel represents the amplitude of the hemodynamic response function (HRF) associated with this voxel. We evaluate this method using some simulation experiments, demonstrating that this method can obtain more reliable spatial resolution images. Furthermore, a statistical analysis is also conducted to help to decide whether a voxel in the field of view is activated or not.

  9. Real-Time Personalized Monitoring to Estimate Occupational Heat Stress in Ambient Assisted Working

    Directory of Open Access Journals (Sweden)

    Pablo Pancardo

    2015-07-01

    Full Text Available Ambient Assisted Working (AAW is a discipline aiming to provide comfort and safety in the workplace through customization and technology. Workers’ comfort may be compromised in many labor situations, including those depending on environmental conditions, like extremely hot weather conduces to heat stress. Occupational heat stress (OHS happens when a worker is in an uninterrupted physical activity and in a hot environment. OHS can produce strain on the body, which leads to discomfort and eventually to heat illness and even death. Related ISO standards contain methods to estimate OHS and to ensure the safety and health of workers, but they are subjective, impersonal, performed a posteriori and even invasive. This paper focuses on the design and development of real-time personalized monitoring for a more effective and objective estimation of OHS, taking into account the individual user profile, fusing data from environmental and unobtrusive body sensors. Formulas employed in this work were taken from different domains and joined in the method that we propose. It is based on calculations that enable continuous surveillance of physical activity performance in a comfortable and healthy manner. In this proposal, we found that OHS can be estimated by satisfying the following criteria: objective, personalized, in situ, in real time, just in time and in an unobtrusive way. This enables timely notice for workers to make decisions based on objective information to control OHS.

  10. SU-E-J-164: Estimation of DVH Variation for PTV Due to Interfraction Organ Motion in Prostate VMAT Using Gaussian Error Function

    International Nuclear Information System (INIS)

    Lewis, C; Jiang, R; Chow, J

    2015-01-01

    Purpose: We developed a method to predict the change of DVH for PTV due to interfraction organ motion in prostate VMAT without repeating the CT scan and treatment planning. The method is based on a pre-calculated patient database with DVH curves of PTV modelled by the Gaussian error function (GEF). Methods: For a group of 30 patients with different prostate sizes, their VMAT plans were recalculated by shifting their PTVs 1 cm with 10 increments in the anterior-posterior, left-right and superior-inferior directions. The DVH curve of PTV in each replan was then fitted by the GEF to determine parameters describing the shape of curve. Information of parameters, varying with the DVH change due to prostate motion for different prostate sizes, was analyzed and stored in a database of a program written by MATLAB. Results: To predict a new DVH for PTV due to prostate interfraction motion, prostate size and shift distance with direction were input to the program. Parameters modelling the DVH for PTV were determined based on the pre-calculated patient dataset. From the new parameters, DVH curves of PTVs with and without considering the prostate motion were plotted for comparison. The program was verified with different prostate cases involving interfraction prostate shifts and replans. Conclusion: Variation of DVH for PTV in prostate VMAT can be predicted using a pre-calculated patient database with DVH curve fitting. The computing time is fast because CT rescan and replan are not required. This quick DVH estimation can help radiation staff to determine if the changed PTV coverage due to prostate shift is tolerable in the treatment. However, it should be noted that the program can only consider prostate interfraction motions along three axes, and is restricted to prostate VMAT plan using the same plan script in the treatment planning system

  11. The Past Sure is Tense: On Interpreting Phylogenetic Divergence Time Estimates.

    Science.gov (United States)

    Brown, Joseph W; Smith, Stephen A

    2018-03-01

    Divergence time estimation-the calibration of a phylogeny to geological time-is an integral first step in modeling the tempo of biological evolution (traits and lineages). However, despite increasingly sophisticated methods to infer divergence times from molecular genetic sequences, the estimated age of many nodes across the tree of life contrast significantly and consistently with timeframes conveyed by the fossil record. This is perhaps best exemplified by crown angiosperms, where molecular clock (Triassic) estimates predate the oldest (Early Cretaceous) undisputed angiosperm fossils by tens of millions of years or more. While the incompleteness of the fossil record is a common concern, issues of data limitation and model inadequacy are viable (if underexplored) alternative explanations. In this vein, Beaulieu et al. (2015) convincingly demonstrated how methods of divergence time inference can be misled by both (i) extreme state-dependent molecular substitution rate heterogeneity and (ii) biased sampling of representative major lineages. These results demonstrate the impact of (potentially common) model violations. Here, we suggest another potential challenge: that the configuration of the statistical inference problem (i.e., the parameters, their relationships, and associated priors) alone may preclude the reconstruction of the paleontological timeframe for the crown age of angiosperms. We demonstrate, through sampling from the joint prior (formed by combining the tree (diversification) prior with the calibration densities specified for fossil-calibrated nodes) that with no data present at all, that an Early Cretaceous crown angiosperms is rejected (i.e., has essentially zero probability). More worrisome, however, is that for the 24 nodes calibrated by fossils, almost all have indistinguishable marginal prior and posterior age distributions when employing routine lognormal fossil calibration priors. These results indicate that there is inadequate information in

  12. Nonparametric trend estimation in the presence of fractal noise: application to fMRI time-series analysis.

    Science.gov (United States)

    Afshinpour, Babak; Hossein-Zadeh, Gholam-Ali; Soltanian-Zadeh, Hamid

    2008-06-30

    Unknown low frequency fluctuations called "trend" are observed in noisy time-series measured for different applications. In some disciplines, they carry primary information while in other fields such as functional magnetic resonance imaging (fMRI) they carry nuisance effects. In all cases, however, it is necessary to estimate them accurately. In this paper, a method for estimating trend in the presence of fractal noise is proposed and applied to fMRI time-series. To this end, a partly linear model (PLM) is fitted to each time-series. The parametric and nonparametric parts of PLM are considered as contributions of hemodynamic response and trend, respectively. Using the whitening property of wavelet transform, the unknown components of the model are estimated in the wavelet domain. The results of the proposed method are compared to those of other parametric trend-removal approaches such as spline and polynomial models. It is shown that the proposed method improves activation detection and decreases variance of the estimated parameters relative to the other methods.

  13. An Adaptive Channel Estimation Algorithm Using Time-Frequency Polynomial Model for OFDM with Fading Multipath Channels

    Directory of Open Access Journals (Sweden)

    Liu KJ Ray

    2002-01-01

    Full Text Available Orthogonal frequency division multiplexing (OFDM is an effective technique for the future 3G communications because of its great immunity to impulse noise and intersymbol interference. The channel estimation is a crucial aspect in the design of OFDM systems. In this work, we propose a channel estimation algorithm based on a time-frequency polynomial model of the fading multipath channels. The algorithm exploits the correlation of the channel responses in both time and frequency domains and hence reduce more noise than the methods using only time or frequency polynomial model. The estimator is also more robust compared to the existing methods based on Fourier transform. The simulation shows that it has more than improvement in terms of mean-squared estimation error under some practical channel conditions. The algorithm needs little prior knowledge about the delay and fading properties of the channel. The algorithm can be implemented recursively and can adjust itself to follow the variation of the channel statistics.

  14. Calculation of the time resolution of the J-PET tomograph using kernel density estimation

    Science.gov (United States)

    Raczyński, L.; Wiślicki, W.; Krzemień, W.; Kowalski, P.; Alfs, D.; Bednarski, T.; Białas, P.; Curceanu, C.; Czerwiński, E.; Dulski, K.; Gajos, A.; Głowacz, B.; Gorgol, M.; Hiesmayr, B.; Jasińska, B.; Kamińska, D.; Korcyl, G.; Kozik, T.; Krawczyk, N.; Kubicz, E.; Mohammed, M.; Pawlik-Niedźwiecka, M.; Niedźwiecki, S.; Pałka, M.; Rudy, Z.; Rundel, O.; Sharma, N. G.; Silarski, M.; Smyrski, J.; Strzelecki, A.; Wieczorek, A.; Zgardzińska, B.; Zieliński, M.; Moskal, P.

    2017-06-01

    In this paper we estimate the time resolution of the J-PET scanner built from plastic scintillators. We incorporate the method of signal processing using the Tikhonov regularization framework and the kernel density estimation method. We obtain simple, closed-form analytical formulae for time resolution. The proposed method is validated using signals registered by means of the single detection unit of the J-PET tomograph built from a 30 cm long plastic scintillator strip. It is shown that the experimental and theoretical results obtained for the J-PET scanner equipped with vacuum tube photomultipliers are consistent.

  15. Synaptic inhibition and excitation estimated via the time constant of membrane potential fluctuations

    DEFF Research Database (Denmark)

    Berg, Rune W.; Ditlevsen, Susanne

    2013-01-01

    When recording the membrane potential, V, of a neuron it is desirable to be able to extract the synaptic input. Critically, the synaptic input is stochastic and non-reproducible so one is therefore often restricted to single trial data. Here, we introduce means of estimating the inhibition and ex...... close to soma (recording site). Though our data is in current-clamp, the method also works in V-clamp recordings, with some minor adaptations. All custom made procedures are provided in Matlab....... and excitation and their confidence limits from single sweep trials. The estimates are based on the mean membrane potential, (V) , and the membrane time constant,τ. The time constant provides the total conductance (G = capacitance/τ) and is extracted from the autocorrelation of V. The synaptic conductances can....... The method gives best results if the synaptic input is large compared to other conductances, the intrinsic conductances have little or no time dependence or are comparably small, the ligand gated kinetics is faster than the membrane time constant, and the majority of synaptic contacts are electrotonically...

  16. Penalized Nonlinear Least Squares Estimation of Time-Varying Parameters in Ordinary Differential Equations

    KAUST Repository

    Cao, Jiguo; Huang, Jianhua Z.; Wu, Hulin

    2012-01-01

    Ordinary differential equations (ODEs) are widely used in biomedical research and other scientific areas to model complex dynamic systems. It is an important statistical problem to estimate parameters in ODEs from noisy observations. In this article we propose a method for estimating the time-varying coefficients in an ODE. Our method is a variation of the nonlinear least squares where penalized splines are used to model the functional parameters and the ODE solutions are approximated also using splines. We resort to the implicit function theorem to deal with the nonlinear least squares objective function that is only defined implicitly. The proposed penalized nonlinear least squares method is applied to estimate a HIV dynamic model from a real dataset. Monte Carlo simulations show that the new method can provide much more accurate estimates of functional parameters than the existing two-step local polynomial method which relies on estimation of the derivatives of the state function. Supplemental materials for the article are available online.

  17. Methodological issues in the estimation of parental time – Analysis of measures in a Canadian time-use survey

    OpenAIRE

    Cara B. Fedick; Shelley Pacholok; Anne H. Gauthier

    2005-01-01

    Extensive small scale studies have documented that when people assume the role of assisting a person with impairments or an older person, care activities account for a significant portion of their daily routines. Nevertheless, little research has investigated the problem of measuring the time that carers spend in care-related activities. This paper contrasts two different measures of care time – an estimated average weekly hours question in the 1998 Australian Survey of Disability, Ageing and...

  18. Imprecision in estimates of dose from ingested 137Cs due to variability in human biological characteristics

    International Nuclear Information System (INIS)

    Schwarz, G.; Dunning, D.E. Jr.

    1982-01-01

    An attempt has been made to quantify the variability in human biological parameters determining dose to man from ingestion of a unit activity of soluble 137 Cs and the resulting imprecision in the predicted total-body dose commitment. The analysis is based on an extensive review of the literature along with the application of statistical methods to determine parameter variability, correlations between parameters, and predictive imprecision. The variability in the principal biological parameters (biological half-time and total-body mass) involved can be described by a geometric standard deviation of 1.2-1.5 for adults and 1.6-1.9 for children/ adolescents of age 0.1-18 yr. The estimated predictive imprecision (using a Monte Carlo technique) in the total-body dose commitment from ingested 137 Cs can be described by a geometric standard deviation on the order of 1.3-1.4, meaning that the 99th percentile of the predicted distribution of dose is within approximately 2.1 times the mean value. The mean dose estimate is 0.009 Sv/MBq (34 mrem/μ Ci) for children/adolescents and 0.01 Sv/MBq (38 mrem/μ Ci) for adults. Little evidence of age dependence in the total-body dose from ingested 137 Cs is observed. (author)

  19. Estimation of Airborne Lidar-Derived Tropical Forest Canopy Height Using Landsat Time Series in Cambodia

    Directory of Open Access Journals (Sweden)

    Tetsuji Ota

    2014-11-01

    Full Text Available In this study, we test and demonstrate the utility of disturbance and recovery information derived from annual Landsat time series to predict current forest vertical structure (as compared to the more common approaches, that consider a sample of airborne Lidar and single-date Landsat derived variables. Mean Canopy Height (MCH was estimated separately using single date, time series, and the combination of single date and time series variables in multiple regression and random forest (RF models. The combination of single date and time series variables, which integrate disturbance history over the entire time series, overall provided better MCH prediction than using either of the two sets of variables separately. In general, the RF models resulted in improved performance in all estimates over those using multiple regression. The lowest validation error was obtained using Landsat time series variables in a RF model (R2 = 0.75 and RMSE = 2.81 m. Combining single date and time series data was more effective when the RF model was used (opposed to multiple regression. The RMSE for RF mean canopy height prediction was reduced by 13.5% when combining the two sets of variables as compared to the 3.6% RMSE decline presented by multiple regression. This study demonstrates the value of airborne Lidar and long term Landsat observations to generate estimates of forest canopy height using the random forest algorithm.

  20. An Estimation of Operator's Diagnostic Time for Feed-And-Bleed Operation under Various Scenarios

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Bo Gyung; Kang, Hyun Gook [KAIST, Daejeon (Korea, Republic of); Yoon, Ho Joon [Khalifa University of Science, Abu Dhabi (United Arab Emirates)

    2015-10-15

    In order to assess realistic safety of plant, effects of interactions between components, operator, and plant condition are needed to be considered in the PSA model. One of the important issues to estimate the CDF is the estimation of human error probability (HEP). When an accident occurs, operators follow the emergency operating procedure and check various alarm, parameters, and signals. In the conventional Korean PSA model, the Korean standard HRA (K-HRA) method is used. In this method, the HEP is the sum of diagnosis error probability and execution error probability. A diagnosis error probability is expressed by the available time for diagnosis and adjusting performance shaping factors, and an execution error probability is a function of task type and stress level. Available time for diagnosis is very important factor of HEP. If the available time for diagnosis is short, the HEP becomes high. In order to obtain the realistic risk assessment results, we first focus on the estimation of HEP considering the plant dynamics under various scenarios. Target operation and scenarios are feed-and bleed operation (F and B operation) and total loss of feedwater (TLOFW) accident with/without loss of coolant accident (LOCA). One of the highest HEP is HEP of FnB operation. In additional, Scenarios, which are related to combination secondary heat removal failure and primary heat removal failure, are most critical core damage scenario of the combined accident except scenarios related to station black out (SBO). In these scenarios, the FnB operation is last resort to prevent core damage. To estimate available operator diagnosis time, we identify the relationship between accidents, mitigation function, and plant condition. Distribution of available time of diagnosis was estimated using the MOSAIQUE. The variables are break size, break timing, trip timing of RCP, and availability of high pressure safety injection (HPSI) pump under the TLOFW accident with LOCA. For Type 1 accident

  1. An estimation of crude oil import demand in Turkey: Evidence from time-varying parameters approach

    International Nuclear Information System (INIS)

    Ozturk, Ilhan; Arisoy, Ibrahim

    2016-01-01

    The aim of this study is to model crude oil import demand and estimate the price and income elasticities of imported crude oil in Turkey based on a time-varying parameters (TVP) approach with the aim of obtaining accurate and more robust estimates of price and income elasticities. This study employs annual time series data of domestic oil consumption, real GDP, and oil price for the period 1966–2012. The empirical results indicate that both the income and price elasticities are in line with the theoretical expectations. However, the income elasticity is statistically significant while the price elasticity is statistically insignificant. The relatively high value of income elasticity (1.182) from this study suggests that crude oil import in Turkey is more responsive to changes in income level. This result indicates that imported crude oil is a normal good and rising income levels will foster higher consumption of oil based equipments, vehicles and services by economic agents. The estimated income elasticity of 1.182 suggests that imported crude oil consumption grows at a higher rate than income. This in turn reduces oil intensity over time. Therefore, crude oil import during the estimation period is substantially driven by income. - Highlights: • We estimated the price and income elasticities of imported crude oil in Turkey. • Income elasticity is statistically significant and it is 1.182. • The price elasticity is statistically insignificant. • Crude oil import in Turkey is more responsive to changes in income level. • Crude oil import during the estimation period is substantially driven by income.

  2. A continuous-time adaptive particle filter for estimations under measurement time uncertainties with an application to a plasma-leucine mixed effects model.

    Science.gov (United States)

    Krengel, Annette; Hauth, Jan; Taskinen, Marja-Riitta; Adiels, Martin; Jirstrand, Mats

    2013-01-19

    When mathematical modelling is applied to many different application areas, a common task is the estimation of states and parameters based on measurements. With this kind of inference making, uncertainties in the time when the measurements have been taken are often neglected, but especially in applications taken from the life sciences, this kind of errors can considerably influence the estimation results. As an example in the context of personalized medicine, the model-based assessment of the effectiveness of drugs is becoming to play an important role. Systems biology may help here by providing good pharmacokinetic and pharmacodynamic (PK/PD) models. Inference on these systems based on data gained from clinical studies with several patient groups becomes a major challenge. Particle filters are a promising approach to tackle these difficulties but are by itself not ready to handle uncertainties in measurement times. In this article, we describe a variant of the standard particle filter (PF) algorithm which allows state and parameter estimation with the inclusion of measurement time uncertainties (MTU). The modified particle filter, which we call MTU-PF, also allows the application of an adaptive stepsize choice in the time-continuous case to avoid degeneracy problems. The modification is based on the model assumption of uncertain measurement times. While the assumption of randomness in the measurements themselves is common, the corresponding measurement times are generally taken as deterministic and exactly known. Especially in cases where the data are gained from measurements on blood or tissue samples, a relatively high uncertainty in the true measurement time seems to be a natural assumption. Our method is appropriate in cases where relatively few data are used from a relatively large number of groups or individuals, which introduce mixed effects in the model. This is a typical setting of clinical studies. We demonstrate the method on a small artificial example

  3. Basic Investigations of Dynamic Travel Time Estimation Model for Traffic Signals Control Using Information from Optical Beacons

    Science.gov (United States)

    Okutani, Iwao; Mitsui, Tatsuro; Nakada, Yusuke

    In this paper put forward are neuron-type models, i.e., neural network model, wavelet neuron model and three layered wavelet neuron model(WV3), for estimating traveling time between signalized intersections in order to facilitate adaptive setting of traffic signal parameters such as green time and offset. Model validation tests using simulated data reveal that compared to other models, WV3 model works very fast in learning process and can produce more accurate estimates of travel time. Also, it is exhibited that up-link information obtainable from optical beacons, i.e., travel time observed during the former cycle time in this case, makes a crucial input variable to the models in that there isn't any substantial difference between the change of estimated and simulated travel time with the change of green time or offset when up-link information is employed as input while there appears big discrepancy between them when not employed.

  4. Estimation of Concrete Corrosion Due to Attack of Chloride Salt

    Directory of Open Access Journals (Sweden)

    V. V. Babitski

    2005-01-01

    Full Text Available The paper provides results of experimental concrete research under conditions of concentrated chloride salt solutions. General principles of forecasting concrete corrosion resistance under salt physical corrosion are given in the paper. Analytical dependences for quantitative estimation of corroded concrete have been obtained.

  5. An estimate by two methods of thyroid absorbed doses due to BRAVO fallout in several Northern Marshall Islands.

    Science.gov (United States)

    Musolino, S V; Greenhouse, N A; Hull, A P

    1997-10-01

    Estimates of the thyroid absorbed doses due to fallout originating from the 1 March 1954 BRAVO thermonuclear test on Bikini Atoll have been made for several inhabited locations in the Northern Marshall Islands. Rongelap, Utirik, Rongerik and Ailinginae Atolls were also inhabited on 1 March 1954, where retrospective thyroid absorbed doses have previously been reconstructed. The current estimates are based primarily on external exposure data, which were recorded shortly after each nuclear test in the Castle Series, and secondarily on soil concentrations of 137Cs in samples collected in 1978 and 1988, along with aerial monitoring done in 1978. The external exposures and 137Cs soil concentrations were representative of the atmospheric transport and deposition patterns of the entire Castle Series tests and show that the BRAVO test was the major contributor to fallout exposure during the Castle series and other test series which were carried out in the Marshall Islands. These data have been used as surrogates for fission product radioiodines and telluriums in order to estimate the range of thyroid absorbed doses that may have occurred throughout the Marshall Islands. Dosimetry based on these two sets of estimates agreed within a factor of 4 at the locations where BRAVO was the dominant contributor to the total exposure and deposition. Both methods indicate that thyroid absorbed doses in the range of 1 Gy (100 rad) may have been incurred in some of the northern locations, whereas the doses at southern locations did not significantly exceed levels comparable to those from worldwide fallout. The results of these estimates indicate that a systematic medical survey for thyroid disease should be conducted, and that a more definitive dose reconstruction should be made for all the populated atolls and islands in the Northern Marshall Islands beyond Rongelap, Utirik, Rongerik and Ailinginae, which were significantly contaminated by BRAVO fallout.

  6. An estimate by two methods of thyroid absorbed doses due to BRAVO fallout in several northern Marshall Islands

    International Nuclear Information System (INIS)

    Musolino, S.V.; Hull, A.P.; Greenhouse, N.A.

    1997-01-01

    Estimates of the thyroid absorbed doses due to fallout originating from the 1 March 1954 BRAVO thermonuclear test on Bikini Atoll have been made for several inhabited locations in the Northern Marshall Islands. Rongelap, Utirik, Rongerik and Ailinginae Atolls were also inhabited on 1 March 1954, where retrospective thyroid absorbed doses have previously been reconstructed. Current estimates are based primarily on external exposure data, which were recorded shortly after each nuclear test in the Castle Series, and secondarily on soil concentrations of 137 Cs in samples collected in 1978 and 1988, along with aerial monitoring done in 1978. External exposures and 137 Cs Soil concentrations were representative of the atmospheric transport and deposition patterns of the entire Castle Series tests and show that the BRAVO test was the major contributor to fallout exposure during the Castle series and other test series which were carried out in the Marshall Islands. These data have been used as surrogates for fission product radioiodines and telluriums in order to estimate the range of thyroid absorbed doses that may have occurred throughout the Marshall Islands. Dosimetry based on these two sets of estimates agreed within a factor of 4 at the locations where BRAVO was the dominant contributor to the total exposure and deposition. Both methods indicate that thyroid absorbed doses in the range of 1 Gy (100 rad) may have been incurred in some of the northern locations, whereas the doses at southern locations did not significantly exceed levels comparable to those from worldwide fallout. The results of these estimates indicate that a systematic medical survey for thyroid disease should be conducted, and that a more definitive dose reconstruction should be made for all the populated atolls and islands in the Northern Marshall Islands beyond Rongelap, Utirik, Rongerik and Ailinginae, which were significantly contaminated by BRAVO fallout. 30 refs., 2 figs., 10 tabs

  7. [Winter wheat area estimation with MODIS-NDVI time series based on parcel].

    Science.gov (United States)

    Li, Le; Zhang, Jin-shui; Zhu, Wen-quan; Hu, Tan-gao; Hou, Dong

    2011-05-01

    Several attributes of MODIS (moderate resolution imaging spectrometer) data, especially the short temporal intervals and the global coverage, provide an extremely efficient way to map cropland and monitor its seasonal change. However, the reliability of their measurement results is challenged because of the limited spatial resolution. The parcel data has clear geo-location and obvious boundary information of cropland. Also, the spectral differences and the complexity of mixed pixels are weak in parcels. All of these make that area estimation based on parcels presents more advantage than on pixels. In the present study, winter wheat area estimation based on MODIS-NDVI time series has been performed with the support of cultivated land parcel in Tongzhou, Beijing. In order to extract the regional winter wheat acreage, multiple regression methods were used to simulate the stable regression relationship between MODIS-NDVI time series data and TM samples in parcels. Through this way, the consistency of the extraction results from MODIS and TM can stably reach up to 96% when the amount of samples accounts for 15% of the whole area. The results shows that the use of parcel data can effectively improve the error in recognition results in MODIS-NDVI based multi-series data caused by the low spatial resolution. Therefore, with combination of moderate and low resolution data, the winter wheat area estimation became available in large-scale region which lacks completed medium resolution images or has images covered with clouds. Meanwhile, it carried out the preliminary experiments for other crop area estimation.

  8. A procedure for estimation of pipe break probabilities due to IGSCC

    International Nuclear Information System (INIS)

    Bergman, M.; Brickstad, B.; Nilsson, F.

    1998-06-01

    A procedure has been developed for estimation of the failure probability of welds joints in nuclear piping susceptible to intergranular stress corrosion cracking. The procedure aims at a robust and rapid estimate of the failure probability for a specific weld with known stress state. Random properties are taken into account of the crack initiation rate, the initial crack length, the in-service inspection efficiency and the leak rate. A computer realization of the procedure has been developed for user friendly applications by design engineers. Some examples are considered to investigate the sensitivity of the failure probability to different input quantities. (au)

  9. Experimental study of geo-acoustic inversion uncertainty due to ocean sound-speed fluctuations.

    NARCIS (Netherlands)

    Siderius, M.; Nielsen, P.L.; Sellschopp, J.; Snellen, M.; Simons, D.G.

    2001-01-01

    Acoustic data measured in the ocean fluctuate due to the complex time-varying properties of the channel. When measured data are used for model-based, geo-acoustic inversion, how do acoustic fluctuations impact estimates for the seabed properties? In May 1999 SACLANT Undersea Research Center and

  10. Estimation of time- and state-dependent delays and other parameters in functional differential equations

    Science.gov (United States)

    Murphy, K. A.

    1990-01-01

    A parameter estimation algorithm is developed which can be used to estimate unknown time- or state-dependent delays and other parameters (e.g., initial condition) appearing within a nonlinear nonautonomous functional differential equation. The original infinite dimensional differential equation is approximated using linear splines, which are allowed to move with the variable delay. The variable delays are approximated using linear splines as well. The approximation scheme produces a system of ordinary differential equations with nice computational properties. The unknown parameters are estimated within the approximating systems by minimizing a least-squares fit-to-data criterion. Convergence theorems are proved for time-dependent delays and state-dependent delays within two classes, which say essentially that fitting the data by using approximations will, in the limit, provide a fit to the data using the original system. Numerical test examples are presented which illustrate the method for all types of delay.

  11. Maximum Likelihood Blind Channel Estimation for Space-Time Coding Systems

    Directory of Open Access Journals (Sweden)

    Hakan A. Çırpan

    2002-05-01

    Full Text Available Sophisticated signal processing techniques have to be developed for capacity enhancement of future wireless communication systems. In recent years, space-time coding is proposed to provide significant capacity gains over the traditional communication systems in fading wireless channels. Space-time codes are obtained by combining channel coding, modulation, transmit diversity, and optional receive diversity in order to provide diversity at the receiver and coding gain without sacrificing the bandwidth. In this paper, we consider the problem of blind estimation of space-time coded signals along with the channel parameters. Both conditional and unconditional maximum likelihood approaches are developed and iterative solutions are proposed. The conditional maximum likelihood algorithm is based on iterative least squares with projection whereas the unconditional maximum likelihood approach is developed by means of finite state Markov process modelling. The performance analysis issues of the proposed methods are studied. Finally, some simulation results are presented.

  12. Estimation of real-time runway surface contamination using flight data recorder parameters

    Science.gov (United States)

    Curry, Donovan

    Within this research effort, the development of an analytic process for friction coefficient estimation is presented. Under static equilibrium, the sum of forces and moments acting on the aircraft, in the aircraft body coordinate system, while on the ground at any instant is equal to zero. Under this premise the longitudinal, lateral and normal forces due to landing are calculated along with the individual deceleration components existent when an aircraft comes to a rest during ground roll. In order to validate this hypothesis a six degree of freedom aircraft model had to be created and landing tests had to be simulated on different surfaces. The simulated aircraft model includes a high fidelity aerodynamic model, thrust model, landing gear model, friction model and antiskid model. Three main surfaces were defined in the friction model; dry, wet and snow/ice. Only the parameters recorded by an FDR are used directly from the aircraft model all others are estimated or known a priori. The estimation of unknown parameters is also presented in the research effort. With all needed parameters a comparison and validation with simulated and estimated data, under different runway conditions, is performed. Finally, this report presents results of a sensitivity analysis in order to provide a measure of reliability of the analytic estimation process. Linear and non-linear sensitivity analysis has been performed in order to quantify the level of uncertainty implicit in modeling estimated parameters and how they can affect the calculation of the instantaneous coefficient of friction. Using the approach of force and moment equilibrium about the CG at landing to reconstruct the instantaneous coefficient of friction appears to be a reasonably accurate estimate when compared to the simulated friction coefficient. This is also true when the FDR and estimated parameters are introduced to white noise and when crosswind is introduced to the simulation. After the linear analysis the

  13. Estimation of transit times in a Karst Aquifer system using environmental tracers: Application on the Jeita Aquifer system-Lebanon.

    Science.gov (United States)

    Doummar, Joanna; Hamdan, Ahmad

    2016-04-01

    Estimating transit times is essential for the assessment of aquifer vulnerability to contaminants. Groundwater in karst aquifer is assumed to be relatively young due to fast preferential pathways; slow flow components are present in water stored in the fissured matrix. Furthermore, transit times are site specific as they depend on recharge rates, temperatures, elevation, and flow media; saturated and unsaturated zones. These differences create significant variation in the groundwater age in karst systems as the water sampled will be a mix of different water that has been transported through different flow pathways (fissured matrix and conduits). Several methods can be applied to estimate water transit time of an aquifer such as artificial tracers, which provide an estimate for fast flow velocities. In this study, groundwater residence times in the Jeita spring aquifer (Lebanon) were estimated using several environmental tracers such as Chlorofluorocarbons (CFCs), Sulfur Hexafluoride (SF6), Helium-Tritium (3H, 3H- 3He). Additional stable isotope and major ion analysis was performed to characterize water types. Groundwater samples were collected from six different wells in the Jeita catchment area (Jurassic Kesrouane aquifer) as well as from the spring and cave itself. The results are reproducible for the Tritium-Helium method, unlike for the CFC/SF6 methods that yielded poor results due to sampling problems. Tritium concentrations in all groundwater samples show nearly the same concentration (~2.73 TU) except for one sample with relatively lower tritium concentration (~2.26 TU). Ages ranging from 0.07 ± 0.07 years to 23.59 ± 0.00 years were obtained. The youngest age is attributed to the spring/ cave while the oldest ages were obtained in wells tapping the fissured matrix. Neon in these samples showed considerable variations and high delta Ne in some samples indicating high excess air. Four (4) samples showed extreme excess air (Delta-Ne is greater than 70 %) and

  14. Detection probabilities for time-domain velocity estimation

    DEFF Research Database (Denmark)

    Jensen, Jørgen Arendt

    1991-01-01

    programs, it is demonstrated that the probability of correct estimation depends on the signal-to-noise ratio, transducer bandwidth, number of A-lines and number of samples used in the correlation estimate. The influence of applying a stationary echo-canceler is explained. The echo canceling can be modeled...

  15. Conformational Network and Residence Time Estimation of Trypsin-Benzamidine Unbinding Pathways

    OpenAIRE

    Dickson, Alex; Lotz, Samuel D.

    2016-01-01

    In this poster we present results from molecular dynamics sampling of benzamidine unbinding from trypsin. We give background on the weighted ensemble technique used (WExplore) and the Markovian state model construction. Our network shows three unique unbinding pathways including a never before observed unbinding pathway. We also estimate residence time to within one order of magnitude to the experimental value.

  16. CLEAR: a model for the calculation of evacuation-time estimates in Emergency Planning Zones

    International Nuclear Information System (INIS)

    McLean, M.A.; Moeller, M.P.; Desrosiers, A.E.

    1983-01-01

    This paper describes the methodology and application of the computer model CLEAR (Calculates Logical Evacuation And Response) which estimates the time required for a specific population density and distribution to evacuate an area using a specific transportation network. The CLEAR model simulates vehicle departure and movement on a transportation network according to the conditions and consequences of traffice flow. These include handling vehicles at intersecting road segments, calculating the velocity of travel on a road segment as a function of its vehicle density, and accounting for the delay of vehicles in traffice queues. The program also models the distribution of times required by individuals to prepare for an evacuation. CLEAR can calculate realistic evacuation time estimates using site specific data and can identify troublesome areas within an Emergency Planning Zone

  17. IEEE 802.15.4 ZigBee-Based Time-of-Arrival Estimation for Wireless Sensor Networks.

    Science.gov (United States)

    Cheon, Jeonghyeon; Hwang, Hyunsu; Kim, Dongsun; Jung, Yunho

    2016-02-05

    Precise time-of-arrival (TOA) estimation is one of the most important techniques in RF-based positioning systems that use wireless sensor networks (WSNs). Because the accuracy of TOA estimation is proportional to the RF signal bandwidth, using broad bandwidth is the most fundamental approach for achieving higher accuracy. Hence, ultra-wide-band (UWB) systems with a bandwidth of 500 MHz are commonly used. However, wireless systems with broad bandwidth suffer from the disadvantages of high complexity and high power consumption. Therefore, it is difficult to employ such systems in various WSN applications. In this paper, we present a precise time-of-arrival (TOA) estimation algorithm using an IEEE 802.15.4 ZigBee system with a narrow bandwidth of 2 MHz. In order to overcome the lack of bandwidth, the proposed algorithm estimates the fractional TOA within the sampling interval. Simulation results show that the proposed TOA estimation algorithm provides an accuracy of 0.5 m at a signal-to-noise ratio (SNR) of 8 dB and achieves an SNR gain of 5 dB as compared with the existing algorithm. In addition, experimental results indicate that the proposed algorithm provides accurate TOA estimation in a real indoor environment.

  18. Estimating the continuous-time dynamics of energy and fat metabolism in mice.

    Science.gov (United States)

    Guo, Juen; Hall, Kevin D

    2009-09-01

    The mouse has become the most popular organism for investigating molecular mechanisms of body weight regulation. But understanding the physiological context by which a molecule exerts its effect on body weight requires knowledge of energy intake, energy expenditure, and fuel selection. Furthermore, measurements of these variables made at an isolated time point cannot explain why body weight has its present value since body weight is determined by the past history of energy and macronutrient imbalance. While food intake and body weight changes can be frequently measured over several weeks (the relevant time scale for mice), correspondingly frequent measurements of energy expenditure and fuel selection are not currently feasible. To address this issue, we developed a mathematical method based on the law of energy conservation that uses the measured time course of body weight and food intake to estimate the underlying continuous-time dynamics of energy output and net fat oxidation. We applied our methodology to male C57BL/6 mice consuming various ad libitum diets during weight gain and loss over several weeks and present the first continuous-time estimates of energy output and net fat oxidation rates underlying the observed body composition changes. We show that transient energy and fat imbalances in the first several days following a diet switch can account for a significant fraction of the total body weight change. We also discovered a time-invariant curve relating body fat and fat-free masses in male C57BL/6 mice, and the shape of this curve determines how diet, fuel selection, and body composition are interrelated.

  19. Critical time delay of the pineal melatonin rhythm in humans due to weak electromagnetic exposure.

    Science.gov (United States)

    Halgamuge, Malka N

    2013-08-01

    Electromagnetic fields (EMFs) can increase free radicals, activate the stress response and alter enzyme reactions. Intracellular signalling is mediated by free radicals and enzyme kinetics is affected by radical pair recombination rates. The magnetic field component of an external EMF can delay the "recombination rate" of free radical pairs. Magnetic fields thus increase radical life-times in biological systems. Although measured in nanoseconds, this extra time increases the potential to do more damage. Melatonin regulates the body's sleep-wake cycle or circadian rhythm. The World Health Organization (WHO) has confirmed that prolonged alterations in sleep patterns suppress the body's ability to make melatonin. Considerable cancer rates have been attributed to the reduction of melatonin production as a result of jet lag and night shift work. In this study, changes in circadian rhythm and melatonin concentration are observed due to the external perturbation of chemical reaction rates. We further analyze the pineal melatonin rhythm and investigate the critical time delay or maturation time of radical pair recombination rates, exploring the impact of the mRNA degradation rate on the critical time delay. The results show that significant melatonin interruption and changes to the circadian rhythm occur due to the perturbation of chemical reaction rates, as also reported in previous studies. The results also show the influence of the mRNA degradation rate on the circadian rhythm's critical time delay or maturation time. The results support the hypothesis that exposure to weak EMFs via melatonin disruption can adversely affect human health.

  20. Parameter Estimation of a Delay Time Model of Wearing Parts Based on Objective Data

    Directory of Open Access Journals (Sweden)

    Y. Tang

    2015-01-01

    Full Text Available The wearing parts of a system have a very high failure frequency, making it necessary to carry out continual functional inspections and maintenance to protect the system from unscheduled downtime. This allows for the collection of a large amount of maintenance data. Taking the unique characteristics of the wearing parts into consideration, we establish their respective delay time models in ideal inspection cases and nonideal inspection cases. The model parameters are estimated entirely using the collected maintenance data. Then, a likelihood function of all renewal events is derived based on their occurring probability functions, and the model parameters are calculated with the maximum likelihood function method, which is solved by the CRM. Finally, using two wearing parts from the oil and gas drilling industry as examples—the filter element and the blowout preventer rubber core—the parameters of the distribution function of the initial failure time and the delay time for each example are estimated, and their distribution functions are obtained. Such parameter estimation based on objective data will contribute to the optimization of the reasonable function inspection interval and will also provide some theoretical models to support the integrity management of equipment or systems.

  1. Dipole estimation errors due to not incorporating anisotropic conductivities in realistic head models for EEG source analysis

    Science.gov (United States)

    Hallez, Hans; Staelens, Steven; Lemahieu, Ignace

    2009-10-01

    EEG source analysis is a valuable tool for brain functionality research and for diagnosing neurological disorders, such as epilepsy. It requires a geometrical representation of the human head or a head model, which is often modeled as an isotropic conductor. However, it is known that some brain tissues, such as the skull or white matter, have an anisotropic conductivity. Many studies reported that the anisotropic conductivities have an influence on the calculated electrode potentials. However, few studies have assessed the influence of anisotropic conductivities on the dipole estimations. In this study, we want to determine the dipole estimation errors due to not taking into account the anisotropic conductivities of the skull and/or brain tissues. Therefore, head models are constructed with the same geometry, but with an anisotropically conducting skull and/or brain tissue compartment. These head models are used in simulation studies where the dipole location and orientation error is calculated due to neglecting anisotropic conductivities of the skull and brain tissue. Results show that not taking into account the anisotropic conductivities of the skull yields a dipole location error between 2 and 25 mm, with an average of 10 mm. When the anisotropic conductivities of the brain tissues are neglected, the dipole location error ranges between 0 and 5 mm. In this case, the average dipole location error was 2.3 mm. In all simulations, the dipole orientation error was smaller than 10°. We can conclude that the anisotropic conductivities of the skull have to be incorporated to improve the accuracy of EEG source analysis. The results of the simulation, as presented here, also suggest that incorporation of the anisotropic conductivities of brain tissues is not necessary. However, more studies are needed to confirm these suggestions.

  2. Dipole estimation errors due to not incorporating anisotropic conductivities in realistic head models for EEG source analysis

    International Nuclear Information System (INIS)

    Hallez, Hans; Staelens, Steven; Lemahieu, Ignace

    2009-01-01

    EEG source analysis is a valuable tool for brain functionality research and for diagnosing neurological disorders, such as epilepsy. It requires a geometrical representation of the human head or a head model, which is often modeled as an isotropic conductor. However, it is known that some brain tissues, such as the skull or white matter, have an anisotropic conductivity. Many studies reported that the anisotropic conductivities have an influence on the calculated electrode potentials. However, few studies have assessed the influence of anisotropic conductivities on the dipole estimations. In this study, we want to determine the dipole estimation errors due to not taking into account the anisotropic conductivities of the skull and/or brain tissues. Therefore, head models are constructed with the same geometry, but with an anisotropically conducting skull and/or brain tissue compartment. These head models are used in simulation studies where the dipole location and orientation error is calculated due to neglecting anisotropic conductivities of the skull and brain tissue. Results show that not taking into account the anisotropic conductivities of the skull yields a dipole location error between 2 and 25 mm, with an average of 10 mm. When the anisotropic conductivities of the brain tissues are neglected, the dipole location error ranges between 0 and 5 mm. In this case, the average dipole location error was 2.3 mm. In all simulations, the dipole orientation error was smaller than 10 deg. We can conclude that the anisotropic conductivities of the skull have to be incorporated to improve the accuracy of EEG source analysis. The results of the simulation, as presented here, also suggest that incorporation of the anisotropic conductivities of brain tissues is not necessary. However, more studies are needed to confirm these suggestions.

  3. Time dependent auto-correlation, autospectrum and decay ratio estimation of transient signals in JET soft X-ray records

    International Nuclear Information System (INIS)

    Por, G.

    1999-08-01

    A program package was developed to estimate the time dependent auto-correlation function (ACF) from the time signals of soft X-ray records taken along the various lines-of-sights in JET-SHOTS, and also to estimate the time dependent Decay Ratio (DR) from that. On the basis of ACF the time dependent auto-power spectral density (APSD) was also calculated. The steps and objectives of this work were: eliminating the white detection noise, trends and slow variation from the time signals, since ordinary methods can give good estimate of the time dependent ACF and DR only for 'nearly' stationary signals, developing an automatic algorithm for finding the maxima and minima of ACF, since they are the basis for DR estimation, evaluating and testing different DR estimators for JET-SHOT, with the aim of finding parts of the signals, where the oscillating character is strong, estimating time dependent ACF and APSD that can follow the relatively fast variation in the time signal. The methods that we have developed for data processing of transient signals are: White detection noise removal and preparation for trend removal - weak components, white detection noise and high frequency components are filtered from the signal using the so-called soft-threshold wavelet filter. Removal of trends and slow variation - Three-point differentiation of the pre-filtered signal is used to remove trends and slow variation. Here we made use of the DERIV function of IDL program language. This leads to a filtered signal that has zero mean value in each time step. Calculation of the time dependent ACF - The signal treated by the two previous steps is used as the input. Calculated ACF value is added in each new time step, but the previously accumulated ACF value is multiplied by a weighting factor. Thus the new sample has 100% contribution, while the contributions from the previous samples are forgotten quickly. DR calculation - DR is a measure of the decay of oscillating ACF. This parameter was shown

  4. A general approach for the estimation of loss of life due to natural and technological disasters

    International Nuclear Information System (INIS)

    Jonkman, S.N.; Lentz, A.; Vrijling, J.K.

    2010-01-01

    In assessing the safety of engineering systems in the context of quantitative risk analysis one of the most important consequence types concerns the loss of life due to accidents and disasters. In this paper, a general approach for loss of life estimation is proposed which includes three elements: (1) the assessment of physical effects associated with the event; (2) determination of the number of exposed persons (taking into account warning and evacuation); and (3) determination of mortality amongst the population exposed. The typical characteristics of and modelling approaches for these three elements are discussed. This paper focuses on 'small probability-large consequences' events within the engineering domain. It is demonstrated how the proposed approach can be applied to various case studies, such as tunnel fires, earthquakes and flood events.

  5. Implementing a combined polar-geostationary algorithm for smoke emissions estimation in near real time

    Science.gov (United States)

    Hyer, E. J.; Schmidt, C. C.; Hoffman, J.; Giglio, L.; Peterson, D. A.

    2013-12-01

    Polar and geostationary satellites are used operationally for fire detection and smoke source estimation by many near-real-time operational users, including operational forecast centers around the globe. The input satellite radiance data are processed by data providers to produce Level-2 and Level -3 fire detection products, but processing these data into spatially and temporally consistent estimates of fire activity requires a substantial amount of additional processing. The most significant processing steps are correction for variable coverage of the satellite observations, and correction for conditions that affect the detection efficiency of the satellite sensors. We describe a system developed by the Naval Research Laboratory (NRL) that uses the full raster information from the entire constellation to diagnose detection opportunities, calculate corrections for factors such as angular dependence of detection efficiency, and generate global estimates of fire activity at spatial and temporal scales suitable for atmospheric modeling. By incorporating these improved fire observations, smoke emissions products, such as NRL's FLAMBE, are able to produce improved estimates of global emissions. This talk provides an overview of the system, demonstrates the achievable improvement over older methods, and describes challenges for near-real-time implementation.

  6. First divergence time estimate of spiders, scorpions, mites and ticks (subphylum: Chelicerata) inferred from mitochondrial phylogeny.

    Science.gov (United States)

    Jeyaprakash, Ayyamperumal; Hoy, Marjorie A

    2009-01-01

    Spiders, scorpions, mites and ticks (chelicerates) form one of the most diverse groups of arthropods on land, but their origin and times of diversification are not yet established. We estimated, for the first time, the molecular divergence times for these chelicerates using complete mitochondrial sequences from 25 taxa. All mitochondrial genes were evaluated individually or after concatenation. Sequences belonging to three missing genes (ND3, 6, and tRNA-Asp) from three taxa, as well as the faster-evolving ribosomal RNAs (12S and 16S), tRNAs, and the third base of each codon from 11 protein-coding genes (PCGs) (COI-III, CYTB, ATP8, 6, ND1-2, 4L, and 4-5), were identified and removed. The remaining concatenated sequences from 11 PCGs produced a completely resolved phylogenetic tree and confirmed that all chelicerates are monophyletic. Removing the third base from each codon was essential to resolve the phylogeny, which allowed deep divergence times to be calculated using three nodes calibrated with upper and lower priors. Our estimates indicate that the orders and classes of spiders, scorpions, mites, and ticks diversified in the late Paleozoic, much earlier than previously reported from fossil date estimates. The divergence time estimated for ticks suggests that their first land hosts could have been amphibians rather than reptiles. Using molecular data, we separated the spider-scorpion clades and estimated their divergence times at 397 +/- 23 million years ago. Algae, fungi, plants, and animals, including insects, were well established on land when these chelicerates diversified. Future analyses, involving mitochondrial sequences from additional chelicerate taxa and the inclusion of nuclear genes (or entire genomes) will provide a more complete picture of the evolution of the Chelicerata, the second most abundant group of animals on earth.

  7. The estimation of time-varying risks in asset pricing modelling using B-Spline method

    Science.gov (United States)

    Nurjannah; Solimun; Rinaldo, Adji

    2017-12-01

    Asset pricing modelling has been extensively studied in the past few decades to explore the risk-return relationship. The asset pricing literature typically assumed a static risk-return relationship. However, several studies found few anomalies in the asset pricing modelling which captured the presence of the risk instability. The dynamic model is proposed to offer a better model. The main problem highlighted in the dynamic model literature is that the set of conditioning information is unobservable and therefore some assumptions have to be made. Hence, the estimation requires additional assumptions about the dynamics of risk. To overcome this problem, the nonparametric estimators can also be used as an alternative for estimating risk. The flexibility of the nonparametric setting avoids the problem of misspecification derived from selecting a functional form. This paper investigates the estimation of time-varying asset pricing model using B-Spline, as one of nonparametric approach. The advantages of spline method is its computational speed and simplicity, as well as the clarity of controlling curvature directly. The three popular asset pricing models will be investigated namely CAPM (Capital Asset Pricing Model), Fama-French 3-factors model and Carhart 4-factors model. The results suggest that the estimated risks are time-varying and not stable overtime which confirms the risk instability anomaly. The results is more pronounced in Carhart’s 4-factors model.

  8. Estimation of modal parameters using bilinear joint time frequency distributions

    Science.gov (United States)

    Roshan-Ghias, A.; Shamsollahi, M. B.; Mobed, M.; Behzad, M.

    2007-07-01

    In this paper, a new method is proposed for modal parameter estimation using time-frequency representations. Smoothed Pseudo Wigner-Ville distribution which is a member of the Cohen's class distributions is used to decouple vibration modes completely in order to study each mode separately. This distribution reduces cross-terms which are troublesome in Wigner-Ville distribution and retains the resolution as well. The method was applied to highly damped systems, and results were superior to those obtained via other conventional methods.

  9. Mental workload and cognitive task automaticity: an evaluation of subjective and time estimation metrics.

    Science.gov (United States)

    Liu, Y; Wickens, C D

    1994-11-01

    The evaluation of mental workload is becoming increasingly important in system design and analysis. The present study examined the structure and assessment of mental workload in performing decision and monitoring tasks by focusing on two mental workload measurements: subjective assessment and time estimation. The task required the assignment of a series of incoming customers to the shortest of three parallel service lines displayed on a computer monitor. The subject was either in charge of the customer assignment (manual mode) or was monitoring an automated system performing the same task (automatic mode). In both cases, the subjects were required to detect the non-optimal assignments that they or the computer had made. Time pressure was manipulated by the experimenter to create fast and slow conditions. The results revealed a multi-dimensional structure of mental workload and a multi-step process of subjective workload assessment. The results also indicated that subjective workload was more influenced by the subject's participatory mode than by the factor of task speed. The time estimation intervals produced while performing the decision and monitoring tasks had significantly greater length and larger variability than those produced while either performing no other tasks or performing a well practised customer assignment task. This result seemed to indicate that time estimation was sensitive to the presence of perceptual/cognitive demands, but not to response related activities to which behavioural automaticity has developed.

  10. Estimating DSGE model parameters in a small open economy: Do real-time data matter?

    Directory of Open Access Journals (Sweden)

    Capek Jan

    2015-03-01

    Full Text Available This paper investigates the differences between parameters estimated using real-time and those estimated with revised data. The models used are New Keynesian DSGE models of the Czech, Polish, Hungarian, Swiss, and Swedish small open economies in interaction with the euro area. The paper also offers an analysis of data revisions of GDP growth and inflation and trend revisions of interest rates.

  11. A Novel Non-Iterative Method for Real-Time Parameter Estimation of the Fricke-Morse Model

    Directory of Open Access Journals (Sweden)

    SIMIC, M.

    2016-11-01

    Full Text Available Parameter estimation of Fricke-Morse model of biological tissue is widely used in bioimpedance data processing and analysis. Complex nonlinear least squares (CNLS data fitting is often used for parameter estimation of the model, but limitations such as high processing time, converging into local minimums, need for good initial guess of model parameters and non-convergence have been reported. Thus, there is strong motivation to develop methods which can solve these flaws. In this paper a novel real-time method for parameter estimation of Fricke-Morse model of biological cells is presented. The proposed method uses the value of characteristic frequency estimated from the measured imaginary part of bioimpedance, whereupon the Fricke-Morse model parameters are calculated using the provided analytical expressions. The proposed method is compared with CNLS in frequency ranges of 1 kHz to 10 MHz (beta-dispersion and 10 kHz to 100 kHz, which is more suitable for low-cost microcontroller-based bioimpedance measurement systems. The obtained results are promising, and in both frequency ranges, CNLS and the proposed method have accuracies suitable for most electrical bioimpedance (EBI applications. However, the proposed algorithm has significantly lower computation complexity, so it was 20-80 times faster than CNLS.

  12. Vision-based online vibration estimation of the in-vessel inspection flexible robot with short-time Fourier transformation

    International Nuclear Information System (INIS)

    Wang, Hesheng; Chen, Weidong; Xu, Lifei; He, Tao

    2015-01-01

    Highlights: • Vision-based online vibration estimation method for a flexible arm is proposed. • The vibration signal is obtained by image processing in unknown environments. • Vibration parameters are estimated by short-time Fourier transformation. - Abstract: The vibration should be suppressed if it happens during the motion of a flexible robot or under the influence of external disturbance caused by its structural features and material properties, because the vibration may affect the positioning accuracy and image quality. In Tokamak environment, we need to get the real-time vibration information on vibration suppression of robotic arm, however, some sensors are not allowed in the extreme Tokamak environment. This paper proposed a vision-based method for online vibration estimation of a flexible manipulator, which is achieved by utilizing the environment image information from the end-effector camera to estimate its vibration. Short-time Fourier Transformation with adaptive window length method is used to estimate vibration parameters of non-stationary vibration signals. Experiments with one-link flexible manipulator equipped with camera are carried out to validate the feasibility of this method in this paper.

  13. Vision-based online vibration estimation of the in-vessel inspection flexible robot with short-time Fourier transformation

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Hesheng [Key Laboratory of System Control and Information Processing, Ministry of Education of China (China); Department of Automation, Shanghai Jiao Tong University, Shanghai 200240 (China); Chen, Weidong, E-mail: wdchen@sjtu.edu.cn [Key Laboratory of System Control and Information Processing, Ministry of Education of China (China); Department of Automation, Shanghai Jiao Tong University, Shanghai 200240 (China); Xu, Lifei; He, Tao [Key Laboratory of System Control and Information Processing, Ministry of Education of China (China); Department of Automation, Shanghai Jiao Tong University, Shanghai 200240 (China)

    2015-10-15

    Highlights: • Vision-based online vibration estimation method for a flexible arm is proposed. • The vibration signal is obtained by image processing in unknown environments. • Vibration parameters are estimated by short-time Fourier transformation. - Abstract: The vibration should be suppressed if it happens during the motion of a flexible robot or under the influence of external disturbance caused by its structural features and material properties, because the vibration may affect the positioning accuracy and image quality. In Tokamak environment, we need to get the real-time vibration information on vibration suppression of robotic arm, however, some sensors are not allowed in the extreme Tokamak environment. This paper proposed a vision-based method for online vibration estimation of a flexible manipulator, which is achieved by utilizing the environment image information from the end-effector camera to estimate its vibration. Short-time Fourier Transformation with adaptive window length method is used to estimate vibration parameters of non-stationary vibration signals. Experiments with one-link flexible manipulator equipped with camera are carried out to validate the feasibility of this method in this paper.

  14. Postmortem time estimation using body temperature and a finite-element computer model

    NARCIS (Netherlands)

    Hartog, E.A. den; Lotens, W.A.

    2004-01-01

    In the Netherlands most murder victims are found 2-24 h after the crime. During this period, body temperature decrease is the most reliable method to estimate the postmortem time (PMT). Recently, two murder cases were analysed in which currently available methods did not provide a su.ciently

  15. A probabilistic model for estimating the waiting time until the simultaneous collapse of two contingencies

    International Nuclear Information System (INIS)

    Barnett, C.S.

    1991-06-01

    The Double Contingency Principle (DCP) is widely applied to criticality safety practice in the United States. Most practitioners base their application of the principle on qualitative, intuitive assessments. The recent trend toward probabilistic safety assessments provides a motive to search for a quantitative, probabilistic foundation for the DCP. A Markov model is tractable and leads to relatively simple results. The model yields estimates of mean time to simultaneous collapse of two contingencies as a function of estimates of mean failure times and mean recovery times of two independent contingencies. The model is a tool that can be used to supplement the qualitative methods now used to assess effectiveness of the DCP. 3 refs., 1 fig

  16. A probabilistic model for estimating the waiting time until the simultaneous collapse of two contingencies

    International Nuclear Information System (INIS)

    Barnett, C.S.

    1991-01-01

    The Double Contingency Principle (DCP) is widely applied to criticality safety practice in the United States. Most practitioners base their application of the principle on qualitative, intuitive assessments. The recent trend toward probabilistic safety assessments provides a motive to search for a quantitative, probabilistic foundation for the DCP. A Markov model is tractable and leads to relatively simple results. The model yields estimates of mean time to simultaneous collapse of two contingencies as a function of estimates of mean failure times and mean recovery times of two independent contingencies. The model is a tool that can be used to supplement the qualitative methods now used to assess effectiveness of the DCP. (Author)

  17. A probabilistic model for estimating the waiting time until the simultaneous collapse of two contingencies

    International Nuclear Information System (INIS)

    Barnett, C.S.

    1992-01-01

    The double contingency principle (DCP) is widely applied to criticality safety practice in the United States. Most practitioners base their application of the principle on qualitative and intuitive assessments. The recent trend toward probabilistic safety assessments provides a motive for a search for a quantitative and probabilistic foundation for the DCP. A Markov model is tractable and leads to relatively simple results. The model yields estimates of mean time to simultaneous collapse of two contingencies, as functions of estimates of mean failure times and mean recovery times of two independent contingencies. The model is a tool that can be used to supplement the qualitative methods now used to assess the effectiveness of the DCP. (Author)

  18. Estimates concentrations in bottled 222Rn of the dose due to mineral waters in Iran

    International Nuclear Information System (INIS)

    Assadi, M. R.; Esmaealnejad, M.; Rahmatinejad, Z.

    2006-01-01

    Radon is a radionuclide that has the main role in exposure. Radon in water causes exposure in whole body but the largest dose being received by the stomach, as EPA (Environmental Protection Agency) estimates that radon in drinking water causes about 168 cancer deaths per year: 89 p ercent f rom lung cancer caused by breathing released to the indoor air from water and 11 p ercent f rom stomach cancer caused by consuming water containing radon. Now days the consumption of bottled mineral waters has become very popular. As is known, some kinds of mineral waters contain naturally occurring radionuclides in higher concentration than the usual drinking (tap) water. Surveys and reports on radon in most surface waters is low compared with radon level in groundwater and mineral water. In our work, the concentration of Rn(222) was determined in some bottled mineral waters available in Iran , and in next step the dose contribution ; due to ingestion ; for 1 l d -1 bottled mineral water consumption.

  19. Learning time-dependent noise to reduce logical errors: real time error rate estimation in quantum error correction

    Science.gov (United States)

    Huo, Ming-Xia; Li, Ying

    2017-12-01

    Quantum error correction is important to quantum information processing, which allows us to reliably process information encoded in quantum error correction codes. Efficient quantum error correction benefits from the knowledge of error rates. We propose a protocol for monitoring error rates in real time without interrupting the quantum error correction. Any adaptation of the quantum error correction code or its implementation circuit is not required. The protocol can be directly applied to the most advanced quantum error correction techniques, e.g. surface code. A Gaussian processes algorithm is used to estimate and predict error rates based on error correction data in the past. We find that using these estimated error rates, the probability of error correction failures can be significantly reduced by a factor increasing with the code distance.

  20. Estimating primary production from oxygen time series: A novel approach in the frequency domain

    NARCIS (Netherlands)

    Cox, T.J.S.; Maris, T.; Soetaert, K.; Kromkamp, J.C.; Meire, P.; Meysman, F.J.R.

    2015-01-01

    Based on an analysis in the frequency domain of the governing equation of oxygen dynamics in aquatic systems, we derive a new method for estimating gross primary production (GPP) from oxygen time series. The central result of this article is a relation between time averaged GPP and the amplitude of

  1. Accuracy of Perceived Estimated Travel Time by EMS to a Trauma Center in San Bernardino County, California

    Directory of Open Access Journals (Sweden)

    Michael M. Neeki

    2016-06-01

    Full Text Available Introduction: Mobilization of trauma resources has the potential to cause ripple effects throughout hospital operations. One major factor affecting efficient utilization of trauma resources is a discrepancy between the prehospital estimated time of arrival (ETA as communicated by emergency medical services (EMS personnel and their actual time of arrival (TOA. The current study aimed to assess the accuracy of the perceived prehospital estimated arrival time by EMS personnel in comparison to their actual arrival time at a Level II trauma center in San Bernardino County, California. Methods: This retrospective study included traumas classified as alerts or activations that were transported to Arrowhead Regional Medical Center in 2013. We obtained estimated arrival time and actual arrival time for each transport from the Surgery Department Trauma Registry. The difference between the median of ETA and actual TOA by EMS crews to the trauma center was calculated for these transports. Additional variables assessed included time of day and month during which the transport took place. Results: A total of 2,454 patients classified as traumas were identified in the Surgery Department Trauma Registry. After exclusion of trauma consults, walk-ins, handoffs between agencies, downgraded traumas, traumas missing information, and traumas transported by agencies other than American Medical Response, Ontario Fire, Rialto Fire or San Bernardino County Fire, we included a final sample size of 555 alert and activation classified traumas in the final analysis. When combining all transports by the included EMS agencies, the median of the ETA was 10 minutes and the median of the actual TOA was 22 minutes (median of difference=9 minutes, p<0.0001. Furthermore, when comparing the difference between trauma alerts and activations, trauma activations demonstrated an equal or larger difference in the median of the estimated and actual time of arrival (p<0.0001. We also found

  2. Novel approaches to estimating the turbulent kinetic energy dissipation rate from low- and moderate-resolution velocity fluctuation time series

    Directory of Open Access Journals (Sweden)

    M. Wacławczyk

    2017-11-01

    Full Text Available In this paper we propose two approaches to estimating the turbulent kinetic energy (TKE dissipation rate, based on the zero-crossing method by Sreenivasan et al. (1983. The original formulation requires a fine resolution of the measured signal, down to the smallest dissipative scales. However, due to finite sampling frequency, as well as measurement errors, velocity time series obtained from airborne experiments are characterized by the presence of effective spectral cutoffs. In contrast to the original formulation the new approaches are suitable for use with signals originating from airborne experiments. The suitability of the new approaches is tested using measurement data obtained during the Physics of Stratocumulus Top (POST airborne research campaign as well as synthetic turbulence data. They appear useful and complementary to existing methods. We show the number-of-crossings-based approaches respond differently to errors due to finite sampling and finite averaging than the classical power spectral method. Hence, their application for the case of short signals and small sampling frequencies is particularly interesting, as it can increase the robustness of turbulent kinetic energy dissipation rate retrieval.

  3. Statistical estimates of absenteeism attributable to seasonal and pandemic influenza from the Canadian Labour Force Survey.

    Science.gov (United States)

    Schanzer, Dena L; Zheng, Hui; Gilmore, Jason

    2011-04-12

    As many respiratory viruses are responsible for influenza like symptoms, accurate measures of the disease burden are not available and estimates are generally based on statistical methods. The objective of this study was to estimate absenteeism rates and hours lost due to seasonal influenza and compare these estimates with estimates of absenteeism attributable to the two H1N1 pandemic waves that occurred in 2009. Key absenteeism variables were extracted from Statistics Canada's monthly labour force survey (LFS). Absenteeism and the proportion of hours lost due to own illness or disability were modelled as a function of trend, seasonality and proxy variables for influenza activity from 1998 to 2009. Hours lost due to the H1N1/09 pandemic strain were elevated compared to seasonal influenza, accounting for a loss of 0.2% of potential hours worked annually. In comparison, an estimated 0.08% of hours worked annually were lost due to seasonal influenza illnesses. Absenteeism rates due to influenza were estimated at 12% per year for seasonal influenza over the 1997/98 to 2008/09 seasons, and 13% for the two H1N1/09 pandemic waves. Employees who took time off due to a seasonal influenza infection took an average of 14 hours off. For the pandemic strain, the average absence was 25 hours. This study confirms that absenteeism due to seasonal influenza has typically ranged from 5% to 20%, with higher rates associated with multiple circulating strains. Absenteeism rates for the 2009 pandemic were similar to those occurring for seasonal influenza. Employees took more time off due to the pandemic strain than was typical for seasonal influenza.

  4. Statistical estimates of absenteeism attributable to seasonal and pandemic influenza from the Canadian Labour Force Survey

    Science.gov (United States)

    2011-01-01

    Background As many respiratory viruses are responsible for influenza like symptoms, accurate measures of the disease burden are not available and estimates are generally based on statistical methods. The objective of this study was to estimate absenteeism rates and hours lost due to seasonal influenza and compare these estimates with estimates of absenteeism attributable to the two H1N1 pandemic waves that occurred in 2009. Methods Key absenteeism variables were extracted from Statistics Canada's monthly labour force survey (LFS). Absenteeism and the proportion of hours lost due to own illness or disability were modelled as a function of trend, seasonality and proxy variables for influenza activity from 1998 to 2009. Results Hours lost due to the H1N1/09 pandemic strain were elevated compared to seasonal influenza, accounting for a loss of 0.2% of potential hours worked annually. In comparison, an estimated 0.08% of hours worked annually were lost due to seasonal influenza illnesses. Absenteeism rates due to influenza were estimated at 12% per year for seasonal influenza over the 1997/98 to 2008/09 seasons, and 13% for the two H1N1/09 pandemic waves. Employees who took time off due to a seasonal influenza infection took an average of 14 hours off. For the pandemic strain, the average absence was 25 hours. Conclusions This study confirms that absenteeism due to seasonal influenza has typically ranged from 5% to 20%, with higher rates associated with multiple circulating strains. Absenteeism rates for the 2009 pandemic were similar to those occurring for seasonal influenza. Employees took more time off due to the pandemic strain than was typical for seasonal influenza. PMID:21486453

  5. Theoretical considerations in measurement of time discrepancies between input and myocardial time-signal intensity curves in estimates of regional myocardial perfusion with first-pass contrast-enhanced MRI.

    Science.gov (United States)

    Natsume, Takahiro; Ishida, Masaki; Kitagawa, Kakuya; Nagata, Motonori; Sakuma, Hajime; Ichihara, Takashi

    2015-11-01

    The purpose of this study was to develop a method to determine time discrepancies between input and myocardial time-signal intensity (TSI) curves for accurate estimation of myocardial perfusion with first-pass contrast-enhanced MRI. Estimation of myocardial perfusion with contrast-enhanced MRI using kinetic models requires faithful recording of contrast content in the blood and myocardium. Typically, the arterial input function (AIF) is obtained by setting a region of interest in the left ventricular cavity. However, there is a small delay between the AIF and the myocardial curves, and such time discrepancies can lead to errors in flow estimation using Patlak plot analysis. In this study, the time discrepancies between the arterial TSI curve and the myocardial tissue TSI curve were estimated based on the compartment model. In the early phase after the arrival of the contrast agent in the myocardium, the relationship between rate constant K1 and the concentrations of Gd-DTPA contrast agent in the myocardium and arterial blood (LV blood) can be described by the equation K1={dCmyo(tpeak)/dt}/Ca(tpeak), where Cmyo(t) and Ca(t) are the relative concentrations of Gd-DTPA contrast agent in the myocardium and in the LV blood, respectively, and tpeak is the time corresponding to the peak of Ca(t). In the ideal case, the time corresponding to the maximum upslope of Cmyo(t), tmax, is equal to tpeak. In practice, however, there is a small difference in the arrival times of the contrast agent into the LV and into the myocardium. This difference was estimated to correspond to the difference between tpeak and tmax. The magnitudes of such time discrepancies and the effectiveness of the correction for these time discrepancies were measured in 18 subjects who underwent myocardial perfusion MRI under rest and stress conditions. The effects of the time discrepancies could be corrected effectively in the myocardial perfusion estimates. Copyright © 2015 Elsevier Inc. All rights

  6. Efficient Estimation for Diffusions Sampled at High Frequency Over a Fixed Time Interval

    DEFF Research Database (Denmark)

    Jakobsen, Nina Munkholt; Sørensen, Michael

    Parametric estimation for diffusion processes is considered for high frequency observations over a fixed time interval. The processes solve stochastic differential equations with an unknown parameter in the diffusion coefficient. We find easily verified conditions on approximate martingale...

  7. Introduction to State Estimation of High-Rate System Dynamics.

    Science.gov (United States)

    Hong, Jonathan; Laflamme, Simon; Dodson, Jacob; Joyce, Bryan

    2018-01-13

    Engineering systems experiencing high-rate dynamic events, including airbags, debris detection, and active blast protection systems, could benefit from real-time observability for enhanced performance. However, the task of high-rate state estimation is challenging, in particular for real-time applications where the rate of the observer's convergence needs to be in the microsecond range. This paper identifies the challenges of state estimation of high-rate systems and discusses the fundamental characteristics of high-rate systems. A survey of applications and methods for estimators that have the potential to produce accurate estimations for a complex system experiencing highly dynamic events is presented. It is argued that adaptive observers are important to this research. In particular, adaptive data-driven observers are advantageous due to their adaptability and lack of dependence on the system model.

  8. Estimating the Dieaway Time and its Precision from Shift Register Data

    Energy Technology Data Exchange (ETDEWEB)

    Croft, S.; Evans, L. G.; Henzlova, D.; Miller, K. A. [Los Alamos National Laboratory, Nuclear Nonproliferation Division, Safeguards Science and Technology Group, Los Alamos (United States)

    2012-06-15

    Neutron coincidence counting is widely used throughout the nuclear fuel cycle to quantify special nuclear materials that fission, for example Pu. Before calibration and setting to work, instruments undergo a range of performance and characterization tests that typically include measurement of the neutron dieaway characteristic. The neutron dieaway characteristic is a general measure of the likelihood that a neutron will be detected following another neutron, and describes time behavior of detected neutrons in the assay system. This paper shows how the effective 1/ecapture dieaway time of the observable coincidence (doubles) rate may be extracted simply and quickly from two coincidence counting rates recorded with different shift register time gate settings. In particular we address how to assign the precision to such estimates including when the rates are correlated.

  9. Estimation of the order of an autoregressive time series: a Bayesian approach

    International Nuclear Information System (INIS)

    Robb, L.J.

    1980-01-01

    Finite-order autoregressive models for time series are often used for prediction and other inferences. Given the order of the model, the parameters of the models can be estimated by least-squares, maximum-likelihood, or Yule-Walker method. The basic problem is estimating the order of the model. The problem of autoregressive order estimation is placed in a Bayesian framework. This approach illustrates how the Bayesian method brings the numerous aspects of the problem together into a coherent structure. A joint prior probability density is proposed for the order, the partial autocorrelation coefficients, and the variance; and the marginal posterior probability distribution for the order, given the data, is obtained. It is noted that the value with maximum posterior probability is the Bayes estimate of the order with respect to a particular loss function. The asymptotic posterior distribution of the order is also given. In conclusion, Wolfer's sunspot data as well as simulated data corresponding to several autoregressive models are analyzed according to Akaike's method and the Bayesian method. Both methods are observed to perform quite well, although the Bayesian method was clearly superior, in most cases

  10. Real-time estimation of small-area populations with human biomarkers in sewage

    International Nuclear Information System (INIS)

    Daughton, Christian G.

    2012-01-01

    urine dilution), the biomarker with the most potential for the SCIM concept for real-time measurement of population was determined to be coprostanol - the major sterol produced by microbial reduction of cholesterol in the colon. - Highlights: ► New concept proposed for estimating small-area human populations. ► Sewage Chemical-Information Mining (SCIM) measures biomarkers in sewage. ► Real-time estimation of populations (accommodating influx and efflux) is possible. ► Coprostanol is identified as a candidate biomarker for estimating population size. ► Composite biomarkers having complementary properties could improve accuracy.

  11. How can streamflow and climate-landscape data be used to estimate baseflow mean response time?

    Science.gov (United States)

    Zhang, Runrun; Chen, Xi; Zhang, Zhicai; Soulsby, Chris; Gao, Man

    2018-02-01

    Mean response time (MRT) is a metric describing the propagation of catchment hydraulic behavior that reflects both hydro-climatic conditions and catchment characteristics. To provide a comprehensive understanding of catchment response over a longer-time scale for hydraulic processes, the MRT function for baseflow generation was derived using an instantaneous unit hydrograph (IUH) model that describes the subsurface response to effective rainfall inputs. IUH parameters were estimated based on the "match test" between the autocorrelation function (ACFs) derived from the filtered base flow time series and from the IUH parameters, under the GLUE framework. Regionalization of MRT was conducted using estimates and hydroclimate-landscape indices in 22 sub-basins of the Jinghe River Basin (JRB) in the Loess Plateau of northwest China. Results indicate there is strong equifinality in determination of the best parameter sets but the median values of the MRT estimates are relatively stable in the acceptable range of the parameters. MRTs vary markedly over the studied sub-basins, ranging from tens of days to more than a year. Climate, topography and geomorphology were identified as three first-order controls on recharge-baseflow response processes. Human activities involving the cultivation of permanent crops may elongate the baseflow MRT and hence increase the dynamic storage. Cross validation suggests the model can be used to estimate MRTs in ungauged catchments in similar regions of throughout the Loess Plateau. The proposed method provides a systematic approach for MRT estimation and regionalization in terms of hydroclimate and catchment characteristics, which is helpful in the sustainable water resources utilization and ecological protection in the Loess Plateau.

  12. Cardiac MRI. Estimation of changes in normalized myocardial gadolinium accumulation over time after contrast injection in patients with acute myocarditis and healthy volunteers

    International Nuclear Information System (INIS)

    Breuckmann, F.; Buhr, C.; Maderwald, S.; Bruder, O.; Schlosser, T.; Nassenstein, K.; Erbel, R.; Barkhausen, J.

    2011-01-01

    An increased normalized gadolinium accumulation (NGA) in the myocardium during early washout has been used for the diagnosis of acute myocarditis (AM). Due to the fact that the pharmacokinetics of contrast agents are complex, time-related changes in NGA after contrast injection are likely. Because knowledge about time-related changes of NGA may improve the diagnostic accuracy of MR, our study aimed to estimate the time course of NGA after contrast injection in patients as well as in healthy volunteers. An ECG-triggered inversion recovery SSFP sequence with incrementally increasing inversion times was repetitively acquired over the 15 minutes after injection of 0.2 Gd-DTPA per kg body weight in a 4-chamber view in 15 patients with AM and 20 volunteers. The T 1relaxation times and the longitudinal relaxation rates (R1) of the myocardium and skeletal musculature were calculated for each point in time after contrast injection. The time course of NGA was estimated based on the linear relationship between R 1 and tissue Gd concentration. NGA decreased over time in the form of a negative power function in patients with AM and in healthy controls. NGA in AM tended to be higher than in controls (p > 0.05). NGA rapidly changes after contrast injection, which must be considered when measuring NGA. Although we observed a trend towards higher NGA values in patients with AM with a maximum difference one minute after contrast injection, NGA did not allow us to differentiate patients with AM from healthy volunteers, because the observed differences did not reach a level of significance. (orig.)

  13. Estimation of the dose to the nursing infant due to direct irradiation from activity present in maternal organs and tissues

    International Nuclear Information System (INIS)

    Hunt, J. G.; Nosske, D.; Dos Santos, D. S.

    2005-01-01

    ) due to the incorporation of 1 Bq of a radionuclide by the mother. This information may be used to provide external dose estimates to the infant in the case of a known or suspected radionuclide incorporation by the mother due to, for example, a nuclear medicine procedure. (authors)

  14. A flexible and coherent test/estimation procedure based on restricted mean survival times for censored time-to-event data in randomized clinical trials.

    Science.gov (United States)

    Horiguchi, Miki; Cronin, Angel M; Takeuchi, Masahiro; Uno, Hajime

    2018-04-22

    In randomized clinical trials where time-to-event is the primary outcome, almost routinely, the logrank test is prespecified as the primary test and the hazard ratio is used to quantify treatment effect. If the ratio of 2 hazard functions is not constant, the logrank test is not optimal and the interpretation of hazard ratio is not obvious. When such a nonproportional hazards case is expected at the design stage, the conventional practice is to prespecify another member of weighted logrank tests, eg, Peto-Prentice-Wilcoxon test. Alternatively, one may specify a robust test as the primary test, which can capture various patterns of difference between 2 event time distributions. However, most of those tests do not have companion procedures to quantify the treatment difference, and investigators have fallen back on reporting treatment effect estimates not associated with the primary test. Such incoherence in the "test/estimation" procedure may potentially mislead clinicians/patients who have to balance risk-benefit for treatment decision. To address this, we propose a flexible and coherent test/estimation procedure based on restricted mean survival time, where the truncation time τ is selected data dependently. The proposed procedure is composed of a prespecified test and an estimation of corresponding robust and interpretable quantitative treatment effect. The utility of the new procedure is demonstrated by numerical studies based on 2 randomized cancer clinical trials; the test is dramatically more powerful than the logrank, Wilcoxon tests, and the restricted mean survival time-based test with a fixed τ, for the patterns of difference seen in these cancer clinical trials. Copyright © 2018 John Wiley & Sons, Ltd.

  15. Nonparametric estimation in an "illness-death" model when all transition times are interval censored

    DEFF Research Database (Denmark)

    Frydman, Halina; Gerds, Thomas; Grøn, Randi

    2013-01-01

    We develop nonparametric maximum likelihood estimation for the parameters of an irreversible Markov chain on states {0,1,2} from the observations with interval censored times of 0 → 1, 0 → 2 and 1 → 2 transitions. The distinguishing aspect of the data is that, in addition to all transition times ...

  16. Cancer incidence due to excess body weight and leisure-time physical inactivity in Canada: implications for prevention.

    Science.gov (United States)

    Brenner, Darren R

    2014-09-01

    This analysis aimed to estimate the number of incident cases of various cancers attributable to excess body weight (overweight, obesity) and leisure-time physical inactivity annually in Canada. The number of attributable cancers was estimated using the population attributable fraction (PAF), risk estimates from recent meta-analyses and population exposure prevalence estimates obtained from the Canadian Community Health Survey (2000). Age-sex-site-specific cancer incidence was obtained from Statistics Canada tables for the most up-to-date year with full national data, 2007. Where the evidence for association has been deemed sufficient, we estimated the number of incident cases of the following cancers attributable to obesity: colon, breast, endometrium, esophagus (adenocarcinomas), gallbladder, pancreas and kidney; and to physical inactivity: colon, breast, endometrium, prostate, lung and/or bronchus, and ovarian. Overall, estimates of all cancer incidence in 2007 suggest that at least 3.5% (n=5771) and 7.9% (n=12,885) are attributed to excess body weight and physical inactivity respectively. For both risk factors the burden of disease was greater among women than among men. Thousands of incident cases of cancer could be prevented annually in Canada as good evidence exists for effective interventions to reduce these risk factors in the population. Copyright © 2014. Published by Elsevier Inc.

  17. On protecting the planet against cosmic attack: Ultrafast real-time estimate of the asteroid's radial velocity

    Science.gov (United States)

    Zakharchenko, V. D.; Kovalenko, I. G.

    2014-05-01

    A new method for the line-of-sight velocity estimation of a high-speed near-Earth object (asteroid, meteorite) is suggested. The method is based on the use of fractional, one-half order derivative of a Doppler signal. The algorithm suggested is much simpler and more economical than the classical one, and it appears preferable for use in orbital weapon systems of threat response. Application of fractional differentiation to quick evaluation of mean frequency location of the reflected Doppler signal is justified. The method allows an assessment of the mean frequency in the time domain without spectral analysis. An algorithm structure for the real-time estimation is presented. The velocity resolution estimates are made for typical asteroids in the X-band. It is shown that the wait time can be shortened by orders of magnitude compared with similar value in the case of a standard spectral processing.

  18. Vehicle Position Estimation Based on Magnetic Markers: Enhanced Accuracy by Compensation of Time Delays

    Directory of Open Access Journals (Sweden)

    Yeun-Sub Byun

    2015-11-01

    Full Text Available The real-time recognition of absolute (or relative position and orientation on a network of roads is a core technology for fully automated or driving-assisted vehicles. This paper presents an empirical investigation of the design, implementation, and evaluation of a self-positioning system based on a magnetic marker reference sensing method for an autonomous vehicle. Specifically, the estimation accuracy of the magnetic sensing ruler (MSR in the up-to-date estimation of the actual position was successfully enhanced by compensating for time delays in signal processing when detecting the vertical magnetic field (VMF in an array of signals. In this study, the signal processing scheme was developed to minimize the effects of the distortion of measured signals when estimating the relative positional information based on magnetic signals obtained using the MSR. In other words, the center point in a 2D magnetic field contour plot corresponding to the actual position of magnetic markers was estimated by tracking the errors between pre-defined reference models and measured magnetic signals. The algorithm proposed in this study was validated by experimental measurements using a test vehicle on a pilot network of roads. From the results, the positioning error was found to be less than 0.04 m on average in an operational test.

  19. Time-varying effect moderation using the structural nested mean model: estimation using inverse-weighted regression with residuals

    Science.gov (United States)

    Almirall, Daniel; Griffin, Beth Ann; McCaffrey, Daniel F.; Ramchand, Rajeev; Yuen, Robert A.; Murphy, Susan A.

    2014-01-01

    This article considers the problem of examining time-varying causal effect moderation using observational, longitudinal data in which treatment, candidate moderators, and possible confounders are time varying. The structural nested mean model (SNMM) is used to specify the moderated time-varying causal effects of interest in a conditional mean model for a continuous response given time-varying treatments and moderators. We present an easy-to-use estimator of the SNMM that combines an existing regression-with-residuals (RR) approach with an inverse-probability-of-treatment weighting (IPTW) strategy. The RR approach has been shown to identify the moderated time-varying causal effects if the time-varying moderators are also the sole time-varying confounders. The proposed IPTW+RR approach provides estimators of the moderated time-varying causal effects in the SNMM in the presence of an additional, auxiliary set of known and measured time-varying confounders. We use a small simulation experiment to compare IPTW+RR versus the traditional regression approach and to compare small and large sample properties of asymptotic versus bootstrap estimators of the standard errors for the IPTW+RR approach. This article clarifies the distinction between time-varying moderators and time-varying confounders. We illustrate the methodology in a case study to assess if time-varying substance use moderates treatment effects on future substance use. PMID:23873437

  20. A Polynomial Estimate of Railway Line Delay

    DEFF Research Database (Denmark)

    Cerreto, Fabrizio; Harrod, Steven; Nielsen, Otto Anker

    2017-01-01

    Railway service may be measured by the aggregate delay over a time horizon or due to an event. Timetables for railway service may dampen aggregate delay by addition of additional process time, either supplement time or buffer time. The evaluation of these variables has previously been performed...... by numerical analysis with simulation. This paper proposes an analytical estimate of aggregate delay with a polynomial form. The function returns the aggregate delay of a railway line resulting from an initial, primary, delay. Analysis of the function demonstrates that there should be a balance between the two...