#### Sample records for positive likelihood ratios

1. Evidence Based Medicine; Positive and Negative Likelihood Ratios of Diagnostic Tests

Directory of Open Access Journals (Sweden)

Alireza Baratloo

2015-10-01

Full Text Available In the previous two parts of educational manuscript series in Emergency, we explained some screening characteristics of diagnostic tests including accuracy, sensitivity, specificity, and positive and negative predictive values. In the 3rd  part we aimed to explain positive and negative likelihood ratio (LR as one of the most reliable performance measures of a diagnostic test. To better understand this characteristic of a test, it is first necessary to fully understand the concept of sensitivity and specificity. So we strongly advise you to review the 1st part of this series again. In short, the likelihood ratios are about the percentage of people with and without a disease but having the same test result. The prevalence of a disease can directly influence screening characteristics of a diagnostic test, especially its sensitivity and specificity. Trying to eliminate this effect, LR was developed. Pre-test probability of a disease multiplied by positive or negative LR can estimate post-test probability. Therefore, LR is the most important characteristic of a test to rule out or rule in a diagnosis. A positive likelihood ratio > 1 means higher probability of the disease to be present in a patient with a positive test. The further from 1, either higher or lower, the stronger the evidence to rule in or rule out the disease, respectively. It is obvious that tests with LR close to one are less practical. On the other hand, LR further from one will have more value for application in medicine. Usually tests with 0.1 < LR > 10 are considered suitable for implication in routine practice.

2. The Laplace Likelihood Ratio Test for Heteroscedasticity

Directory of Open Access Journals (Sweden)

J. Martin van Zyl

2011-01-01

Full Text Available It is shown that the likelihood ratio test for heteroscedasticity, assuming the Laplace distribution, gives good results for Gaussian and fat-tailed data. The likelihood ratio test, assuming normality, is very sensitive to any deviation from normality, especially when the observations are from a distribution with fat tails. Such a likelihood test can also be used as a robust test for a constant variance in residuals or a time series if the data is partitioned into groups.

3. Likelihood-ratio-based biometric verification

NARCIS (Netherlands)

Bazen, A.M.; Veldhuis, Raymond N.J.

2002-01-01

This paper presents results on optimal similarity measures for biometric verification based on fixed-length feature vectors. First, we show that the verification of a single user is equivalent to the detection problem, which implies that for single-user verification the likelihood ratio is optimal.

4. Likelihood Ratio-Based Biometric Verification

NARCIS (Netherlands)

Bazen, A.M.; Veldhuis, Raymond N.J.

The paper presents results on optimal similarity measures for biometric verification based on fixed-length feature vectors. First, we show that the verification of a single user is equivalent to the detection problem, which implies that, for single-user verification, the likelihood ratio is optimal.

5. Obtaining reliable Likelihood Ratio tests from simulated likelihood functions

DEFF Research Database (Denmark)

Andersen, Laura Mørch

It is standard practice by researchers and the default option in many statistical programs to base test statistics for mixed models on simulations using asymmetric draws (e.g. Halton draws). This paper shows that when the estimated likelihood functions depend on standard deviations of mixed param...

6. The behavior of the likelihood ratio test for testing missingness

OpenAIRE

Hens, Niel; Aerts, Marc; Molenberghs, Geert; Thijs, Herbert

2003-01-01

To asses the sensitivity of conclusions to model choices in the context of selection models for non-random dropout, one can oppose the different missing mechanisms to each other; e.g. by the likelihood ratio tests. The finite sample behavior of the null distribution and the power of the likelihood ratio test is studied under a variety of missingness mechanisms. missing data; sensitivity analysis; likelihood ratio test; missing mechanisms

7. Posterior distributions for likelihood ratios in forensic science.

Science.gov (United States)

van den Hout, Ardo; Alberink, Ivo

2016-09-01

Evaluation of evidence in forensic science is discussed using posterior distributions for likelihood ratios. Instead of eliminating the uncertainty by integrating (Bayes factor) or by conditioning on parameter values, uncertainty in the likelihood ratio is retained by parameter uncertainty derived from posterior distributions. A posterior distribution for a likelihood ratio can be summarised by the median and credible intervals. Using the posterior mean of the distribution is not recommended. An analysis of forensic data for body height estimation is undertaken. The posterior likelihood approach has been criticised both theoretically and with respect to applicability. This paper addresses the latter and illustrates an interesting application area. Copyright © 2016 The Chartered Society of Forensic Sciences. Published by Elsevier Ireland Ltd. All rights reserved.

8. Nearly Efficient Likelihood Ratio Tests of the Unit Root Hypothesis

DEFF Research Database (Denmark)

Jansson, Michael; Nielsen, Morten Ørregaard

Seemingly absent from the arsenal of currently available "nearly efficient" testing procedures for the unit root hypothesis, i.e. tests whose local asymptotic power functions are indistinguishable from the Gaussian power envelope, is a test admitting a (quasi-)likelihood ratio interpretation. We...... show that the likelihood ratio unit root test derived in a Gaussian AR(1) model with standard normal innovations is nearly efficient in that model. Moreover, these desirable properties carry over to more complicated models allowing for serially correlated and/or non-Gaussian innovations....

9. Nearly Efficient Likelihood Ratio Tests for Seasonal Unit Roots

DEFF Research Database (Denmark)

Jansson, Michael; Nielsen, Morten Ørregaard

In an important generalization of zero frequency autore- gressive unit root tests, Hylleberg, Engle, Granger, and Yoo (1990) developed regression-based tests for unit roots at the seasonal frequencies in quarterly time series. We develop likelihood ratio tests for seasonal unit roots and show...... that these tests are "nearly efficient" in the sense of Elliott, Rothenberg, and Stock (1996), i.e. that their local asymptotic power functions are indistinguishable from the Gaussian power envelope. Currently available nearly efficient testing procedures for seasonal unit roots are regression-based and require...... the choice of a GLS detrending parameter, which our likelihood ratio tests do not....

10. Likelihood ratio decisions in memory: three implied regularities.

Science.gov (United States)

Glanzer, Murray; Hilford, Andrew; Maloney, Laurence T

2009-06-01

We analyze four general signal detection models for recognition memory that differ in their distributional assumptions. Our analyses show that a basic assumption of signal detection theory, the likelihood ratio decision axis, implies three regularities in recognition memory: (1) the mirror effect, (2) the variance effect, and (3) the z-ROC length effect. For each model, we present the equations that produce the three regularities and show, in computed examples, how they do so. We then show that the regularities appear in data from a range of recognition studies. The analyses and data in our study support the following generalization: Individuals make efficient recognition decisions on the basis of likelihood ratios.

11. Likelihood ratio sequential sampling models of recognition memory.

Science.gov (United States)

Osth, Adam F; Dennis, Simon; Heathcote, Andrew

2017-02-01

12. Understanding the properties of diagnostic tests - Part 2: Likelihood ratios.

Science.gov (United States)

Ranganathan, Priya; Aggarwal, Rakesh

2018-01-01

Diagnostic tests are used to identify subjects with and without disease. In a previous article in this series, we examined some attributes of diagnostic tests - sensitivity, specificity, and predictive values. In this second article, we look at likelihood ratios, which are useful for the interpretation of diagnostic test results in everyday clinical practice.

13. A simplification of the likelihood ratio test statistic for testing ...

African Journals Online (AJOL)

The traditional likelihood ratio test statistic for testing hypothesis about goodness of fit of multinomial probabilities in one, two and multi – dimensional contingency table was simplified. Advantageously, using the simplified version of the statistic to test the null hypothesis is easier and faster because calculating the expected ...

14. Exclusion probabilities and likelihood ratios with applications to mixtures.

Science.gov (United States)

Slooten, Klaas-Jan; Egeland, Thore

2016-01-01

The statistical evidence obtained from mixed DNA profiles can be summarised in several ways in forensic casework including the likelihood ratio (LR) and the Random Man Not Excluded (RMNE) probability. The literature has seen a discussion of the advantages and disadvantages of likelihood ratios and exclusion probabilities, and part of our aim is to bring some clarification to this debate. In a previous paper, we proved that there is a general mathematical relationship between these statistics: RMNE can be expressed as a certain average of the LR, implying that the expected value of the LR, when applied to an actual contributor to the mixture, is at least equal to the inverse of the RMNE. While the mentioned paper presented applications for kinship problems, the current paper demonstrates the relevance for mixture cases, and for this purpose, we prove some new general properties. We also demonstrate how to use the distribution of the likelihood ratio for donors of a mixture, to obtain estimates for exceedance probabilities of the LR for non-donors, of which the RMNE is a special case corresponding to L R>0. In order to derive these results, we need to view the likelihood ratio as a random variable. In this paper, we describe how such a randomization can be achieved. The RMNE is usually invoked only for mixtures without dropout. In mixtures, artefacts like dropout and drop-in are commonly encountered and we address this situation too, illustrating our results with a basic but widely implemented model, a so-called binary model. The precise definitions, modelling and interpretation of the required concepts of dropout and drop-in are not entirely obvious, and we attempt to clarify them here in a general likelihood framework for a binary model.

15. Scale invariant for one-sided multivariate likelihood ratio tests

Directory of Open Access Journals (Sweden)

Samruam Chongcharoen

2010-07-01

Full Text Available Suppose 1 2 , ,..., n X X X is a random sample from Np ( ,V distribution. Consider 0 1 2 : ... 0 p H      and1 : 0 for 1, 2,..., i H   i  p , let 1 0 H  H denote the hypothesis that 1 H holds but 0 H does not, and let ~ 0 H denote thehypothesis that 0 H does not hold. Because the likelihood ratio test (LRT of 0 H versus 1 0 H  H is complicated, severalad hoc tests have been proposed. Tang, Gnecco and Geller (1989 proposed an approximate LRT, Follmann (1996 suggestedrejecting 0 H if the usual test of 0 H versus ~ 0 H rejects 0 H with significance level 2 and a weighted sum of the samplemeans is positive, and Chongcharoen, Singh and Wright (2002 modified Follmann’s test to include information about thecorrelation structure in the sum of the sample means. Chongcharoen and Wright (2007, 2006 give versions of the Tang-Gnecco-Geller tests and Follmann-type tests, respectively, with invariance properties. With LRT’s scale invariant desiredproperty, we investigate its powers by using Monte Carlo techniques and compare them with the tests which we recommendin Chongcharoen and Wright (2007, 2006.

16. Multiple Improvements of Multiple Imputation Likelihood Ratio Tests

OpenAIRE

Chan, Kin Wai; Meng, Xiao-Li

2017-01-01

Multiple imputation (MI) inference handles missing data by first properly imputing the missing values $m$ times, and then combining the $m$ analysis results from applying a complete-data procedure to each of the completed datasets. However, the existing method for combining likelihood ratio tests has multiple defects: (i) the combined test statistic can be negative in practice when the reference null distribution is a standard $F$ distribution; (ii) it is not invariant to re-parametrization; ...

17. Similar tests and the standardized log likelihood ratio statistic

DEFF Research Database (Denmark)

Jensen, Jens Ledet

1986-01-01

When testing an affine hypothesis in an exponential family the 'ideal' procedure is to calculate the exact similar test, or an approximation to this, based on the conditional distribution given the minimal sufficient statistic under the null hypothesis. By contrast to this there is a 'primitive......' approach in which the marginal distribution of a test statistic considered and any nuisance parameter appearing in the test statistic is replaced by an estimate. We show here that when using standardized likelihood ratio statistics the 'primitive' procedure is in fact an 'ideal' procedure to order O(n -3...

18. Transfer Entropy as a Log-Likelihood Ratio

Science.gov (United States)

Barnett, Lionel; Bossomaier, Terry

2012-09-01

Transfer entropy, an information-theoretic measure of time-directed information transfer between joint processes, has steadily gained popularity in the analysis of complex stochastic dynamics in diverse fields, including the neurosciences, ecology, climatology, and econometrics. We show that for a broad class of predictive models, the log-likelihood ratio test statistic for the null hypothesis of zero transfer entropy is a consistent estimator for the transfer entropy itself. For finite Markov chains, furthermore, no explicit model is required. In the general case, an asymptotic χ2 distribution is established for the transfer entropy estimator. The result generalizes the equivalence in the Gaussian case of transfer entropy and Granger causality, a statistical notion of causal influence based on prediction via vector autoregression, and establishes a fundamental connection between directed information transfer and causality in the Wiener-Granger sense.

19. Likelihood ratio model for classification of forensic evidence

Energy Technology Data Exchange (ETDEWEB)

Zadora, G., E-mail: gzadora@ies.krakow.pl [Institute of Forensic Research, Westerplatte 9, 31-033 Krakow (Poland); Neocleous, T., E-mail: tereza@stats.gla.ac.uk [University of Glasgow, Department of Statistics, 15 University Gardens, Glasgow G12 8QW (United Kingdom)

2009-05-29

One of the problems of analysis of forensic evidence such as glass fragments, is the determination of their use-type category, e.g. does a glass fragment originate from an unknown window or container? Very small glass fragments arise during various accidents and criminal offences, and could be carried on the clothes, shoes and hair of participants. It is therefore necessary to obtain information on their physicochemical composition in order to solve the classification problem. Scanning Electron Microscopy coupled with an Energy Dispersive X-ray Spectrometer and the Glass Refractive Index Measurement method are routinely used in many forensic institutes for the investigation of glass. A natural form of glass evidence evaluation for forensic purposes is the likelihood ratio-LR = p(E|H{sub 1})/p(E|H{sub 2}). The main aim of this paper was to study the performance of LR models for glass object classification which considered one or two sources of data variability, i.e. between-glass-object variability and(or) within-glass-object variability. Within the proposed model a multivariate kernel density approach was adopted for modelling the between-object distribution and a multivariate normal distribution was adopted for modelling within-object distributions. Moreover, a graphical method of estimating the dependence structure was employed to reduce the highly multivariate problem to several lower-dimensional problems. The performed analysis showed that the best likelihood model was the one which allows to include information about between and within-object variability, and with variables derived from elemental compositions measured by SEM-EDX, and refractive values determined before (RI{sub b}) and after (RI{sub a}) the annealing process, in the form of dRI = log{sub 10}|RI{sub a} - RI{sub b}|. This model gave better results than the model with only between-object variability considered. In addition, when dRI and variables derived from elemental compositions were used, this

20. Likelihood ratio model for classification of forensic evidence

International Nuclear Information System (INIS)

Zadora, G.; Neocleous, T.

2009-01-01

One of the problems of analysis of forensic evidence such as glass fragments, is the determination of their use-type category, e.g. does a glass fragment originate from an unknown window or container? Very small glass fragments arise during various accidents and criminal offences, and could be carried on the clothes, shoes and hair of participants. It is therefore necessary to obtain information on their physicochemical composition in order to solve the classification problem. Scanning Electron Microscopy coupled with an Energy Dispersive X-ray Spectrometer and the Glass Refractive Index Measurement method are routinely used in many forensic institutes for the investigation of glass. A natural form of glass evidence evaluation for forensic purposes is the likelihood ratio-LR = p(E|H 1 )/p(E|H 2 ). The main aim of this paper was to study the performance of LR models for glass object classification which considered one or two sources of data variability, i.e. between-glass-object variability and(or) within-glass-object variability. Within the proposed model a multivariate kernel density approach was adopted for modelling the between-object distribution and a multivariate normal distribution was adopted for modelling within-object distributions. Moreover, a graphical method of estimating the dependence structure was employed to reduce the highly multivariate problem to several lower-dimensional problems. The performed analysis showed that the best likelihood model was the one which allows to include information about between and within-object variability, and with variables derived from elemental compositions measured by SEM-EDX, and refractive values determined before (RI b ) and after (RI a ) the annealing process, in the form of dRI = log 10 |RI a - RI b |. This model gave better results than the model with only between-object variability considered. In addition, when dRI and variables derived from elemental compositions were used, this model outperformed two other

1. Physician Bayesian updating from personal beliefs about the base rate and likelihood ratio.

Science.gov (United States)

Rottman, Benjamin Margolin

2017-02-01

Whether humans can accurately make decisions in line with Bayes' rule has been one of the most important yet contentious topics in cognitive psychology. Though a number of paradigms have been used for studying Bayesian updating, rarely have subjects been allowed to use their own preexisting beliefs about the prior and the likelihood. A study is reported in which physicians judged the posttest probability of a diagnosis for a patient vignette after receiving a test result, and the physicians' posttest judgments were compared to the normative posttest calculated from their own beliefs in the sensitivity and false positive rate of the test (likelihood ratio) and prior probability of the diagnosis. On the one hand, the posttest judgments were strongly related to the physicians' beliefs about both the prior probability as well as the likelihood ratio, and the priors were used considerably more strongly than in previous research. On the other hand, both the prior and the likelihoods were still not used quite as much as they should have been, and there was evidence of other nonnormative aspects to the updating, such as updating independent of the likelihood beliefs. By focusing on how physicians use their own prior beliefs for Bayesian updating, this study provides insight into how well experts perform probabilistic inference in settings in which they rely upon their own prior beliefs rather than experimenter-provided cues. It suggests that there is reason to be optimistic about experts' abilities, but that there is still considerable need for improvement.

2. Likelihood ratio data to report the validation of a forensic fingerprint evaluation method

NARCIS (Netherlands)

Ramos, Daniel; Haraksim, Rudolf; Meuwly, Didier

2017-01-01

Data to which the authors refer to throughout this article are likelihood ratios (LR) computed from the comparison of 5–12 minutiae fingermarks with fingerprints. These LRs data are used for the validation of a likelihood ratio (LR) method in forensic evidence evaluation. These data present a

3. A guideline for the validation of likelihood ratio methods used for forensic evidence evaluation

NARCIS (Netherlands)

Meuwly, Didier; Ramos, Daniel; Haraksim, Rudolf

2017-01-01

This Guideline proposes a protocol for the validation of forensic evaluation methods at the source level, using the Likelihood Ratio framework as defined within the Bayes’ inference model. In the context of the inference of identity of source, the Likelihood Ratio is used to evaluate the strength of

4. Maximum likelihood positioning algorithm for high-resolution PET scanners

International Nuclear Information System (INIS)

Gross-Weege, Nicolas; Schug, David; Hallen, Patrick; Schulz, Volkmar

2016-01-01

Purpose: In high-resolution positron emission tomography (PET), lightsharing elements are incorporated into typical detector stacks to read out scintillator arrays in which one scintillator element (crystal) is smaller than the size of the readout channel. In order to identify the hit crystal by means of the measured light distribution, a positioning algorithm is required. One commonly applied positioning algorithm uses the center of gravity (COG) of the measured light distribution. The COG algorithm is limited in spatial resolution by noise and intercrystal Compton scatter. The purpose of this work is to develop a positioning algorithm which overcomes this limitation. Methods: The authors present a maximum likelihood (ML) algorithm which compares a set of expected light distributions given by probability density functions (PDFs) with the measured light distribution. Instead of modeling the PDFs by using an analytical model, the PDFs of the proposed ML algorithm are generated assuming a single-gamma-interaction model from measured data. The algorithm was evaluated with a hot-rod phantom measurement acquired with the preclinical HYPERION II D PET scanner. In order to assess the performance with respect to sensitivity, energy resolution, and image quality, the ML algorithm was compared to a COG algorithm which calculates the COG from a restricted set of channels. The authors studied the energy resolution of the ML and the COG algorithm regarding incomplete light distributions (missing channel information caused by detector dead time). Furthermore, the authors investigated the effects of using a filter based on the likelihood values on sensitivity, energy resolution, and image quality. Results: A sensitivity gain of up to 19% was demonstrated in comparison to the COG algorithm for the selected operation parameters. Energy resolution and image quality were on a similar level for both algorithms. Additionally, the authors demonstrated that the performance of the ML

5. Zero-inflated Poisson model based likelihood ratio test for drug safety signal detection.

Science.gov (United States)

Huang, Lan; Zheng, Dan; Zalkikar, Jyoti; Tiwari, Ram

2017-02-01

In recent decades, numerous methods have been developed for data mining of large drug safety databases, such as Food and Drug Administration's (FDA's) Adverse Event Reporting System, where data matrices are formed by drugs such as columns and adverse events as rows. Often, a large number of cells in these data matrices have zero cell counts and some of them are "true zeros" indicating that the drug-adverse event pairs cannot occur, and these zero counts are distinguished from the other zero counts that are modeled zero counts and simply indicate that the drug-adverse event pairs have not occurred yet or have not been reported yet. In this paper, a zero-inflated Poisson model based likelihood ratio test method is proposed to identify drug-adverse event pairs that have disproportionately high reporting rates, which are also called signals. The maximum likelihood estimates of the model parameters of zero-inflated Poisson model based likelihood ratio test are obtained using the expectation and maximization algorithm. The zero-inflated Poisson model based likelihood ratio test is also modified to handle the stratified analyses for binary and categorical covariates (e.g. gender and age) in the data. The proposed zero-inflated Poisson model based likelihood ratio test method is shown to asymptotically control the type I error and false discovery rate, and its finite sample performance for signal detection is evaluated through a simulation study. The simulation results show that the zero-inflated Poisson model based likelihood ratio test method performs similar to Poisson model based likelihood ratio test method when the estimated percentage of true zeros in the database is small. Both the zero-inflated Poisson model based likelihood ratio test and likelihood ratio test methods are applied to six selected drugs, from the 2006 to 2011 Adverse Event Reporting System database, with varying percentages of observed zero-count cells.

6. Sampling variability in forensic likelihood-ratio computation: A simulation study

NARCIS (Netherlands)

Ali, Tauseef; Spreeuwers, Lieuwe Jan; Veldhuis, Raymond N.J.; Meuwly, Didier

2015-01-01

Recently, in the forensic biometric community, there is a growing interest to compute a metric called “likelihood- ratio‿ when a pair of biometric specimens is compared using a biometric recognition system. Generally, a biomet- ric recognition system outputs a score and therefore a likelihood-ratio

7. Performances of the likelihood-ratio classifier based on different data modelings

NARCIS (Netherlands)

Chen, C.; Veldhuis, Raymond N.J.

2008-01-01

The classical likelihood ratio classifier easily collapses in many biometric applications especially with independent training-test subjects. The reason lies in the inaccurate estimation of the underlying user-specific feature density. Firstly, the feature density estimation suffers from

8. Parameter estimation in astronomy through application of the likelihood ratio. [satellite data analysis techniques

Science.gov (United States)

Cash, W.

1979-01-01

Many problems in the experimental estimation of parameters for models can be solved through use of the likelihood ratio test. Applications of the likelihood ratio, with particular attention to photon counting experiments, are discussed. The procedures presented solve a greater range of problems than those currently in use, yet are no more difficult to apply. The procedures are proved analytically, and examples from current problems in astronomy are discussed.

9. Tests and Confidence Intervals for an Extended Variance Component Using the Modified Likelihood Ratio Statistic

DEFF Research Database (Denmark)

Christensen, Ole Fredslund; Frydenberg, Morten; Jensen, Jens Ledet

2005-01-01

The large deviation modified likelihood ratio statistic is studied for testing a variance component equal to a specified value. Formulas are presented in the general balanced case, whereas in the unbalanced case only the one-way random effects model is studied. Simulation studies are presented......, showing that the normal approximation to the large deviation modified likelihood ratio statistic gives confidence intervals for variance components with coverage probabilities very close to the nominal confidence coefficient....

10. The likelihood ratio as a random variable for linked markers in kinship analysis.

Science.gov (United States)

Egeland, Thore; Slooten, Klaas

2016-11-01

The likelihood ratio is the fundamental quantity that summarizes the evidence in forensic cases. Therefore, it is important to understand the theoretical properties of this statistic. This paper is the last in a series of three, and the first to study linked markers. We show that for all non-inbred pairwise kinship comparisons, the expected likelihood ratio in favor of a type of relatedness depends on the allele frequencies only via the number of alleles, also for linked markers, and also if the true relationship is another one than is tested for by the likelihood ratio. Exact expressions for the expectation and variance are derived for all these cases. Furthermore, we show that the expected likelihood ratio is a non-increasing function if the recombination rate increases between 0 and 0.5 when the actual relationship is the one investigated by the LR. Besides being of theoretical interest, exact expressions such as obtained here can be used for software validation as they allow to verify the correctness up to arbitrary precision. The paper also presents results and advice of practical importance. For example, we argue that the logarithm of the likelihood ratio behaves in a fundamentally different way than the likelihood ratio itself in terms of expectation and variance, in agreement with its interpretation as weight of evidence. Equipped with the results presented and freely available software, one may check calculations and software and also do power calculations.

11. Statistical analysis of COMPTEL maximum likelihood-ratio distributions: evidence for a signal from previously undetected AGN

International Nuclear Information System (INIS)

Williams, O. R.; Bennett, K.; Much, R.; Schoenfelder, V.; Blom, J. J.; Ryan, J.

1997-01-01

The maximum likelihood-ratio method is frequently used in COMPTEL analysis to determine the significance of a point source at a given location. In this paper we do not consider whether the likelihood-ratio at a particular location indicates a detection, but rather whether distributions of likelihood-ratios derived from many locations depart from that expected for source free data. We have constructed distributions of likelihood-ratios by reading values from standard COMPTEL maximum-likelihood ratio maps at positions corresponding to the locations of different categories of AGN. Distributions derived from the locations of Seyfert galaxies are indistinguishable, according to a Kolmogorov-Smirnov test, from those obtained from ''random'' locations, but differ slightly from those obtained from the locations of flat spectrum radio loud quasars, OVVs, and BL Lac objects. This difference is not due to known COMPTEL sources, since regions near these sources are excluded from the analysis. We suggest that it might arise from a number of sources with fluxes below the COMPTEL detection threshold

12. Uncertainty about the true source. A note on the likelihood ratio at the activity level.

Science.gov (United States)

Taroni, Franco; Biedermann, Alex; Bozza, Silvia; Comte, Jennifer; Garbolino, Paolo

2012-07-10

This paper focuses on likelihood ratio based evaluations of fibre evidence in cases in which there is uncertainty about whether or not the reference item available for analysis - that is, an item typically taken from the suspect or seized at his home - is the item actually worn at the time of the offence. A likelihood ratio approach is proposed that, for situations in which certain categorical assumptions can be made about additionally introduced parameters, converges to formula described in existing literature. The properties of the proposed likelihood ratio approach are analysed through sensitivity analyses and discussed with respect to possible argumentative implications that arise in practice. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

13. Derivation of LDA log likelihood ratio one-to-one classifier

NARCIS (Netherlands)

Spreeuwers, Lieuwe Jan

2014-01-01

The common expression for the Likelihood Ratio classifier using LDA assumes that the reference class mean is available. In biometrics, this is often not the case and only a single sample of the reference class is available. In this paper expressions are derived for biometric comparison between

14. Comparison of IRT Likelihood Ratio Test and Logistic Regression DIF Detection Procedures

Science.gov (United States)

Atar, Burcu; Kamata, Akihito

2011-01-01

The Type I error rates and the power of IRT likelihood ratio test and cumulative logit ordinal logistic regression procedures in detecting differential item functioning (DIF) for polytomously scored items were investigated in this Monte Carlo simulation study. For this purpose, 54 simulation conditions (combinations of 3 sample sizes, 2 sample…

15. Robust Biometric Score Fusion by Naive Likelihood Ratio via Receiver Operating Characteristics

NARCIS (Netherlands)

Tao, Q.; Veldhuis, Raymond N.J.

This paper presents a novel method of fusing multiple biometrics on the matching score level. We estimate the likelihood ratios of the fused biometric scores, via individual receiver operating characteristics (ROC) which construct the Naive Bayes classifier. Using a limited number of operation

16. Validation of software for calculating the likelihood ratio for parentage and kinship.

Science.gov (United States)

Drábek, J

2009-03-01

Although the likelihood ratio is a well-known statistical technique, commercial off-the-shelf (COTS) software products for its calculation are not sufficiently validated to suit general requirements for the competence of testing and calibration laboratories (EN/ISO/IEC 17025:2005 norm) per se. The software in question can be considered critical as it directly weighs the forensic evidence allowing judges to decide on guilt or innocence or to identify person or kin (i.e.: in mass fatalities). For these reasons, accredited laboratories shall validate likelihood ratio software in accordance with the above norm. To validate software for calculating the likelihood ratio in parentage/kinship scenarios I assessed available vendors, chose two programs (Paternity Index and familias) for testing, and finally validated them using tests derived from elaboration of the available guidelines for the field of forensics, biomedicine, and software engineering. MS Excel calculation using known likelihood ratio formulas or peer-reviewed results of difficult paternity cases were used as a reference. Using seven testing cases, it was found that both programs satisfied the requirements for basic paternity cases. However, only a combination of two software programs fulfills the criteria needed for our purpose in the whole spectrum of functions under validation with the exceptions of providing algebraic formulas in cases of mutation and/or silent allele.

17. Validation of Likelihood Ratio Methods Used for Forensic Evidence Evaluation: Application in Forensic Fingerprints

NARCIS (Netherlands)

Haraksim, Rudolf

2014-01-01

In this chapter the Likelihood Ratio (LR) inference model will be introduced, the theoretical aspects of probabilities will be discussed and the validation framework for LR methods used for forensic evidence evaluation will be presented. Prior to introducing the validation framework, following

18. Validation of DNA-based identification software by computation of pedigree likelihood ratios.

Science.gov (United States)

Slooten, K

2011-08-01

Disaster victim identification (DVI) can be aided by DNA-evidence, by comparing the DNA-profiles of unidentified individuals with those of surviving relatives. The DNA-evidence is used optimally when such a comparison is done by calculating the appropriate likelihood ratios. Though conceptually simple, the calculations can be quite involved, especially with large pedigrees, precise mutation models etc. In this article we describe a series of test cases designed to check if software designed to calculate such likelihood ratios computes them correctly. The cases include both simple and more complicated pedigrees, among which inbred ones. We show how to calculate the likelihood ratio numerically and algebraically, including a general mutation model and possibility of allelic dropout. In Appendix A we show how to derive such algebraic expressions mathematically. We have set up these cases to validate new software, called Bonaparte, which performs pedigree likelihood ratio calculations in a DVI context. Bonaparte has been developed by SNN Nijmegen (The Netherlands) for the Netherlands Forensic Institute (NFI). It is available free of charge for non-commercial purposes (see www.dnadvi.nl for details). Commercial licenses can also be obtained. The software uses Bayesian networks and the junction tree algorithm to perform its calculations. Copyright © 2010 Elsevier Ireland Ltd. All rights reserved.

19. Validation of DNA-based identification software by computation of pedigree likelihood ratios

NARCIS (Netherlands)

Slooten, K.

Disaster victim identification (DVI) can be aided by DNA-evidence, by comparing the DNA-profiles of unidentified individuals with those of surviving relatives. The DNA-evidence is used optimally when such a comparison is done by calculating the appropriate likelihood ratios. Though conceptually

20. Inference for the Sharpe Ratio Using a Likelihood-Based Approach

Directory of Open Access Journals (Sweden)

Ying Liu

2012-01-01

Full Text Available The Sharpe ratio is the prominent risk-adjusted performance measure used by practitioners. Statistical testing of this ratio using its asymptotic distribution has lagged behind its use. In this paper, highly accurate likelihood analysis is applied for inference on the Sharpe ratio. Both the one- and two-sample problems are considered. The methodology has O(n−3/2 distributional accuracy and can be implemented using any parametric return distribution structure. Simulations are provided to demonstrate the method's superior accuracy over existing methods used for testing in the literature.

1. Adjusted Empirical Likelihood Method in the Presence of Nuisance Parameters with Application to the Sharpe Ratio

Directory of Open Access Journals (Sweden)

Yuejiao Fu

2018-04-01

Full Text Available The Sharpe ratio is a widely used risk-adjusted performance measurement in economics and finance. Most of the known statistical inferential methods devoted to the Sharpe ratio are based on the assumption that the data are normally distributed. In this article, without making any distributional assumption on the data, we develop the adjusted empirical likelihood method to obtain inference for a parameter of interest in the presence of nuisance parameters. We show that the log adjusted empirical likelihood ratio statistic is asymptotically distributed as the chi-square distribution. The proposed method is applied to obtain inference for the Sharpe ratio. Simulation results illustrate that the proposed method is comparable to Jobson and Korkie’s method (1981 and outperforms the empirical likelihood method when the data are from a symmetric distribution. In addition, when the data are from a skewed distribution, the proposed method significantly outperforms all other existing methods. A real-data example is analyzed to exemplify the application of the proposed method.

2. A scaling transformation for classifier output based on likelihood ratio: Applications to a CAD workstation for diagnosis of breast cancer

International Nuclear Information System (INIS)

Horsch, Karla; Pesce, Lorenzo L.; Giger, Maryellen L.; Metz, Charles E.; Jiang Yulei

2012-01-01

Purpose: The authors developed scaling methods that monotonically transform the output of one classifier to the ''scale'' of another. Such transformations affect the distribution of classifier output while leaving the ROC curve unchanged. In particular, they investigated transformations between radiologists and computer classifiers, with the goal of addressing the problem of comparing and interpreting case-specific values of output from two classifiers. Methods: Using both simulated and radiologists' rating data of breast imaging cases, the authors investigated a likelihood-ratio-scaling transformation, based on ''matching'' classifier likelihood ratios. For comparison, three other scaling transformations were investigated that were based on matching classifier true positive fraction, false positive fraction, or cumulative distribution function, respectively. The authors explored modifying the computer output to reflect the scale of the radiologist, as well as modifying the radiologist's ratings to reflect the scale of the computer. They also evaluated how dataset size affects the transformations. Results: When ROC curves of two classifiers differed substantially, the four transformations were found to be quite different. The likelihood-ratio scaling transformation was found to vary widely from radiologist to radiologist. Similar results were found for the other transformations. Our simulations explored the effect of database sizes on the accuracy of the estimation of our scaling transformations. Conclusions: The likelihood-ratio-scaling transformation that the authors have developed and evaluated was shown to be capable of transforming computer and radiologist outputs to a common scale reliably, thereby allowing the comparison of the computer and radiologist outputs on the basis of a clinically relevant statistic.

3. Efficient estimators for likelihood ratio sensitivity indices of complex stochastic dynamics

Energy Technology Data Exchange (ETDEWEB)

Arampatzis, Georgios; Katsoulakis, Markos A.; Rey-Bellet, Luc [Department of Mathematics and Statistics, University of Massachusetts, Amherst, Massachusetts 01003 (United States)

2016-03-14

We demonstrate that centered likelihood ratio estimators for the sensitivity indices of complex stochastic dynamics are highly efficient with low, constant in time variance and consequently they are suitable for sensitivity analysis in long-time and steady-state regimes. These estimators rely on a new covariance formulation of the likelihood ratio that includes as a submatrix a Fisher information matrix for stochastic dynamics and can also be used for fast screening of insensitive parameters and parameter combinations. The proposed methods are applicable to broad classes of stochastic dynamics such as chemical reaction networks, Langevin-type equations and stochastic models in finance, including systems with a high dimensional parameter space and/or disparate decorrelation times between different observables. Furthermore, they are simple to implement as a standard observable in any existing simulation algorithm without additional modifications.

4. Multimodal Personal Verification Using Likelihood Ratio for the Match Score Fusion

Directory of Open Access Journals (Sweden)

Long Binh Tran

2017-01-01

Full Text Available In this paper, the authors present a novel personal verification system based on the likelihood ratio test for fusion of match scores from multiple biometric matchers (face, fingerprint, hand shape, and palm print. In the proposed system, multimodal features are extracted by Zernike Moment (ZM. After matching, the match scores from multiple biometric matchers are fused based on the likelihood ratio test. A finite Gaussian mixture model (GMM is used for estimating the genuine and impostor densities of match scores for personal verification. Our approach is also compared to some different famous approaches such as the support vector machine and the sum rule with min-max. The experimental results have confirmed that the proposed system can achieve excellent identification performance for its higher level in accuracy than different famous approaches and thus can be utilized for more application related to person verification.

5. Diagonal Likelihood Ratio Test for Equality of Mean Vectors in High-Dimensional Data

KAUST Repository

Hu, Zongliang; Tong, Tiejun; Genton, Marc G.

2017-01-01

We propose a likelihood ratio test framework for testing normal mean vectors in high-dimensional data under two common scenarios: the one-sample test and the two-sample test with equal covariance matrices. We derive the test statistics under the assumption that the covariance matrices follow a diagonal matrix structure. In comparison with the diagonal Hotelling's tests, our proposed test statistics display some interesting characteristics. In particular, they are a summation of the log-transformed squared t-statistics rather than a direct summation of those components. More importantly, to derive the asymptotic normality of our test statistics under the null and local alternative hypotheses, we do not require the assumption that the covariance matrix follows a diagonal matrix structure. As a consequence, our proposed test methods are very flexible and can be widely applied in practice. Finally, simulation studies and a real data analysis are also conducted to demonstrate the advantages of our likelihood ratio test method.

6. On-line validation of linear process models using generalized likelihood ratios

International Nuclear Information System (INIS)

Tylee, J.L.

1981-12-01

A real-time method for testing the validity of linear models of nonlinear processes is described and evaluated. Using generalized likelihood ratios, the model dynamics are continually monitored to see if the process has moved far enough away from the nominal linear model operating point to justify generation of a new linear model. The method is demonstrated using a seventh-order model of a natural circulation steam generator

7. A likelihood ratio test for species membership based on DNA sequence data

DEFF Research Database (Denmark)

Matz, Mikhail V.; Nielsen, Rasmus

2005-01-01

DNA barcoding as an approach for species identification is rapidly increasing in popularity. However, it remains unclear which statistical procedures should accompany the technique to provide a measure of uncertainty. Here we describe a likelihood ratio test which can be used to test if a sampled...... sequence is a member of an a priori specified species. We investigate the performance of the test using coalescence simulations, as well as using the real data from butterflies and frogs representing two kinds of challenge for DNA barcoding: extremely low and extremely high levels of sequence variability....

8. Analytic confidence level calculations using the likelihood ratio and fourier transform

International Nuclear Information System (INIS)

Hu Hongbo; Nielsen, J.

2000-01-01

The interpretation of new particle search results involves a confidence level calculation on either the discovery hypothesis or the background-only ('null') hypothesis. A typical approach uses toy Monte Carlo experiments to build an expected experiment estimator distribution against which an observed experiment's estimator may be compared. In this note, a new approach is presented which calculates analytically the experiment estimator distribution via a Fourier transform, using the likelihood ratio as an ordering estimator. The analytic approach enjoys an enormous speed advantage over the toy Monte Carlo method, making it possible to quickly and precisely calculate confidence level results

9. Diagonal Likelihood Ratio Test for Equality of Mean Vectors in High-Dimensional Data

KAUST Repository

Hu, Zongliang

2017-10-27

We propose a likelihood ratio test framework for testing normal mean vectors in high-dimensional data under two common scenarios: the one-sample test and the two-sample test with equal covariance matrices. We derive the test statistics under the assumption that the covariance matrices follow a diagonal matrix structure. In comparison with the diagonal Hotelling\\'s tests, our proposed test statistics display some interesting characteristics. In particular, they are a summation of the log-transformed squared t-statistics rather than a direct summation of those components. More importantly, to derive the asymptotic normality of our test statistics under the null and local alternative hypotheses, we do not require the assumption that the covariance matrix follows a diagonal matrix structure. As a consequence, our proposed test methods are very flexible and can be widely applied in practice. Finally, simulation studies and a real data analysis are also conducted to demonstrate the advantages of our likelihood ratio test method.

10. Steady state likelihood ratio sensitivity analysis for stiff kinetic Monte Carlo simulations.

Science.gov (United States)

Núñez, M; Vlachos, D G

2015-01-28

Kinetic Monte Carlo simulation is an integral tool in the study of complex physical phenomena present in applications ranging from heterogeneous catalysis to biological systems to crystal growth and atmospheric sciences. Sensitivity analysis is useful for identifying important parameters and rate-determining steps, but the finite-difference application of sensitivity analysis is computationally demanding. Techniques based on the likelihood ratio method reduce the computational cost of sensitivity analysis by obtaining all gradient information in a single run. However, we show that disparity in time scales of microscopic events, which is ubiquitous in real systems, introduces drastic statistical noise into derivative estimates for parameters affecting the fast events. In this work, the steady-state likelihood ratio sensitivity analysis is extended to singularly perturbed systems by invoking partial equilibration for fast reactions, that is, by working on the fast and slow manifolds of the chemistry. Derivatives on each time scale are computed independently and combined to the desired sensitivity coefficients to considerably reduce the noise in derivative estimates for stiff systems. The approach is demonstrated in an analytically solvable linear system.

11. Lay understanding of forensic statistics: Evaluation of random match probabilities, likelihood ratios, and verbal equivalents.

Science.gov (United States)

Thompson, William C; Newman, Eryn J

2015-08-01

Forensic scientists have come under increasing pressure to quantify the strength of their evidence, but it is not clear which of several possible formats for presenting quantitative conclusions will be easiest for lay people, such as jurors, to understand. This experiment examined the way that people recruited from Amazon's Mechanical Turk (n = 541) responded to 2 types of forensic evidence--a DNA comparison and a shoeprint comparison--when an expert explained the strength of this evidence 3 different ways: using random match probabilities (RMPs), likelihood ratios (LRs), or verbal equivalents of likelihood ratios (VEs). We found that verdicts were sensitive to the strength of DNA evidence regardless of how the expert explained it, but verdicts were sensitive to the strength of shoeprint evidence only when the expert used RMPs. The weight given to DNA evidence was consistent with the predictions of a Bayesian network model that incorporated the perceived risk of a false match from 3 causes (coincidence, a laboratory error, and a frame-up), but shoeprint evidence was undervalued relative to the same Bayesian model. Fallacious interpretations of the expert's testimony (consistent with the source probability error and the defense attorney's fallacy) were common and were associated with the weight given to the evidence and verdicts. The findings indicate that perceptions of forensic science evidence are shaped by prior beliefs and expectations as well as expert testimony and consequently that the best way to characterize and explain forensic evidence may vary across forensic disciplines. (c) 2015 APA, all rights reserved).

12. Investigation of various factors influencing Raman spectra interpretation with the use of likelihood ratio approach.

Science.gov (United States)

Michalska, Aleksandra; Martyna, Agnieszka; Zadora, Grzegorz

2018-01-01

The main aim of this study was to verify whether selected analytical parameters may affect solving the comparison problem of Raman spectra with the use of the likelihood ratio (LR) approach. Firstly the LR methodologies developed for Raman spectra of blue automotive paints obtained with the use of 785nm laser source (results published by the authors previously) were implemented for good quality spectra recorded for these paints with the use of 514.5nm laser source. For LR models construction two types of variables were used i.e. areas under selected pigments bands and coefficients derived from discrete wavelet transform procedure (DWT). Few experiments were designed for 785nm and 514.5nm Raman spectra databases after constructing well performing LR models (low rates of false positive and false negative answers and acceptable results of empirical cross entropy approach). In order to verify whether objective magnification described by its numerical aperture affects spectra interpretation, three objective magnifications -20×(N.A.=0.4.), 50×(N.A.=0.75) and 100×(N.A.=0.85) within each of the applied laser sources (514.5nm and 785nm) were tested for a group of blue solid and metallic automotive paints having the same sets of pigments depending on the applied laser source. The findings obtained by two types of LR models indicate the importance of this parameter for solving the comparison problem of both solid and metallic automotive paints regardless of the laser source used for measuring Raman signal. Hence, the same objective magnification, preferably 50× (established based on the analysis of within- and between-samples variability and F-factor value), should be used when focusing the laser on samples during Raman measurements. Then the influence of parameters (laser power and time of irradiation) of one of the recommended fluorescence suppression techniques, namely photobleaching, was under investigation. Analysis performed on a group of solid automotive paint

13. Estimating negative likelihood ratio confidence when test sensitivity is 100%: A bootstrapping approach.

Science.gov (United States)

Marill, Keith A; Chang, Yuchiao; Wong, Kim F; Friedman, Ari B

2017-08-01

Objectives Assessing high-sensitivity tests for mortal illness is crucial in emergency and critical care medicine. Estimating the 95% confidence interval (CI) of the likelihood ratio (LR) can be challenging when sample sensitivity is 100%. We aimed to develop, compare, and automate a bootstrapping method to estimate the negative LR CI when sample sensitivity is 100%. Methods The lowest population sensitivity that is most likely to yield sample sensitivity 100% is located using the binomial distribution. Random binomial samples generated using this population sensitivity are then used in the LR bootstrap. A free R program, "bootLR," automates the process. Extensive simulations were performed to determine how often the LR bootstrap and comparator method 95% CIs cover the true population negative LR value. Finally, the 95% CI was compared for theoretical sample sizes and sensitivities approaching and including 100% using: (1) a technique of individual extremes, (2) SAS software based on the technique of Gart and Nam, (3) the Score CI (as implemented in the StatXact, SAS, and R PropCI package), and (4) the bootstrapping technique. Results The bootstrapping approach demonstrates appropriate coverage of the nominal 95% CI over a spectrum of populations and sample sizes. Considering a study of sample size 200 with 100 patients with disease, and specificity 60%, the lowest population sensitivity with median sample sensitivity 100% is 99.31%. When all 100 patients with disease test positive, the negative LR 95% CIs are: individual extremes technique (0,0.073), StatXact (0,0.064), SAS Score method (0,0.057), R PropCI (0,0.062), and bootstrap (0,0.048). Similar trends were observed for other sample sizes. Conclusions When study samples demonstrate 100% sensitivity, available methods may yield inappropriately wide negative LR CIs. An alternative bootstrapping approach and accompanying free open-source R package were developed to yield realistic estimates easily. This

14. Maximum likelihood positioning for gamma-ray imaging detectors with depth of interaction measurement

International Nuclear Information System (INIS)

Lerche, Ch.W.; Ros, A.; Monzo, J.M.; Aliaga, R.J.; Ferrando, N.; Martinez, J.D.; Herrero, V.; Esteve, R.; Gadea, R.; Colom, R.J.; Toledo, J.; Mateo, F.; Sebastia, A.; Sanchez, F.; Benlloch, J.M.

2009-01-01

The center of gravity algorithm leads to strong artifacts for gamma-ray imaging detectors that are based on monolithic scintillation crystals and position sensitive photo-detectors. This is a consequence of using the centroids as position estimates. The fact that charge division circuits can also be used to compute the standard deviation of the scintillation light distribution opens a way out of this drawback. We studied the feasibility of maximum likelihood estimation for computing the true gamma-ray photo-conversion position from the centroids and the standard deviation of the light distribution. The method was evaluated on a test detector that consists of the position sensitive photomultiplier tube H8500 and a monolithic LSO crystal (42mmx42mmx10mm). Spatial resolution was measured for the centroids and the maximum likelihood estimates. The results suggest that the maximum likelihood positioning is feasible and partially removes the strong artifacts of the center of gravity algorithm.

15. Maximum likelihood positioning for gamma-ray imaging detectors with depth of interaction measurement

Energy Technology Data Exchange (ETDEWEB)

Lerche, Ch.W. [Grupo de Sistemas Digitales, ITACA, Universidad Politecnica de Valencia, 46022 Valencia (Spain)], E-mail: lerche@ific.uv.es; Ros, A. [Grupo de Fisica Medica Nuclear, IFIC, Universidad de Valencia-Consejo Superior de Investigaciones Cientificas, 46980 Paterna (Spain); Monzo, J.M.; Aliaga, R.J.; Ferrando, N.; Martinez, J.D.; Herrero, V.; Esteve, R.; Gadea, R.; Colom, R.J.; Toledo, J.; Mateo, F.; Sebastia, A. [Grupo de Sistemas Digitales, ITACA, Universidad Politecnica de Valencia, 46022 Valencia (Spain); Sanchez, F.; Benlloch, J.M. [Grupo de Fisica Medica Nuclear, IFIC, Universidad de Valencia-Consejo Superior de Investigaciones Cientificas, 46980 Paterna (Spain)

2009-06-01

The center of gravity algorithm leads to strong artifacts for gamma-ray imaging detectors that are based on monolithic scintillation crystals and position sensitive photo-detectors. This is a consequence of using the centroids as position estimates. The fact that charge division circuits can also be used to compute the standard deviation of the scintillation light distribution opens a way out of this drawback. We studied the feasibility of maximum likelihood estimation for computing the true gamma-ray photo-conversion position from the centroids and the standard deviation of the light distribution. The method was evaluated on a test detector that consists of the position sensitive photomultiplier tube H8500 and a monolithic LSO crystal (42mmx42mmx10mm). Spatial resolution was measured for the centroids and the maximum likelihood estimates. The results suggest that the maximum likelihood positioning is feasible and partially removes the strong artifacts of the center of gravity algorithm.

16. Forensic Automatic Speaker Recognition Based on Likelihood Ratio Using Acoustic-phonetic Features Measured Automatically

Directory of Open Access Journals (Sweden)

Huapeng Wang

2015-01-01

Full Text Available Forensic speaker recognition is experiencing a remarkable paradigm shift in terms of the evaluation framework and presentation of voice evidence. This paper proposes a new method of forensic automatic speaker recognition using the likelihood ratio framework to quantify the strength of voice evidence. The proposed method uses a reference database to calculate the within- and between-speaker variability. Some acoustic-phonetic features are extracted automatically using the software VoiceSauce. The effectiveness of the approach was tested using two Mandarin databases: A mobile telephone database and a landline database. The experiment's results indicate that these acoustic-phonetic features do have some discriminating potential and are worth trying in discrimination. The automatic acoustic-phonetic features have acceptable discriminative performance and can provide more reliable results in evidence analysis when fused with other kind of voice features.

17. Searching for degenerate Higgs bosons using a profile likelihood ratio method

CERN Document Server

Heikkilä, Jaana

ATLAS and CMS collaborations at the Large Hadron Collider have observed a new resonance con- sistent with the standard model Higgs boson. However, it has been suggested that the observed signal could also be produced by multiple nearly mass-degenerate states that couple differently to the standard model particles. In this work, a method to discriminate between the hypothesis of a single Higgs boson and that of multiple mass-degenerate Higgs bosons was developed. Using the matrix of measured signal strengths in different production and decay modes, parametrizations for the two hypotheses were constructed as a general rank 1 matrix and the most general $5 \\times 4$ matrix, respectively. The test statistic was defined as a ratio of profile likelihoods for the two hypotheses. The method was applied to the CMS measurements. The expected test statistic distribution was estimated twice by generating pseudo-experiments according to both the standard model hypothesis and the single Higgs boson hypothesis best fitting...

18. Orthogonal series generalized likelihood ratio test for failure detection and isolation. [for aircraft control

Science.gov (United States)

Hall, Steven R.; Walker, Bruce K.

1990-01-01

A new failure detection and isolation algorithm for linear dynamic systems is presented. This algorithm, the Orthogonal Series Generalized Likelihood Ratio (OSGLR) test, is based on the assumption that the failure modes of interest can be represented by truncated series expansions. This assumption leads to a failure detection algorithm with several desirable properties. Computer simulation results are presented for the detection of the failures of actuators and sensors of a C-130 aircraft. The results show that the OSGLR test generally performs as well as the GLR test in terms of time to detect a failure and is more robust to failure mode uncertainty. However, the OSGLR test is also somewhat more sensitive to modeling errors than the GLR test.

19. Improved anomaly detection using multi-scale PLS and generalized likelihood ratio test

KAUST Repository

2017-02-16

Process monitoring has a central role in the process industry to enhance productivity, efficiency, and safety, and to avoid expensive maintenance. In this paper, a statistical approach that exploit the advantages of multiscale PLS models (MSPLS) and those of a generalized likelihood ratio (GLR) test to better detect anomalies is proposed. Specifically, to consider the multivariate and multi-scale nature of process dynamics, a MSPLS algorithm combining PLS and wavelet analysis is used as modeling framework. Then, GLR hypothesis testing is applied using the uncorrelated residuals obtained from MSPLS model to improve the anomaly detection abilities of these latent variable based fault detection methods even further. Applications to a simulated distillation column data are used to evaluate the proposed MSPLS-GLR algorithm.

20. Improved anomaly detection using multi-scale PLS and generalized likelihood ratio test

KAUST Repository

Madakyaru, Muddu; Harrou, Fouzi; Sun, Ying

2017-01-01

Process monitoring has a central role in the process industry to enhance productivity, efficiency, and safety, and to avoid expensive maintenance. In this paper, a statistical approach that exploit the advantages of multiscale PLS models (MSPLS) and those of a generalized likelihood ratio (GLR) test to better detect anomalies is proposed. Specifically, to consider the multivariate and multi-scale nature of process dynamics, a MSPLS algorithm combining PLS and wavelet analysis is used as modeling framework. Then, GLR hypothesis testing is applied using the uncorrelated residuals obtained from MSPLS model to improve the anomaly detection abilities of these latent variable based fault detection methods even further. Applications to a simulated distillation column data are used to evaluate the proposed MSPLS-GLR algorithm.

1. Likelihood ratio meta-analysis: New motivation and approach for an old method.

Science.gov (United States)

Dormuth, Colin R; Filion, Kristian B; Platt, Robert W

2016-03-01

A 95% confidence interval (CI) in an updated meta-analysis may not have the expected 95% coverage. If a meta-analysis is simply updated with additional data, then the resulting 95% CI will be wrong because it will not have accounted for the fact that the earlier meta-analysis failed or succeeded to exclude the null. This situation can be avoided by using the likelihood ratio (LR) as a measure of evidence that does not depend on type-1 error. We show how an LR-based approach, first advanced by Goodman, can be used in a meta-analysis to pool data from separate studies to quantitatively assess where the total evidence points. The method works by estimating the log-likelihood ratio (LogLR) function from each study. Those functions are then summed to obtain a combined function, which is then used to retrieve the total effect estimate, and a corresponding 'intrinsic' confidence interval. Using as illustrations the CAPRIE trial of clopidogrel versus aspirin in the prevention of ischemic events, and our own meta-analysis of higher potency statins and the risk of acute kidney injury, we show that the LR-based method yields the same point estimate as the traditional analysis, but with an intrinsic confidence interval that is appropriately wider than the traditional 95% CI. The LR-based method can be used to conduct both fixed effect and random effects meta-analyses, it can be applied to old and new meta-analyses alike, and results can be presented in a format that is familiar to a meta-analytic audience. Copyright © 2016 Elsevier Inc. All rights reserved.

2. Likelihood ratio data to report the validation of a forensic fingerprint evaluation method

Directory of Open Access Journals (Sweden)

Daniel Ramos

2017-02-01

Full Text Available Data to which the authors refer to throughout this article are likelihood ratios (LR computed from the comparison of 5–12 minutiae fingermarks with fingerprints. These LRs data are used for the validation of a likelihood ratio (LR method in forensic evidence evaluation. These data present a necessary asset for conducting validation experiments when validating LR methods used in forensic evidence evaluation and set up validation reports. These data can be also used as a baseline for comparing the fingermark evidence in the same minutiae configuration as presented in (D. Meuwly, D. Ramos, R. Haraksim, [1], although the reader should keep in mind that different feature extraction algorithms and different AFIS systems used may produce different LRs values. Moreover, these data may serve as a reproducibility exercise, in order to train the generation of validation reports of forensic methods, according to [1]. Alongside the data, a justification and motivation for the use of methods is given. These methods calculate LRs from the fingerprint/mark data and are subject to a validation procedure. The choice of using real forensic fingerprint in the validation and simulated data in the development is described and justified. Validation criteria are set for the purpose of validation of the LR methods, which are used to calculate the LR values from the data and the validation report. For privacy and data protection reasons, the original fingerprint/mark images cannot be shared. But these images do not constitute the core data for the validation, contrarily to the LRs that are shared.

3. A guideline for the validation of likelihood ratio methods used for forensic evidence evaluation.

Science.gov (United States)

Meuwly, Didier; Ramos, Daniel; Haraksim, Rudolf

2017-07-01

This Guideline proposes a protocol for the validation of forensic evaluation methods at the source level, using the Likelihood Ratio framework as defined within the Bayes' inference model. In the context of the inference of identity of source, the Likelihood Ratio is used to evaluate the strength of the evidence for a trace specimen, e.g. a fingermark, and a reference specimen, e.g. a fingerprint, to originate from common or different sources. Some theoretical aspects of probabilities necessary for this Guideline were discussed prior to its elaboration, which started after a workshop of forensic researchers and practitioners involved in this topic. In the workshop, the following questions were addressed: "which aspects of a forensic evaluation scenario need to be validated?", "what is the role of the LR as part of a decision process?" and "how to deal with uncertainty in the LR calculation?". The questions: "what to validate?" focuses on the validation methods and criteria and "how to validate?" deals with the implementation of the validation protocol. Answers to these questions were deemed necessary with several objectives. First, concepts typical for validation standards [1], such as performance characteristics, performance metrics and validation criteria, will be adapted or applied by analogy to the LR framework. Second, a validation strategy will be defined. Third, validation methods will be described. Finally, a validation protocol and an example of validation report will be proposed, which can be applied to the forensic fields developing and validating LR methods for the evaluation of the strength of evidence at source level under the following propositions. Copyright © 2016. Published by Elsevier B.V.

4. The reaction index and positivity ratio revisited

DEFF Research Database (Denmark)

Andersen, Klaus Ejner; Andersen, Flemming

2008-01-01

BACKGROUND AND OBJECTIVES: Assessing the quality of patch test preparations continues to be a challenge. 2 parameters, the reaction index (RI) and positivity ratio (PR), have been proposed as quality indicators by the Information Network of Departments of Dermatology (IVDK). The value of these st......BACKGROUND AND OBJECTIVES: Assessing the quality of patch test preparations continues to be a challenge. 2 parameters, the reaction index (RI) and positivity ratio (PR), have been proposed as quality indicators by the Information Network of Departments of Dermatology (IVDK). The value...

5. A Space Object Detection Algorithm using Fourier Domain Likelihood Ratio Test

Science.gov (United States)

Becker, D.; Cain, S.

Space object detection is of great importance in the highly dependent yet competitive and congested space domain. Detection algorithms employed play a crucial role in fulfilling the detection component in the situational awareness mission to detect, track, characterize and catalog unknown space objects. Many current space detection algorithms use a matched filter or a spatial correlator to make a detection decision at a single pixel point of a spatial image based on the assumption that the data follows a Gaussian distribution. This paper explores the potential for detection performance advantages when operating in the Fourier domain of long exposure images of small and/or dim space objects from ground based telescopes. A binary hypothesis test is developed based on the joint probability distribution function of the image under the hypothesis that an object is present and under the hypothesis that the image only contains background noise. The detection algorithm tests each pixel point of the Fourier transformed images to make the determination if an object is present based on the criteria threshold found in the likelihood ratio test. Using simulated data, the performance of the Fourier domain detection algorithm is compared to the current algorithm used in space situational awareness applications to evaluate its value.

6. Aircraft control surface failure detection and isolation using the OSGLR test. [orthogonal series generalized likelihood ratio

Science.gov (United States)

Bonnice, W. F.; Motyka, P.; Wagner, E.; Hall, S. R.

1986-01-01

The performance of the orthogonal series generalized likelihood ratio (OSGLR) test in detecting and isolating commercial aircraft control surface and actuator failures is evaluated. A modification to incorporate age-weighting which significantly reduces the sensitivity of the algorithm to modeling errors is presented. The steady-state implementation of the algorithm based on a single linear model valid for a cruise flight condition is tested using a nonlinear aircraft simulation. A number of off-nominal no-failure flight conditions including maneuvers, nonzero flap deflections, different turbulence levels and steady winds were tested. Based on the no-failure decision functions produced by off-nominal flight conditions, the failure detection and isolation performance at the nominal flight condition was determined. The extension of the algorithm to a wider flight envelope by scheduling on dynamic pressure and flap deflection is examined. Based on this testing, the OSGLR algorithm should be capable of detecting control surface failures that would affect the safe operation of a commercial aircraft. Isolation may be difficult if there are several surfaces which produce similar effects on the aircraft. Extending the algorithm over the entire operating envelope of a commercial aircraft appears feasible.

7. Near-exact distributions for the block equicorrelation and equivariance likelihood ratio test statistic

Science.gov (United States)

Coelho, Carlos A.; Marques, Filipe J.

2013-09-01

In this paper the authors combine the equicorrelation and equivariance test introduced by Wilks [13] with the likelihood ratio test (l.r.t.) for independence of groups of variables to obtain the l.r.t. of block equicorrelation and equivariance. This test or its single block version may find applications in many areas as in psychology, education, medicine, genetics and they are important "in many tests of multivariate analysis, e.g. in MANOVA, Profile Analysis, Growth Curve analysis, etc" [12, 9]. By decomposing the overall hypothesis into the hypotheses of independence of groups of variables and the hypothesis of equicorrelation and equivariance we are able to obtain the expressions for the overall l.r.t. statistic and its moments. From these we obtain a suitable factorization of the characteristic function (c.f.) of the logarithm of the l.r.t. statistic, which enables us to develop highly manageable and precise near-exact distributions for the test statistic.

8. Empirical Correction to the Likelihood Ratio Statistic for Structural Equation Modeling with Many Variables.

Science.gov (United States)

Yuan, Ke-Hai; Tian, Yubin; Yanagihara, Hirokazu

2015-06-01

Survey data typically contain many variables. Structural equation modeling (SEM) is commonly used in analyzing such data. The most widely used statistic for evaluating the adequacy of a SEM model is T ML, a slight modification to the likelihood ratio statistic. Under normality assumption, T ML approximately follows a chi-square distribution when the number of observations (N) is large and the number of items or variables (p) is small. However, in practice, p can be rather large while N is always limited due to not having enough participants. Even with a relatively large N, empirical results show that T ML rejects the correct model too often when p is not too small. Various corrections to T ML have been proposed, but they are mostly heuristic. Following the principle of the Bartlett correction, this paper proposes an empirical approach to correct T ML so that the mean of the resulting statistic approximately equals the degrees of freedom of the nominal chi-square distribution. Results show that empirically corrected statistics follow the nominal chi-square distribution much more closely than previously proposed corrections to T ML, and they control type I errors reasonably well whenever N ≥ max(50,2p). The formulations of the empirically corrected statistics are further used to predict type I errors of T ML as reported in the literature, and they perform well.

9. A flexible spatial scan statistic with a restricted likelihood ratio for detecting disease clusters.

Science.gov (United States)

Tango, Toshiro; Takahashi, Kunihiko

2012-12-30

Spatial scan statistics are widely used tools for detection of disease clusters. Especially, the circular spatial scan statistic proposed by Kulldorff (1997) has been utilized in a wide variety of epidemiological studies and disease surveillance. However, as it cannot detect noncircular, irregularly shaped clusters, many authors have proposed different spatial scan statistics, including the elliptic version of Kulldorff's scan statistic. The flexible spatial scan statistic proposed by Tango and Takahashi (2005) has also been used for detecting irregularly shaped clusters. However, this method sets a feasible limitation of a maximum of 30 nearest neighbors for searching candidate clusters because of heavy computational load. In this paper, we show a flexible spatial scan statistic implemented with a restricted likelihood ratio proposed by Tango (2008) to (1) eliminate the limitation of 30 nearest neighbors and (2) to have surprisingly much less computational time than the original flexible spatial scan statistic. As a side effect, it is shown to be able to detect clusters with any shape reasonably well as the relative risk of the cluster becomes large via Monte Carlo simulation. We illustrate the proposed spatial scan statistic with data on mortality from cerebrovascular disease in the Tokyo Metropolitan area, Japan. Copyright © 2012 John Wiley & Sons, Ltd.

10. Sensitivity, specificity and likelihood ratios of PCR in the diagnosis of syphilis: a systematic review and meta-analysis.

Science.gov (United States)

Gayet-Ageron, Angèle; Lautenschlager, Stephan; Ninet, Béatrice; Perneger, Thomas V; Combescure, Christophe

2013-05-01

To systematically review and estimate pooled sensitivity and specificity of the polymerase chain reaction (PCR) technique compared to recommended reference tests in the diagnosis of suspected syphilis at various stages and in various biological materials. Systematic review and meta-analysis. Search of three electronic bibliographic databases from January 1990 to January 2012 and the abstract books of five congresses specialized in the infectious diseases' field (1999-2011). Search key terms included syphilis, Treponema pallidum or neurosyphilis and molecular amplification, polymerase chain reaction or PCR. We included studies that used both reference tests to diagnose syphilis plus PCR and we presented pooled estimates of PCR sensitivity, specificity, and positive and negative likelihood ratios (LR) per syphilis stages and biological materials. Of 1160 identified abstracts, 69 were selected and 46 studies used adequate reference tests to diagnose syphilis. Sensitivity was highest in the swabs from primary genital or anal chancres (78.4%; 95% CI: 68.2-86.0) and in blood from neonates with congenital syphilis (83.0%; 55.0-95.2). Most pooled specificities were ∼95%, except those in blood. A positive PCR is highly informative with a positive LR around 20 in ulcers or skin lesions. In the blood, the positive LR was syphilis diagnosis in lesions. PCR is a useful diagnostic tool in ulcers, especially when serology is still negative and in medical settings with a high prevalence of syphilis.

11. Likelihood ratio and posterior odds in forensic genetics: Two sides of the same coin.

Science.gov (United States)

Caliebe, Amke; Walsh, Susan; Liu, Fan; Kayser, Manfred; Krawczak, Michael

2017-05-01

It has become widely accepted in forensics that, owing to a lack of sensible priors, the evidential value of matching DNA profiles in trace donor identification or kinship analysis is most sensibly communicated in the form of a likelihood ratio (LR). This restraint does not abate the fact that the posterior odds (PO) would be the preferred basis for returning a verdict. A completely different situation holds for Forensic DNA Phenotyping (FDP), which is aimed at predicting externally visible characteristics (EVCs) of a trace donor from DNA left behind at the crime scene. FDP is intended to provide leads to the police investigation helping them to find unknown trace donors that are unidentifiable by DNA profiling. The statistical models underlying FDP typically yield posterior odds (PO) for an individual possessing a certain EVC. This apparent discrepancy has led to confusion as to when LR or PO is the appropriate outcome of forensic DNA analysis to be communicated to the investigating authorities. We thus set out to clarify the distinction between LR and PO in the context of forensic DNA profiling and FDP from a statistical point of view. In so doing, we also addressed the influence of population affiliation on LR and PO. In contrast to the well-known population dependency of the LR in DNA profiling, the PO as obtained in FDP may be widely population-independent. The actual degree of independence, however, is a matter of (i) how much of the causality of the respective EVC is captured by the genetic markers used for FDP and (ii) by the extent to which non-genetic such as environmental causal factors of the same EVC are distributed equally throughout populations. The fact that an LR should be communicated in cases of DNA profiling whereas the PO are suitable for FDP does not conflict with theory, but rather reflects the immanent differences between these two forensic applications of DNA information. Copyright © 2017 Elsevier B.V. All rights reserved.

12. Likelihood ratio-based integrated personal risk assessment of type 2 diabetes.

Science.gov (United States)

Sato, Noriko; Htun, Nay Chi; Daimon, Makoto; Tamiya, Gen; Kato, Takeo; Kubota, Isao; Ueno, Yoshiyuki; Yamashita, Hidetoshi; Fukao, Akira; Kayama, Takamasa; Muramatsu, Masaaki

2014-01-01

To facilitate personalized health care for multifactorial diseases, risks of genetic and clinical/environmental factors should be assessed together for each individual in an integrated fashion. This approach is possible with the likelihood ratio (LR)-based risk assessment system, as this system can incorporate manifold tests. We examined the usefulness of this system for assessing type 2 diabetes (T2D). Our system employed 29 genetic susceptibility variants, body mass index (BMI), and hypertension as risk factors whose LRs can be estimated from openly available T2D association data for the Japanese population. The pretest probability was set at a sex- and age-appropriate population average of diabetes prevalence. The classification performance of our LR-based risk assessment was compared to that of a non-invasive screening test for diabetes called TOPICS (with score based on age, sex, family history, smoking, BMI, and hypertension) using receiver operating characteristic analysis with a community cohort (n = 1263). The area under the receiver operating characteristic curve (AUC) for the LR-based assessment and TOPICS was 0.707 (95% CI 0.665-0.750) and 0.719 (0.675-0.762), respectively. These AUCs were much higher than that of a genetic risk score constructed using the same genetic susceptibility variants, 0.624 (0.574-0.674). The use of ethnically matched LRs is necessary for proper personal risk assessment. In conclusion, although LR-based integrated risk assessment for T2D still requires additional tests that evaluate other factors, such as risks involved in missing heritability, our results indicate the potential usability of LR-based assessment system and stress the importance of stratified epidemiological investigations in personalized medicine.

13. Geochemical wolframite fingerprinting - the likelihood ratio approach for laser ablation ICP-MS data.

Science.gov (United States)

Martyna, Agnieszka; Gäbler, Hans-Eike; Bahr, Andreas; Zadora, Grzegorz

2018-05-01

Wolframite has been specified as a 'conflict mineral' by a U.S. Government Act, which obliges companies that use these minerals to report their origin. Minerals originating from conflict regions in the Democratic Republic of the Congo shall be excluded from the market as their illegal mining, trading, and taxation are supposed to fuel ongoing violent conflicts. The German Federal Institute for Geosciences and Natural Resources (BGR) developed a geochemical fingerprinting method for wolframite based on laser ablation inductively coupled plasma-mass spectrometry. Concentrations of 46 elements in about 5300 wolframite grains from 64 mines were determined. The issue of verifying the declared origins of the wolframite samples may be framed as a forensic problem by considering two contrasting hypotheses: the examined sample and a sample collected from the declared mine originate from the same mine (H 1 ), and the two samples come from different mines (H 2 ). The solution is found using the likelihood ratio (LR) theory. On account of the multidimensionality, the lack of normal distribution of data within each sample, and the huge within-sample dispersion in relation to the dispersion between samples, the classic LR models had to be modified. Robust principal component analysis and linear discriminant analysis were used to characterize samples. The similarity of two samples was expressed by Kolmogorov-Smirnov distances, which were interpreted in view of H 1 and H 2 hypotheses within the LR framework. The performance of the models, controlled by the levels of incorrect responses and the empirical cross entropy, demonstrated that the proposed LR models are successful in verifying the authenticity of the wolframite samples. Graphical abstract Geochemical wolframite fingerprinting.

14. A comparison of likelihood ratio tests and Rao's score test for three separable covariance matrix structures.

Science.gov (United States)

Filipiak, Katarzyna; Klein, Daniel; Roy, Anuradha

2017-01-01

The problem of testing the separability of a covariance matrix against an unstructured variance-covariance matrix is studied in the context of multivariate repeated measures data using Rao's score test (RST). The RST statistic is developed with the first component of the separable structure as a first-order autoregressive (AR(1)) correlation matrix or an unstructured (UN) covariance matrix under the assumption of multivariate normality. It is shown that the distribution of the RST statistic under the null hypothesis of any separability does not depend on the true values of the mean or the unstructured components of the separable structure. A significant advantage of the RST is that it can be performed for small samples, even smaller than the dimension of the data, where the likelihood ratio test (LRT) cannot be used, and it outperforms the standard LRT in a number of contexts. Monte Carlo simulations are then used to study the comparative behavior of the null distribution of the RST statistic, as well as that of the LRT statistic, in terms of sample size considerations, and for the estimation of the empirical percentiles. Our findings are compared with existing results where the first component of the separable structure is a compound symmetry (CS) correlation matrix. It is also shown by simulations that the empirical null distribution of the RST statistic converges faster than the empirical null distribution of the LRT statistic to the limiting χ 2 distribution. The tests are implemented on a real dataset from medical studies. © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

15. Likelihood analysis of the chalcone synthase genes suggests the role of positive selection in morning glories (Ipomoea).

Science.gov (United States)

Yang, Ji; Gu, Hongya; Yang, Ziheng

2004-01-01

Chalcone synthase (CHS) is a key enzyme in the biosynthesis of flavonoides, which are important for the pigmentation of flowers and act as attractants to pollinators. Genes encoding CHS constitute a multigene family in which the copy number varies among plant species and functional divergence appears to have occurred repeatedly. In morning glories (Ipomoea), five functional CHS genes (A-E) have been described. Phylogenetic analysis of the Ipomoea CHS gene family revealed that CHS A, B, and C experienced accelerated rates of amino acid substitution relative to CHS D and E. To examine whether the CHS genes of the morning glories underwent adaptive evolution, maximum-likelihood models of codon substitution were used to analyze the functional sequences in the Ipomoea CHS gene family. These models used the nonsynonymous/synonymous rate ratio (omega = d(N)/ d(S)) as an indicator of selective pressure and allowed the ratio to vary among lineages or sites. Likelihood ratio test suggested significant variation in selection pressure among amino acid sites, with a small proportion of them detected to be under positive selection along the branches ancestral to CHS A, B, and C. Positive Darwinian selection appears to have promoted the divergence of subfamily ABC and subfamily DE and is at least partially responsible for a rate increase following gene duplication.

16. An assessment of the information content of likelihood ratios derived from complex mixtures.

Science.gov (United States)

Marsden, Clare D; Rudin, Norah; Inman, Keith; Lohmueller, Kirk E

2016-05-01

With the increasing sensitivity of DNA typing methodologies, as well as increasing awareness by law enforcement of the perceived capabilities of DNA typing, complex mixtures consisting of DNA from two or more contributors are increasingly being encountered. However, insufficient research has been conducted to characterize the ability to distinguish a true contributor (TC) from a known non-contributor (KNC) in these complex samples, and under what specific conditions. In order to investigate this question, sets of six 15-locus Caucasian genotype profiles were simulated and used to create mixtures containing 2-5 contributors. Likelihood ratios were computed for various situations, including varying numbers of contributors and unknowns in the evidence profile, as well as comparisons of the evidence profile to TCs and KNCs. This work was intended to illustrate the best-case scenario, in which all alleles from the TC were detected in the simulated evidence samples. Therefore the possibility of drop-out was not modeled in this study. The computer program DNAMIX was then used to compute LRs comparing the evidence profile to TCs and KNCs. This resulted in 140,000 LRs for each of the two scenarios. These complex mixture simulations show that, even when all alleles are detected (i.e. no drop-out), TCs can generate LRs less than 1 across a 15-locus profile. However, this outcome was rare, 7 of 140,000 replicates (0.005%), and associated only with mixtures comprising 5 contributors in which the numerator hypothesis includes one or more unknown contributors. For KNCs, LRs were found to be greater than 1 in a small number of replicates (75 of 140,000 replicates, or 0.05%). These replicates were limited to 4 and 5 person mixtures with 1 or more unknowns in the numerator. Only 5 of these 75 replicates (0.004%) yielded an LR greater than 1,000. Thus, overall, these results imply that the weight of evidence that can be derived from complex mixtures containing up to 5 contributors

17. Measures of effect size for chi-squared and likelihood-ratio goodness-of-fit tests.

Science.gov (United States)

Johnston, Janis E; Berry, Kenneth J; Mielke, Paul W

2006-10-01

A fundamental shift in editorial policy for psychological journals was initiated when the fourth edition of the Publication Manual of the American Psychological Association (1994) placed emphasis on reporting measures of effect size. This paper presents measures of effect size for the chi-squared and the likelihood-ratio goodness-of-fit statistic tests.

18. Noncentral Chi-Square versus Normal Distributions in Describing the Likelihood Ratio Statistic: The Univariate Case and Its Multivariate Implication

Science.gov (United States)

Yuan, Ke-Hai

2008-01-01

In the literature of mean and covariance structure analysis, noncentral chi-square distribution is commonly used to describe the behavior of the likelihood ratio (LR) statistic under alternative hypothesis. Due to the inaccessibility of the rather technical literature for the distribution of the LR statistic, it is widely believed that the…

19. A preliminary evaluation of the generalized likelihood ratio for detecting and identifying control element failures in a transport aircraft

Science.gov (United States)

Bundick, W. T.

1985-01-01

The application of the Generalized Likelihood Ratio technique to the detection and identification of aircraft control element failures has been evaluated in a linear digital simulation of the longitudinal dynamics of a B-737 aircraft. Simulation results show that the technique has potential but that the effects of wind turbulence and Kalman filter model errors are problems which must be overcome.

20. Evaluating score- and feature-based likelihood ratio models for multivariate continuous data: applied to forensic MDMA comparison

NARCIS (Netherlands)

Bolck, A.; Ni, H.; Lopatka, M.

2015-01-01

Likelihood ratio (LR) models are moving into the forefront of forensic evidence evaluation as these methods are adopted by a diverse range of application areas in forensic science. We examine the fundamentally different results that can be achieved when feature- and score-based methodologies are

1. Further Evaluation of Covariate Analysis using Empirical Bayes Estimates in Population Pharmacokinetics: the Perception of Shrinkage and Likelihood Ratio Test.

Science.gov (United States)

Xu, Xu Steven; Yuan, Min; Yang, Haitao; Feng, Yan; Xu, Jinfeng; Pinheiro, Jose

2017-01-01

Covariate analysis based on population pharmacokinetics (PPK) is used to identify clinically relevant factors. The likelihood ratio test (LRT) based on nonlinear mixed effect model fits is currently recommended for covariate identification, whereas individual empirical Bayesian estimates (EBEs) are considered unreliable due to the presence of shrinkage. The objectives of this research were to investigate the type I error for LRT and EBE approaches, to confirm the similarity of power between the LRT and EBE approaches from a previous report and to explore the influence of shrinkage on LRT and EBE inferences. Using an oral one-compartment PK model with a single covariate impacting on clearance, we conducted a wide range of simulations according to a two-way factorial design. The results revealed that the EBE-based regression not only provided almost identical power for detecting a covariate effect, but also controlled the false positive rate better than the LRT approach. Shrinkage of EBEs is likely not the root cause for decrease in power or inflated false positive rate although the size of the covariate effect tends to be underestimated at high shrinkage. In summary, contrary to the current recommendations, EBEs may be a better choice for statistical tests in PPK covariate analysis compared to LRT. We proposed a three-step covariate modeling approach for population PK analysis to utilize the advantages of EBEs while overcoming their shortcomings, which allows not only markedly reducing the run time for population PK analysis, but also providing more accurate covariate tests.

2. Validation of SmartRank: A likelihood ratio software for searching national DNA databases with complex DNA profiles.

Science.gov (United States)

Benschop, Corina C G; van de Merwe, Linda; de Jong, Jeroen; Vanvooren, Vanessa; Kempenaers, Morgane; Kees van der Beek, C P; Barni, Filippo; Reyes, Eusebio López; Moulin, Léa; Pene, Laurent; Haned, Hinda; Sijen, Titia

2017-07-01

Searching a national DNA database with complex and incomplete profiles usually yields very large numbers of possible matches that can present many candidate suspects to be further investigated by the forensic scientist and/or police. Current practice in most forensic laboratories consists of ordering these 'hits' based on the number of matching alleles with the searched profile. Thus, candidate profiles that share the same number of matching alleles are not differentiated and due to the lack of other ranking criteria for the candidate list it may be difficult to discern a true match from the false positives or notice that all candidates are in fact false positives. SmartRank was developed to put forward only relevant candidates and rank them accordingly. The SmartRank software computes a likelihood ratio (LR) for the searched profile and each profile in the DNA database and ranks database entries above a defined LR threshold according to the calculated LR. In this study, we examined for mixed DNA profiles of variable complexity whether the true donors are retrieved, what the number of false positives above an LR threshold is and the ranking position of the true donors. Using 343 mixed DNA profiles over 750 SmartRank searches were performed. In addition, the performance of SmartRank and CODIS were compared regarding DNA database searches and SmartRank was found complementary to CODIS. We also describe the applicable domain of SmartRank and provide guidelines. The SmartRank software is open-source and freely available. Using the best practice guidelines, SmartRank enables obtaining investigative leads in criminal cases lacking a suspect. Copyright © 2017 Elsevier B.V. All rights reserved.

3. An Accurate Method for Inferring Relatedness in Large Datasets of Unphased Genotypes via an Embedded Likelihood-Ratio Test

KAUST Repository

Rodriguez, Jesse M.

2013-01-01

Studies that map disease genes rely on accurate annotations that indicate whether individuals in the studied cohorts are related to each other or not. For example, in genome-wide association studies, the cohort members are assumed to be unrelated to one another. Investigators can correct for individuals in a cohort with previously-unknown shared familial descent by detecting genomic segments that are shared between them, which are considered to be identical by descent (IBD). Alternatively, elevated frequencies of IBD segments near a particular locus among affected individuals can be indicative of a disease-associated gene. As genotyping studies grow to use increasingly large sample sizes and meta-analyses begin to include many data sets, accurate and efficient detection of hidden relatedness becomes a challenge. To enable disease-mapping studies of increasingly large cohorts, a fast and accurate method to detect IBD segments is required. We present PARENTE, a novel method for detecting related pairs of individuals and shared haplotypic segments within these pairs. PARENTE is a computationally-efficient method based on an embedded likelihood ratio test. As demonstrated by the results of our simulations, our method exhibits better accuracy than the current state of the art, and can be used for the analysis of large genotyped cohorts. PARENTE\\'s higher accuracy becomes even more significant in more challenging scenarios, such as detecting shorter IBD segments or when an extremely low false-positive rate is required. PARENTE is publicly and freely available at http://parente.stanford.edu/. © 2013 Springer-Verlag.

4. Does lying in the recovery position increase the likelihood of not delivering cardiopulmonary resuscitation?

Science.gov (United States)

Freire-Tellado, Miguel; Navarro-Patón, Rubén; Pavón-Prieto, Maria Del Pilar; Fernández-López, Marta; Mateos-Lorenzo, Javier; López-Fórneas, Ivan

2017-06-01

5. Performance and sensitivity analysis of the generalized likelihood ratio method for failure detection. M.S. Thesis

Science.gov (United States)

Bueno, R. A.

1977-01-01

Results of the generalized likelihood ratio (GLR) technique for the detection of failures in aircraft application are presented, and its relationship to the properties of the Kalman-Bucy filter is examined. Under the assumption that the system is perfectly modeled, the detectability and distinguishability of four failure types are investigated by means of analysis and simulations. Detection of failures is found satisfactory, but problems in identifying correctly the mode of a failure may arise. These issues are closely examined as well as the sensitivity of GLR to modeling errors. The advantages and disadvantages of this technique are discussed, and various modifications are suggested to reduce its limitations in performance and computational complexity.

6. Enhancing their likelihood for a positive future: the perspective of inner-city youth.

Science.gov (United States)

Ginsburg, Kenneth R; Alexander, Penny M; Hunt, Jean; Sullivan, Maisha; Zhao, Huaqing; Cnaan, Avital

2002-06-01

Inner-city youth must overcome many environmental challenges as they strive for success. Their outcome is influenced by the interplay of protective forces and risk factors. To learn directly from youth what solutions they believe would most influence their likelihood of achieving a positive future. In-school 8th-, 9th-, and 12th-graders in north Philadelphia generated, prioritized, and explained their own solutions through a 4-stage hierarchical process facilitated by AmeriCorps workers. In Stage 1, 60 randomly selected students participated in 8 focus groups to develop the study question. In Stage 2, youth in Nominal Group Technique sessions generated and prioritized solutions. In Stage 3, a survey for each grade that included their top prioritized ideas was distributed, and youth rated each idea on a Likert scale (5= Definitely would make me more likely to have a positive future to 1 = Would definitely not.). One thousand twenty-two ninth-graders (69% of in-school youth at 5 high schools) returned usable surveys. Ninety-three percent of responders were 14 to 16 years old, 44% were male, 54% were black, and 32% were Latino. Four hundred seventeen 8th-graders and 322 12th-graders returned usable surveys. In Stage 4, youth in 10 focus groups added meaning and context to the ideas. The highest rated items in all grades were solutions that promoted education or increased job opportunities. Ninth-graders ranked helping youth get into college first by the Marginal Homogeneity Test. The creation of more jobs was ranked second. Third rank was shared by more job training, keeping youth from dropping out of school, and better books for schools. The next tier of items focused mostly on opportunities for youth to spend their free time productively and to have interactions with adults. Many items calling for the reduction of risk behaviors or disruptive surroundings were rated lower. The Kruskal-Wallis test found little variation in rating of the ideas by gender, race, or

7. An empirical likelihood ratio test robust to individual heterogeneity for differential expression analysis of RNA-seq.

Science.gov (United States)

Xu, Maoqi; Chen, Liang

2018-01-01

8. An Adjusted Likelihood Ratio Approach Analysing Distribution of Food Products to Assist the Investigation of Foodborne Outbreaks

Science.gov (United States)

Norström, Madelaine; Kristoffersen, Anja Bråthen; Görlach, Franziska Sophie; Nygård, Karin; Hopp, Petter

2015-01-01

In order to facilitate foodborne outbreak investigations there is a need to improve the methods for identifying the food products that should be sampled for laboratory analysis. The aim of this study was to examine the applicability of a likelihood ratio approach previously developed on simulated data, to real outbreak data. We used human case and food product distribution data from the Norwegian enterohaemorrhagic Escherichia coli outbreak in 2006. The approach was adjusted to include time, space smoothing and to handle missing or misclassified information. The performance of the adjusted likelihood ratio approach on the data originating from the HUS outbreak and control data indicates that the adjusted approach is promising and indicates that the adjusted approach could be a useful tool to assist and facilitate the investigation of food borne outbreaks in the future if good traceability are available and implemented in the distribution chain. However, the approach needs to be further validated on other outbreak data and also including other food products than meat products in order to make a more general conclusion of the applicability of the developed approach. PMID:26237468

9. The Likelihood Ratio Test of Common Factors under Non-Ideal Conditions

Directory of Open Access Journals (Sweden)

Ana M. Angulo

2011-01-01

Full Text Available El modelo espacial de Durbin ocupa una posición interesante en econometría espacial. Es la forma reducida de un modelo de corte transversal con dependencia en los errores y puede ser utilizado como ecuación de anidación en un enfoque más general de selección de modelos. En concreto, a partir de esta ecuación puede obtenerse el Ratio de Verosimilitudes conocido como test de Factores Comunes (LRCOM. Como se muestra en Mur y Angulo (2006, este test tiene buenas propiedades si el modelo está correctamente especificado. Sin embargo, por lo que sabemos, no hay referencias en la literatura sobre el comportamiento de este test bajo condiciones no ideales. En concreto, estudiamos el comportamiento del test en los casos de heterocedasticidad, no normalidad, endogeneidad, matrices de contactos densas y no-linealidad. Nuestros resultados ofrecen una visión positiva del test de Factores Comunes que parece una técnica útil en el instrumental propio de la econometría espacial contemporánea.

10. Approximate Likelihood

CERN Multimedia

CERN. Geneva

2015-01-01

Most physics results at the LHC end in a likelihood ratio test. This includes discovery and exclusion for searches as well as mass, cross-section, and coupling measurements. The use of Machine Learning (multivariate) algorithms in HEP is mainly restricted to searches, which can be reduced to classification between two fixed distributions: signal vs. background. I will show how we can extend the use of ML classifiers to distributions parameterized by physical quantities like masses and couplings as well as nuisance parameters associated to systematic uncertainties. This allows for one to approximate the likelihood ratio while still using a high dimensional feature vector for the data. Both the MEM and ABC approaches mentioned above aim to provide inference on model parameters (like cross-sections, masses, couplings, etc.). ABC is fundamentally tied Bayesian inference and focuses on the “likelihood free” setting where only a simulator is available and one cannot directly compute the likelihood for the dat...

11. Nuclear Power Plant Thermocouple Sensor-Fault Detection and Classification Using Deep Learning and Generalized Likelihood Ratio Test

Science.gov (United States)

Mandal, Shyamapada; Santhi, B.; Sridhar, S.; Vinolia, K.; Swaminathan, P.

2017-06-01

In this paper, an online fault detection and classification method is proposed for thermocouples used in nuclear power plants. In the proposed method, the fault data are detected by the classification method, which classifies the fault data from the normal data. Deep belief network (DBN), a technique for deep learning, is applied to classify the fault data. The DBN has a multilayer feature extraction scheme, which is highly sensitive to a small variation of data. Since the classification method is unable to detect the faulty sensor; therefore, a technique is proposed to identify the faulty sensor from the fault data. Finally, the composite statistical hypothesis test, namely generalized likelihood ratio test, is applied to compute the fault pattern of the faulty sensor signal based on the magnitude of the fault. The performance of the proposed method is validated by field data obtained from thermocouple sensors of the fast breeder test reactor.

12. Maximum likelihood estimation of the position of a radiating source in a waveguide

International Nuclear Information System (INIS)

Hinich, M.J.

1979-01-01

An array of sensors is receiving radiation from a source of interest. The source and the array are in a one- or two-dimensional waveguide. The maximum-likelihood estimators of the coordinates of the source are analyzed under the assumptions that the noise field is Gaussian. The Cramer-Rao lower bound is of the order of the number of modes which define the source excitation function. The results show that the accuracy of the maximum likelihood estimator of source depth using a vertical array in a infinite horizontal waveguide (such as the ocean) is limited by the number of modes detected by the array regardless of the array size

13. Elastic properties of a material composed of alternating layers of negative and positive Poisson's ratio

International Nuclear Information System (INIS)

Kocer, C.; McKenzie, D.R.; Bilek, M.M.

2009-01-01

The theory of elasticity predicts a variety of phenomena associated with solids that possess a negative Poisson's ratio. The fabrication of metamaterials with a 'designed' microstructure that exhibit a Poisson's ratio approaching the thermodynamic limits of 1/2 and -1 increases the likelihood of realising these phenomena for applications. In this work, we investigate the properties of a layered composite, with alternating layers of materials with negative and positive Poisson's ratio approaching the thermodynamic limits. Using the finite element method to simulate uniaxial loading and indentation of a free standing composite, we observed an increase in the resistance to mechanical deformation above the average value of the two materials. Even though the greatest increase in stiffness is gained as the thermodynamic limits are approached, a significant amount of added stiffness can be attained, provided that the Young's modulus of the negative Poisson's ratio material is not less than that of the positive Poisson's ratio material

14. Likelihood ratio-based differentiation of nodular Hashimoto thyroiditis and papillary thyroid carcinoma in patients with sonographically evident diffuse hashimoto thyroiditis: preliminary study.

Science.gov (United States)

Wang, Liang; Xia, Yu; Jiang, Yu-Xin; Dai, Qing; Li, Xiao-Yi

2012-11-01

To assess the efficacy of sonography for discriminating nodular Hashimoto thyroiditis from papillary thyroid carcinoma in patients with sonographically evident diffuse Hashimoto thyroiditis. This study included 20 patients with 24 surgically confirmed Hashimoto thyroiditis nodules and 40 patients with 40 papillary thyroid carcinoma nodules; all had sonographically evident diffuse Hashimoto thyroiditis. A retrospective review of the sonograms was performed, and significant benign and malignant sonographic features were selected by univariate and multivariate analyses. The combined likelihood ratio was calculated as the product of each feature's likelihood ratio for papillary thyroid carcinoma. We compared the abilities of the original sonographic features and combined likelihood ratios in diagnosing nodular Hashimoto thyroiditis and papillary thyroid carcinoma by their sensitivity, specificity, and Youden index. The diagnostic capabilities of the sonographic features varied greatly, with Youden indices ranging from 0.175 to 0.700. Compared with single features, combinations of features were unable to improve the Youden indices effectively because the sensitivity and specificity usually changed in opposite directions. For combined likelihood ratios, however, the sensitivity improved greatly without an obvious reduction in specificity, which resulted in the maximum Youden index (0.825). With a combined likelihood ratio greater than 7.00 as the diagnostic criterion for papillary thyroid carcinoma, sensitivity reached 82.5%, whereas specificity remained at 100.0%. With a combined likelihood ratio less than 1.00 for nodular Hashimoto thyroiditis, sensitivity and specificity were 90.0% and 92.5%, respectively. Several sonographic features of nodular Hashimoto thyroiditis and papillary thyroid carcinoma in a background of diffuse Hashimoto thyroiditis were significantly different. The combined likelihood ratio may be superior to original sonographic features for

15. Tolerance to missing data using a likelihood ratio based classifier for computer-aided classification of breast cancer

International Nuclear Information System (INIS)

Bilska-Wolak, Anna O; Floyd, Carey E Jr

2004-01-01

While mammography is a highly sensitive method for detecting breast tumours, its ability to differentiate between malignant and benign lesions is low, which may result in as many as 70% of unnecessary biopsies. The purpose of this study was to develop a highly specific computer-aided diagnosis algorithm to improve classification of mammographic masses. A classifier based on the likelihood ratio was developed to accommodate cases with missing data. Data for development included 671 biopsy cases (245 malignant), with biopsy-proved outcome. Sixteen features based on the BI-RADS TM lexicon and patient history had been recorded for the cases, with 1.3 ± 1.1 missing feature values per case. Classifier evaluation methods included receiver operating characteristic and leave-one-out bootstrap sampling. The classifier achieved 32% specificity at 100% sensitivity on the 671 cases with 16 features that had missing values. Utilizing just the seven features present for all cases resulted in decreased performance at 100% sensitivity with average 19% specificity. No cases and no feature data were omitted during classifier development, showing that it is more beneficial to utilize cases with missing values than to discard incomplete cases that cannot be handled by many algorithms. Classification of mammographic masses was commendable at high sensitivity levels, indicating that benign cases could be potentially spared from biopsy

16. (Re)evaluating the Implications of the Autoregressive Latent Trajectory Model Through Likelihood Ratio Tests of Its Initial Conditions.

Science.gov (United States)

Ou, Lu; Chow, Sy-Miin; Ji, Linying; Molenaar, Peter C M

2017-01-01

The autoregressive latent trajectory (ALT) model synthesizes the autoregressive model and the latent growth curve model. The ALT model is flexible enough to produce a variety of discrepant model-implied change trajectories. While some researchers consider this a virtue, others have cautioned that this may confound interpretations of the model's parameters. In this article, we show that some-but not all-of these interpretational difficulties may be clarified mathematically and tested explicitly via likelihood ratio tests (LRTs) imposed on the initial conditions of the model. We show analytically the nested relations among three variants of the ALT model and the constraints needed to establish equivalences. A Monte Carlo simulation study indicated that LRTs, particularly when used in combination with information criterion measures, can allow researchers to test targeted hypotheses about the functional forms of the change process under study. We further demonstrate when and how such tests may justifiably be used to facilitate our understanding of the underlying process of change using a subsample (N = 3,995) of longitudinal family income data from the National Longitudinal Survey of Youth.

17. A likelihood ratio-based method to predict exact pedigrees for complex families from next-generation sequencing data.

Science.gov (United States)

Heinrich, Verena; Kamphans, Tom; Mundlos, Stefan; Robinson, Peter N; Krawitz, Peter M

2017-01-01

Next generation sequencing technology considerably changed the way we screen for pathogenic mutations in rare Mendelian disorders. However, the identification of the disease-causing mutation amongst thousands of variants of partly unknown relevance is still challenging and efficient techniques that reduce the genomic search space play a decisive role. Often segregation- or linkage analysis are used to prioritize candidates, however, these approaches require correct information about the degree of relationship among the sequenced samples. For quality assurance an automated control of pedigree structures and sample assignment is therefore highly desirable in order to detect label mix-ups that might otherwise corrupt downstream analysis. We developed an algorithm based on likelihood ratios that discriminates between different classes of relationship for an arbitrary number of genotyped samples. By identifying the most likely class we are able to reconstruct entire pedigrees iteratively, even for highly consanguineous families. We tested our approach on exome data of different sequencing studies and achieved high precision for all pedigree predictions. By analyzing the precision for varying degrees of relatedness or inbreeding we could show that a prediction is robust down to magnitudes of a few hundred loci. A java standalone application that computes the relationships between multiple samples as well as a Rscript that visualizes the pedigree information is available for download as well as a web service at www.gene-talk.de CONTACT: heinrich@molgen.mpg.deSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.

18. Log-Likelihood Ratio Calculation for Iterative Decoding on Rayleigh Fading Channels Using Padé Approximation

Directory of Open Access Journals (Sweden)

Gou Hosoya

2013-01-01

Full Text Available Approximate calculation of channel log-likelihood ratio (LLR for wireless channels using Padé approximation is presented. LLR is used as an input of iterative decoding for powerful error-correcting codes such as low-density parity-check (LDPC codes or turbo codes. Due to the lack of knowledge of the channel state information of a wireless fading channel, such as uncorrelated fiat Rayleigh fading channels, calculations of exact LLR for these channels are quite complicated for a practical implementation. The previous work, an LLR calculation using the Taylor approximation, quickly becomes inaccurate as the channel output leaves some derivative point. This becomes a big problem when higher order modulation scheme is employed. To overcome this problem, a new LLR approximation using Padé approximation, which expresses the original function by a rational form of two polynomials with the same total number of coefficients of the Taylor series and can accelerate the Taylor approximation, is devised. By applying the proposed approximation to the iterative decoding and the LDPC codes with some modulation schemes, we show the effectiveness of the proposed methods by simulation results and analysis based on the density evolution.

19. Evaluation of direct and indirect ethanol biomarkers using a likelihood ratio approach to identify chronic alcohol abusers for forensic purposes.

Science.gov (United States)

Alladio, Eugenio; Martyna, Agnieszka; Salomone, Alberto; Pirro, Valentina; Vincenti, Marco; Zadora, Grzegorz

2017-02-01

The detection of direct ethanol metabolites, such as ethyl glucuronide (EtG) and fatty acid ethyl esters (FAEEs), in scalp hair is considered the optimal strategy to effectively recognize chronic alcohol misuses by means of specific cut-offs suggested by the Society of Hair Testing. However, several factors (e.g. hair treatments) may alter the correlation between alcohol intake and biomarkers concentrations, possibly introducing bias in the interpretative process and conclusions. 125 subjects with various drinking habits were subjected to blood and hair sampling to determine indirect (e.g. CDT) and direct alcohol biomarkers. The overall data were investigated using several multivariate statistical methods. A likelihood ratio (LR) approach was used for the first time to provide predictive models for the diagnosis of alcohol abuse, based on different combinations of direct and indirect alcohol biomarkers. LR strategies provide a more robust outcome than the plain comparison with cut-off values, where tiny changes in the analytical results can lead to dramatic divergence in the way they are interpreted. An LR model combining EtG and FAEEs hair concentrations proved to discriminate non-chronic from chronic consumers with ideal correct classification rates, whereas the contribution of indirect biomarkers proved to be negligible. Optimal results were observed using a novel approach that associates LR methods with multivariate statistics. In particular, the combination of LR approach with either Principal Component Analysis (PCA) or Linear Discriminant Analysis (LDA) proved successful in discriminating chronic from non-chronic alcohol drinkers. These LR models were subsequently tested on an independent dataset of 43 individuals, which confirmed their high efficiency. These models proved to be less prone to bias than EtG and FAEEs independently considered. In conclusion, LR models may represent an efficient strategy to sustain the diagnosis of chronic alcohol consumption

20. Prevalence and likelihood ratio of symptoms in patients with good therapeutic response to Lycopodium clavatum. A retrospective study.

Science.gov (United States)

Eizayaga, José Enrique; Pozzi, María Isabel; Canan, María Clara; Saravia, Laura

2016-02-01

Assessment of the likelihood ratio (LR) of symptoms has been proposed as a rational means for detecting indicators to homeopathic medicines. To investigate the prevalence and LR of symptoms commonly attributed to the homeopathic medicine Lycopodium clavatum (Lyc). Secondarily, to answer the question if experienced homeopaths could intuitively infer which the main symptoms of this medicine are. The presence of 35 selected symptoms, prescribed medicines and therapeutic response were assessed retrospectively. The symptoms' prevalence in the Lyc responding population and the LR of the symptoms compared to their prevalence in the remainder of the population were calculated. Two hundred and two Lyc and 550 non Lyc cases (total 752) were included for analysis. Twenty-two symptoms were confirmed as pertaining to Lyc's semiology (prevalence %; LR): contemptuous (3.3; 6.7), urinary stones history (2.7; 5.4), egotism (5.6; 3.6), dictatorial (33.3; 3.4), haughty (8.7; 3.3), sleeps on abdomen (3.3; 3.3), intolerance to clothing in abdomen (12.0; 3.1), reproaches (4.0; 3.0), helplessness (24.0; 2.7), fear of failure (10.7; 2.6), irritability on waking in the morning (16.7; 2.5), constipation alternating with diarrhea (8.7; 2.5), intolerant to contradiction (59.3; 2.3), want of self confidence (30.0; 2.4), abdominal distension after eating (23.3; 2.1); ailments from anticipation (32.0; 1.9), irritability before menses (23.3; 1.8), conscientious (26.0; 1.6), desire of sweets (52.0; 1.6), desire of chocolate (16.7; 1.6), lack of vital heat (41.3; 1.3), and flatterer (1.3; ∞). Surveyed homeopaths' intuitive inferences correlated well with symptoms' prevalence but not with their LR. Lycopodium's main symptoms are well known by homeopaths, but their knowledge correlates well with the symptoms' prevalence and not with their LR. Retrospective assessment of prevalence and LR of symptoms in good responders might be a means for better selection of symptoms for prospective studies

1. Searching for degenerate Higgs bosons a profile likelihood ratio method to test for mass-degenerate states in the presence of censored data and uncertainties

CERN Document Server

David, André; Petrucciani, Giovanni

2015-01-01

Using the likelihood ratio test statistic, we present a method which can be employed to test the hypothesis of a single Higgs boson using the matrix of measured signal strengths. This method can be applied in the presence of censored data and takes into account uncertainties on the measurements. The p-value against the hypothesis of a single Higgs boson is defined from the expected distribution of the test statistic, generated using pseudo-experiments. The applicability of the likelihood-based test is demonstrated using numerical examples with uncertainties and missing matrix elements.

2. Adaptive wave filtering for dynamic positioning of marine vessels using maximum likelihood identification: Theory and experiments

Digital Repository Service at National Institute of Oceanography (India)

Hassani, V.; Sorensen, A.J.; Pascoal, A.M.

This paper addresses a filtering problem that arises in the design of dynamic positioning systems for ships and offshore rigs subjected to the influence of sea waves. The dynamic model of the vessel captures explicitly the sea state as an uncertain...

3. Testing for adaptive evolution of the female reproductive protein ZPC in mammals, birds and fishes reveals problems with the M7-M8 likelihood ratio test.

Science.gov (United States)

Berlin, Sofia; Smith, Nick G C

2005-11-10

4. The complex dynamics of wishful thinking: the critical positivity ratio.

Science.gov (United States)

Brown, Nicholas J L; Sokal, Alan D; Friedman, Harris L

2013-12-01

We examine critically the claims made by Fredrickson and Losada (2005) concerning the construct known as the "positivity ratio." We find no theoretical or empirical justification for the use of differential equations drawn from fluid dynamics, a subfield of physics, to describe changes in human emotions over time; furthermore, we demonstrate that the purported application of these equations contains numerous fundamental conceptual and mathematical errors. The lack of relevance of these equations and their incorrect application lead us to conclude that Fredrickson and Losada's claim to have demonstrated the existence of a critical minimum positivity ratio of 2.9013 is entirely unfounded. More generally, we urge future researchers to exercise caution in the use of advanced mathematical tools, such as nonlinear dynamics, and in particular to verify that the elementary conditions for their valid application have been met. PsycINFO Database Record (c) 2013 APA, all rights reserved.

5. Receiver-operating characteristic curves and likelihood ratios: improvements over traditional methods for the evaluation and application of veterinary clinical pathology tests

DEFF Research Database (Denmark)

Gardner, Ian A.; Greiner, Matthias

2006-01-01

Receiver-operating characteristic (ROC) curves provide a cutoff-independent method for the evaluation of continuous or ordinal tests used in clinical pathology laboratories. The area under the curve is a useful overall measure of test accuracy and can be used to compare different tests (or...... different equipment) used by the same tester, as well as the accuracy of different diagnosticians that use the same test material. To date, ROC analysis has not been widely used in veterinary clinical pathology studies, although it should be considered a useful complement to estimates of sensitivity...... and specificity in test evaluation studies. In addition, calculation of likelihood ratios can potentially improve the clinical utility of such studies because likelihood ratios provide an indication of how the post-test probability changes as a function of the magnitude of the test results. For ordinal test...

6. Log-ratio circuit for beam position monitoring

International Nuclear Information System (INIS)

Wells, F.D.; Shafer, R.E.; Gilpatrick, J.D.; Shurter, R.B.

1990-01-01

A synopsis is given of work in progress on a new signal processing technique for obtaining real-time normalized beam position information from sensing electrodes in accelerator beam pipes. The circuit employs wideband logarithmic amplifiers in a configuration that converts pickup electrode signals to position signals that are substantially independent of beam current. The circuit functions as a ratio detector that computes the logarithm of (A/B) as (Log A-Log B), and presents the result in a video (real-time analog) format representing beam position. It has potential benefits of greater dynamic range and better linearity than other techniques currently used and it may be able to operate at substantially higher frequencies. 4 refs., 8 figs

7. Likelihood ratio (razón de verosimilitud): definición y aplicación en Radiología

OpenAIRE

C. Silva Fuente-Alba; M. Molina Villagra

2017-01-01

Los parámetros de medición de validez más conocidos en las pruebas diagnósticas son la sensibilidad, la especificidad y los valores predictivos (positivo y negativo). Las razones de verosimilitud o likelihood ratio conforman otra herramienta que resume la precisión de una prueba diagnóstica, permitiendo un uso clínico más intuitivo.

8. Positive smoking cessation-related interactions with HIV care providers increase the likelihood of interest in cessation among HIV-positive cigarette smokers.

Science.gov (United States)

Pacek, Lauren R; Rass, Olga; Johnson, Matthew W

2017-10-01

Smoking cessation has proven to be a challenge for HIV-positive smokers. Patient and provider characteristics may provide barriers to smoking cessation. We aimed to identify characteristics associated with interest in cessation as well as characterize use of, current interest in, and provider recommendations for smoking cessation modalities. Data came from 275 HIV-positive smokers recruited online. Half (49.1%) of the sample was interested in quitting; daily smoking was associated with decreased likelihood of interest in cessation, whereas making a lifetime quit attempt, receiving encouragement to quit from an HIV care provider, and greater frequency of discussions regarding cessation with HIV care providers were associated with increased likelihood of interest in cessation. Nicotine replacement therapy was the most commonly used (42.9%), generated the most interest (59.1%), and was the most commonly clinician-recommended (70.7%) cessation modality. Findings emphasize the importance of the healthcare provider-patient relationship for smoking cessation promotion in HIV-positive smokers.

9. Likelihood of treatment in a coronary care unit for a first-time myocardial infarction in relation to sex, country of birth and socioeconomic position in Sweden.

Science.gov (United States)

Yang, Dong; James, Stefan; de Faire, Ulf; Alfredsson, Lars; Jernberg, Tomas; Moradi, Tahereh

2013-01-01

To examine the relationship between sex, country of birth, level of education as an indicator of socioeconomic position, and the likelihood of treatment in a coronary care unit (CCU) for a first-time myocardial infarction. Nationwide register based study. Sweden. 199 906 patients (114 387 men and 85,519 women) of all ages who were admitted to hospital for first-time myocardial infarction between 2001 and 2009. Admission to a coronary care unit due to myocardial infarction. Despite the observed increasing access to coronary care units over time, the proportion of women treated in a coronary care unit was 13% less than for men. As compared with men, the multivariable adjusted odds ratio among women was 0.80 (95% confidence interval 0.77 to 0.82). This lower proportion of women treated in a CCU varied by age and year of diagnosis and country of birth. Overall, there was no evidence of a difference in likelihood of treatment in a coronary care unit between Sweden-born and foreign-born patients. As compared with patients with high education, the adjusted odds ratio among patients with a low level of education was 0.93 (95% confidence interval 0.89 to 0.96). Foreign-born and Sweden-born first-time myocardial infarction patients had equal opportunity of being treated in a coronary care unit in Sweden; this is in contrast to the situation in many other countries with large immigrant populations. However, the apparent lower rate of coronary care unit admission after first-time myocardial infarction among women and patients with low socioeconomic position warrants further investigation.

10. Predicting Likelihood of Having Four or More Positive Nodes in Patient With Sentinel Lymph Node-Positive Breast Cancer: A Nomogram Validation Study

International Nuclear Information System (INIS)

Unal, Bulent; Gur, Akif Serhat; Beriwal, Sushil; Tang Gong; Johnson, Ronald; Ahrendt, Gretchen; Bonaventura, Marguerite; Soran, Atilla

2009-01-01

Purpose: Katz suggested a nomogram for predicting having four or more positive nodes in sentinel lymph node (SLN)-positive breast cancer patients. The findings from this formula might influence adjuvant radiotherapy decisions. Our goal was to validate the accuracy of the Katz nomogram. Methods and Materials: We reviewed the records of 309 patients with breast cancer who had undergone completion axillary lymph node dissection. The factors associated with the likelihood of having four or more positive axillary nodes were evaluated in patients with one to three positive SLNs. The nomogram developed by Katz was applied to our data set. The area under the curve of the corresponding receiver operating characteristics curve was calculated for the nomogram. Results: Of the 309 patients, 80 (25.9%) had four or more positive axillary lymph nodes. On multivariate analysis, the number of positive SLNs (p < .0001), overall metastasis size (p = .019), primary tumor size (p = .0001), and extracapsular extension (p = .01) were significant factors predicting for four or more positive nodes. For patients with <5% probability, 90.3% had fewer than four positive nodes and 9.7% had four or more positive nodes. The negative predictive value was 91.7%, and sensitivity was 80%. The nomogram was accurate and discriminating (area under the curve, .801). Conclusion: The probability of four or more involved nodes is significantly greater in patients who have an increased number of positive SLNs, increased overall metastasis size, increased tumor size, and extracapsular extension. The Katz nomogram was validated in our patients. This nomogram will be helpful to clinicians making adjuvant treatment recommendations to their patients.

11. Application of a stratum-specific likelihood ratio analysis in a screen for depression among a community-dwelling population in Japan

Directory of Open Access Journals (Sweden)

Sugawara N

2017-09-01

Full Text Available Norio Sugawara,1,2 Ayako Kaneda,2 Ippei Takahashi,3 Shigeyuki Nakaji,3 Norio Yasui-Furukori2 1Department of Clinical Epidemiology, Translational Medical Center, National Center of Neurology and Psychiatry, Kodaira, Tokyo, 2Department of Neuropsychiatry, Hirosaki University School of Medicine, Hirosaki, 3Department of Social Medicine, Hirosaki University School of Medicine, Hirosaki, Japan Background: Efficient screening for depression is important in community mental health. In this study, we applied a stratum-specific likelihood ratio (SSLR analysis, which is independent of the prevalence of the target disease, to screen for depression among community-dwelling individuals.Method: The Center for Epidemiologic Studies Depression Scale (CES-D and the Mini International Neuropsychiatric Interview (MINI were administered to 789 individuals (19–87 years of age who participated in the Iwaki Health Promotion Project 2011. Major depressive disorder (MDD was assessed using the MINI.Results: For MDD, the SSLRs were 0.13 (95% CI 0.04–0.40, 3.68 (95% CI 1.37–9.89, and 24.77 (95% CI 14.97–40.98 for CES–D scores of 0–16, 17–20, and above 21, respectively.Conclusion: The validity of the CES-D is confirmed, and SSLR analysis is recommended for its practical value for the detection of individuals with the risk of MDD in the Japanese community. Keywords: screening, depression, Center for Epidemiologic Studies Depression Scale, stratum-specific likelihood ratio

12. Speech perception at positive signal-to-noise ratios using adaptive adjustment of time compression.

Science.gov (United States)

Schlueter, Anne; Brand, Thomas; Lemke, Ulrike; Nitzschner, Stefan; Kollmeier, Birger; Holube, Inga

2015-11-01

Positive signal-to-noise ratios (SNRs) characterize listening situations most relevant for hearing-impaired listeners in daily life and should therefore be considered when evaluating hearing aid algorithms. For this, a speech-in-noise test was developed and evaluated, in which the background noise is presented at fixed positive SNRs and the speech rate (i.e., the time compression of the speech material) is adaptively adjusted. In total, 29 younger and 12 older normal-hearing, as well as 24 older hearing-impaired listeners took part in repeated measurements. Younger normal-hearing and older hearing-impaired listeners conducted one of two adaptive methods which differed in adaptive procedure and step size. Analysis of the measurements with regard to list length and estimation strategy for thresholds resulted in a practical method measuring the time compression for 50% recognition. This method uses time-compression adjustment and step sizes according to Versfeld and Dreschler [(2002). J. Acoust. Soc. Am. 111, 401-408], with sentence scoring, lists of 30 sentences, and a maximum likelihood method for threshold estimation. Evaluation of the procedure showed that older participants obtained higher test-retest reliability compared to younger participants. Depending on the group of listeners, one or two lists are required for training prior to data collection.

13. Empirical likelihood

CERN Document Server

Owen, Art B

2001-01-01

Empirical likelihood provides inferences whose validity does not depend on specifying a parametric model for the data. Because it uses a likelihood, the method has certain inherent advantages over resampling methods: it uses the data to determine the shape of the confidence regions, and it makes it easy to combined data from multiple sources. It also facilitates incorporating side information, and it simplifies accounting for censored, truncated, or biased sampling.One of the first books published on the subject, Empirical Likelihood offers an in-depth treatment of this method for constructing confidence regions and testing hypotheses. The author applies empirical likelihood to a range of problems, from those as simple as setting a confidence region for a univariate mean under IID sampling, to problems defined through smooth functions of means, regression models, generalized linear models, estimating equations, or kernel smooths, and to sampling with non-identically distributed data. Abundant figures offer vi...

14. Implementation and assessment of a likelihood ratio approach for the evaluation of LA-ICP-MS evidence in forensic glass analysis.

Science.gov (United States)

van Es, Andrew; Wiarda, Wim; Hordijk, Maarten; Alberink, Ivo; Vergeer, Peter

2017-05-01

For the comparative analysis of glass fragments, a method using Laser Ablation Inductively Coupled Plasma Mass Spectrometry (LA-ICP-MS) is in use at the NFI, giving measurements of the concentration of 18 elements. An important question is how to evaluate the results as evidence that a glass sample originates from a known glass source or from an arbitrary different glass source. One approach is the use of matching criteria e.g. based on a t-test or overlap of confidence intervals. An important drawback of this method is the fact that the rarity of the glass composition is not taken into account. A similar match can have widely different evidential values. In addition the use of fixed matching criteria can give rise to a "fall off the cliff" effect. Small differences may result in a match or a non-match. In this work a likelihood ratio system is presented, largely based on the two-level model as proposed by Aitken and Lucy [1], and Aitken, Zadora and Lucy [2]. Results show that the output from the two-level model gives good discrimination between same and different source hypotheses, but a post-hoc calibration step is necessary to improve the accuracy of the likelihood ratios. Subsequently, the robustness and performance of the LR system are studied. Results indicate that the output of the LR system is robust to the sample properties of the dataset used for calibration. Furthermore, the empirical upper and lower bound method [3], designed to deal with extrapolation errors in the density models, results in minimum and maximum values of the LR outputted by the system of 3.1×10 -3 and 3.4×10 4 . Calibration of the system, as measured by empirical cross-entropy, shows good behavior over the complete prior range. Rates of misleading evidence are small: for same-source comparisons, 0.3% of LRs support a different-source hypothesis; for different-source comparisons, 0.2% supports a same-source hypothesis. The authors use the LR system in reporting of glass cases to

15. Computing power and sample size for case-control association studies with copy number polymorphism: application of mixture-based likelihood ratio test.

Directory of Open Access Journals (Sweden)

Wonkuk Kim

Full Text Available Recent studies suggest that copy number polymorphisms (CNPs may play an important role in disease susceptibility and onset. Currently, the detection of CNPs mainly depends on microarray technology. For case-control studies, conventionally, subjects are assigned to a specific CNP category based on the continuous quantitative measure produced by microarray experiments, and cases and controls are then compared using a chi-square test of independence. The purpose of this work is to specify the likelihood ratio test statistic (LRTS for case-control sampling design based on the underlying continuous quantitative measurement, and to assess its power and relative efficiency (as compared to the chi-square test of independence on CNP counts. The sample size and power formulas of both methods are given. For the latter, the CNPs are classified using the Bayesian classification rule. The LRTS is more powerful than this chi-square test for the alternatives considered, especially alternatives in which the at-risk CNP categories have low frequencies. An example of the application of the LRTS is given for a comparison of CNP distributions in individuals of Caucasian or Taiwanese ethnicity, where the LRTS appears to be more powerful than the chi-square test, possibly due to misclassification of the most common CNP category into a less common category.

16. Fingermark evidence evaluation based on automated fingerprint identification system matching scores: the effect of different types of conditioning on likelihood ratios.

Science.gov (United States)

Alberink, Ivo; de Jongh, Arent; Rodriguez, Crystal

2014-01-01

In recent studies, the evidential value of the similarity of minutiae configurations of fingermarks and fingerprints, for example expressed by automated fingerprint identification systems (AFIS), is determined by likelihood ratios (LRs). The paper explores whether there is an effect on LRs if conditioning takes place on specified fingers, fingerprints, or fingermarks under competing hypotheses: In addition, an approach is explored where conditioning is asymmetric. Comparisons between fingerprints and simulated fingermarks with eight minutiae are performed to produce similarity score distributions for each type of conditioning, given a fixed AFIS matching algorithm. Both similarity scores and LRs are significantly different if the conditioning changes. Given a common-source scenario, "LRs" resulting from asymmetric conditioning are on average higher. The difference may reach a factor of 2000. As conditioning on a suspect's finger(print) is labor-intensive and requires a cooperating suspect, it is recommended to just condition on the number of minutiae in the fingermark. © 2013 American Academy of Forensic Sciences.

17. Evaluation of an Improved Branch-Site Likelihood Method for Detecting Positive Selection at the Molecular Level

DEFF Research Database (Denmark)

Zhang, Jianzhi; Nielsen, Rasmus; Yang, Ziheng

2005-01-01

of interest, while test 2 had acceptable false-positive rates and appeared robust against violations of model assumptions. As test 2 is a direct test of positive selection on the lineages of interest, it is referred to as the branch-site test of positive selection and is recommended for use in real data......Detecting positive Darwinian selection at the DNA sequence level has been a subject of considerable interest. However, positive selection is difficult to detect because it often operates episodically on a few amino acid sites, and the signal may be masked by negative selection. Several methods have...... been developed to test positive selection that acts on given branches (branch methods) or on a subset of sites (site methods). Recently, Yang, Z., and R. Nielsen (2002. Codon-substitution models for detecting molecular adaptation at individual sites along specific lineages. Mol. Biol. Evol. 19...

18. Small Intestinal Bacterial Overgrowth May Increase the Likelihood of Lactose and Sorbitol but not Fructose Intolerance False Positive Diagnosis.

Science.gov (United States)

Perets, Tsachi Tsadok; Hamouda, Dalal; Layfer, Olga; Ashorov, Olga; Boltin, Doron; Levy, Sigal; Niv, Yaron; Dickman, Ram

2017-08-01

Small intestinal bacterial overgrowth (SIBO) is defined as a bacterial count of more than 10 5 colony-forming units per milliliter in duodenal aspirate. It shares many symptoms with carbohydrate intolerance, which makes the clinical distinction of the disorders difficult. The aim of the study was to examine the relationship between a positive carbohydrate breath test and the presence of SIBO suggested by a positive lactulose hydrogen breath test. The electronic database of the gastroenterology laboratory of a tertiary medical center was searched for all patients clinically tested for SIBO in 2012-2013 for whom previous results for lactose, fructose, and/or sorbitol breath test were available. The correlation between positive findings for carbohydrate intolerance and for SIBO was statistically analyzed. The study group included 349 patients, 231 female and 118 male, of mean age 53±19 years. All had symptoms of abdominal bloating and gas. There was a statistically significant difference in rates of a positive breath test for lactose and sorbitol at ≤90 minutes between patients who were positive and negative for SIBO [χ 2 (1)=12.8, p <0.01 and χ 2 (1)=9.5, p <0.01 respectively]. Findings for fructose were not significant. There was no effect of age or gender. SIBO may represent an important reversible cause of carbohydrate intolerance. It may be especially prudent to exclude SIBO patients with an early peak (≤90 minutes) in H 2 excretion. © 2017 by the Association of Clinical Scientists, Inc.

19. Impact of intermediate mammography assessment on the likelihood of false-positive results in breast cancer screening programmes

Energy Technology Data Exchange (ETDEWEB)

Ascunce, Nieves [Public Health Institute, CIBERESP, Navarra Breast Cancer Screening Programme, Pamplona (Spain); Instituto de Salud Publica, Navarra Breast Cancer Screening Programme, Pamplona (Spain); Ederra, Maria; Delfrade, Josu; Erdozain, Nieves [Public Health Institute, CIBERESP, Navarra Breast Cancer Screening Programme, Pamplona (Spain); Baroja, Araceli [Fundacion Rioja Salud, Logrono (Spain); Zubizarreta, Raquel [Public Health and Planning Directorate, Health Office, Galician Breast Cancer Screening Programme, Galicia (Spain); Salas, Dolores [General Directorate Public Health and Centre for Public Health Research (CSISP), Valencia (Spain); Castells, Xavier [Mar Teaching Hospital, CIBERESP, Department of Clinical Epidemiology, Barcelona (Spain)

2012-02-15

Breast cancer screening is offered to 100% of the target population in Spain and intermediate mammograms (IMs) are sometimes indicated. This study was aimed at analysing the frequency of IMs, the factors determining their recommendation, and their impact on the risk of false-positive results and the detection rate. Data from 3,471,307 mammograms from Spanish breast cancer screening programmes were included. 3.36% of the mammograms were IMs. The factors associated with the use of IMs were age, initial screening, previous invasive tests, a familial history of breast cancer and use of hormone replacement therapy. In screening episodes with an IM, the probability of a false-positive result was 13.74% (95% CI: 13.43-14.05), almost double that in episodes without IMs (6.02%, 95% CI 5.99-6.05). In young women with previous invasive procedures, a familial history of breast cancer or hormone replacement therapy use who were undergoing their initial screen, this probability was lower when IMs were performed. IMs always increased the detection rate. The factors prompting IMs should be characterised so that radiologists can systematise their recommendations according to the presence of the factors maximising the benefits and minimising the adverse effects of this procedure. (orig.)

20. Impact of intermediate mammography assessment on the likelihood of false-positive results in breast cancer screening programmes

International Nuclear Information System (INIS)

Ascunce, Nieves; Ederra, Maria; Delfrade, Josu; Erdozain, Nieves; Baroja, Araceli; Zubizarreta, Raquel; Salas, Dolores; Castells, Xavier

2012-01-01

Breast cancer screening is offered to 100% of the target population in Spain and intermediate mammograms (IMs) are sometimes indicated. This study was aimed at analysing the frequency of IMs, the factors determining their recommendation, and their impact on the risk of false-positive results and the detection rate. Data from 3,471,307 mammograms from Spanish breast cancer screening programmes were included. 3.36% of the mammograms were IMs. The factors associated with the use of IMs were age, initial screening, previous invasive tests, a familial history of breast cancer and use of hormone replacement therapy. In screening episodes with an IM, the probability of a false-positive result was 13.74% (95% CI: 13.43-14.05), almost double that in episodes without IMs (6.02%, 95% CI 5.99-6.05). In young women with previous invasive procedures, a familial history of breast cancer or hormone replacement therapy use who were undergoing their initial screen, this probability was lower when IMs were performed. IMs always increased the detection rate. The factors prompting IMs should be characterised so that radiologists can systematise their recommendations according to the presence of the factors maximising the benefits and minimising the adverse effects of this procedure. (orig.)

1. Predictors of positive health in disability pensioners: a population-based questionnaire study using Positive Odds Ratio

Directory of Open Access Journals (Sweden)

Edén Lena

2002-09-01

Full Text Available Abstract Background Determinants of ill-health have been studied far more than determinants of good and improving health. Health promotion measures are important even among individuals with chronic diseases. The aim of this study was to find predictors of positive subjective health among disability pensioners (DPs with musculoskeletal disorders. Methods Two questionnaire surveys were performed among 352 DPs with musculoskeletal disorders. Two groups were defined: DPs with positive health and negative health, respectively. In consequence with the health perspective in this study the conception Positive Odds Ratio was defined and used in the logistic regression analyses instead of the commonly used odds ratio. Results Positive health was associated with age ≥ 55 years, not being an immigrant, not having fibromyalgia as the main diagnosis for granting an early retirement, no regular use of analgesics, a high ADL capacity, a positive subjective health preceding the study period, and good quality of life. Conclusion Positive odds ratio is a concept well adapted to theories of health promotion. It can be used in relation to positive outcomes instead of risks. Suggested health promotion and secondary prevention efforts among individuals with musculoskeletal disorders are 1 to avoid a disability pension for individuals

2. Obtaining reliable likelihood ratio tests from simulated likelihood functions

DEFF Research Database (Denmark)

Andersen, Laura Mørch

2014-01-01

Mixed models: Models allowing for continuous heterogeneity by assuming that value of one or more parameters follow a specified distribution have become increasingly popular. This is known as ‘mixing’ parameters, and it is standard practice by researchers - and the default option in many statistic...

3. Mother's prior intrauterine position affects the sex ratio of her offspring in house mice.

Science.gov (United States)

Vandenbergh, J G; Huggett, C L

1994-11-08

Sex ratio alterations related to environmental factors occur in several mammals, but no mechanism has been identified to explain the adjustment. Intrauterine position (IUP) may provide the context in which such alterations occur. Previous studies on house mice and gerbils reveal that the position of a fetus in the uterus in relation to the sex of its neighbors influences its later anatomy, physiology, and behavior. The anogenital distance (AGD) of females located between two males (2M) is longer than that of females not between two males (OM). We have found that the IUP, as determined by cesarean section and by an index of the AGD, correlates with the sex ratio of the litters produced by female mice. The sex ratio of the first litter born to 2M females was 58% males, for 1M females was 51% males and for OM females was 42% males. The effect on sex ratio continues into the second litter. The number of pups produced by mothers of different IUPs in her first two litters did not differ, suggesting that the sex ratio adjustment occurs prior to parturition. These results provide a basis for the natural variability observed in sex ratios of litter-bearing mammals and suggest that one or more intrauterine mechanisms may be responsible for environmentally related sex ratio alterations.

4. Mother's prior intrauterine position affects the sex ratio of her offspring in house mice.

OpenAIRE

Vandenbergh, J G; Huggett, C L

1994-01-01

Sex ratio alterations related to environmental factors occur in several mammals, but no mechanism has been identified to explain the adjustment. Intrauterine position (IUP) may provide the context in which such alterations occur. Previous studies on house mice and gerbils reveal that the position of a fetus in the uterus in relation to the sex of its neighbors influences its later anatomy, physiology, and behavior. The anogenital distance (AGD) of females located between two males (2M) is lon...

5. Simultaneous fabrication of very high aspect ratio positive nano- to milliscale structures.

Science.gov (United States)

Chen, Long Qing; Chan-Park, Mary B; Zhang, Qing; Chen, Peng; Li, Chang Ming; Li, Sai

2009-05-01

A simple and inexpensive technique for the simultaneous fabrication of positive (i.e., protruding), very high aspect (>10) ratio nanostructures together with micro- or millistructures is developed. The method involves using residual patterns of thin-film over-etching (RPTO) to produce sub-micro-/nanoscale features. The residual thin-film nanopattern is used as an etching mask for Si deep reactive ion etching. The etched Si structures are further reduced in size by Si thermal oxidation to produce amorphous SiO(2), which is subsequently etched away by HF. Two arrays of positive Si nanowalls are demonstrated with this combined RPTO-SiO(2)-HF technique. One array has a feature size of 150 nm and an aspect ratio of 26.7 and another has a feature size of 50 nm and an aspect ratio of 15. No other parallel reduction technique can achieve such a very high aspect ratio for 50-nm-wide nanowalls. As a demonstration of the technique to simultaneously achieve nano- and milliscale features, a simple Si nanofluidic master mold with positive features with dimensions varying continuously from 1 mm to 200 nm and a highest aspect ratio of 6.75 is fabricated; the narrow 200-nm section is 4.5 mm long. This Si master mold is then used as a mold for UV embossing. The embossed open channels are then closed by a cover with glue bonding. A high aspect ratio is necessary to produce unblocked closed channels after the cover bonding process of the nanofluidic chip. The combined method of RPTO, Si thermal oxidation, and HF etching can be used to make complex nanofluidic systems and nano-/micro-/millistructures for diverse applications.

6. Feasibility study of one axis three positions tracking solar PV with low concentration ratio reflector

International Nuclear Information System (INIS)

Huang, B.J.; Sun, F.S.

2007-01-01

A new PV design, called 'one axis three position sun tracking PV module', with low concentration ratio reflector was proposed in the present study. Every PV module is designed with a low concentration ratio reflector and is mounted on an individual sun tracking frame. The one axis tracking mechanism adjusts the PV position only at three fixed angles (three position tracking): morning, noon and afternoon. This 'one axis three position sun tracking PV module' can be designed in a simple structure with low cost. A design analysis was performed in the present study. The analytical results show that the optimal stopping angle β in the morning or afternoon is about 50 o from the solar noon position and the optimal switching angle that controls the best time for changing the attitude of the PV module is half of the stopping angle, i.e. θ H = β/2, and both are independent of the latitude. The power generation increases by approximately 24.5% as compared to a fixed PV module for latitude φ o . The analysis also shows that the effect of installation misalignment away from the true south direction is negligible ( o . An experiment performed in the present study indicates that the PV power generation can increase by about 23% using low concentration (2X) reflectors. Hence, combining with the power output increase of 24.5%, by using one axis three position tracking, the total increase in power generation is about 56%. The economic analysis shows that the price reduction is between 20% and 30% for the various market prices of flat plate PV modules

7. Pathways to Adolescents' Flourishing: Linking Self-Control Skills and Positivity Ratio through Social Support

Science.gov (United States)

Orkibi, Hod; Hamama, Liat; Gavriel-Fried, Belle; Ronen, Tammie

2018-01-01

This study focused on the ability to experience a high ratio of positive to negative emotions in 807 Israeli adolescents aged 12 to 15 years (50% girls). While considering possible gender differences, we tested a model positing that adolescents' self-control skills would link to their positivity ratio and indirectly through perceived social…

8. Likelihood Ratio Based Mixed Resolution Facial Comparison

NARCIS (Netherlands)

Peng, Y.; Spreeuwers, Lieuwe Jan; Veldhuis, Raymond N.J.

2015-01-01

In this paper, we propose a novel method for low-resolution face recognition. It is especially useful for a common situation in forensic search where faces of low resolution, e.g. on surveillance footage or in a crowd, must be compared to a high-resolution reference. This method is based on the

9. Optimism and spontaneous self-affirmation are associated with lower likelihood of cognitive impairment and greater positive affect among cancer survivors

Science.gov (United States)

Taber, Jennifer M.; Klein, William M. P.; Ferrer, Rebecca A.; Kent, Erin E.; Harris, Peter R.

2016-01-01

Background Optimism and self-affirmation promote adaptive coping, goal achievement, and better health. Purpose To examine the associations of optimism and spontaneous self-affirmation (SSA) with physical, mental, and cognitive health and information seeking among cancer survivors. Methods Cancer survivors (n=326) completed the Health Information National Trends Survey 2013, a national survey of U.S. adults. Participants reported optimism, SSA, cognitive and physical impairment, affect, health status, and information seeking. Results Participants higher in optimism reported better health on nearly all indices examined, even when controlling for SSA. Participants higher in SSA reported lower likelihood of cognitive impairment, greater happiness and hopefulness, and greater likelihood of cancer information seeking. SSA remained significantly associated with greater hopefulness and cancer information seeking when controlling for optimism. Conclusions Optimism and SSA may be associated with beneficial health-related outcomes among cancer survivors. Given the demonstrated malleability of self-affirmation, these findings represent important avenues for future research. PMID:26497697

10. Optimism and Spontaneous Self-affirmation are Associated with Lower Likelihood of Cognitive Impairment and Greater Positive Affect among Cancer Survivors.

Science.gov (United States)

Taber, Jennifer M; Klein, William M P; Ferrer, Rebecca A; Kent, Erin E; Harris, Peter R

2016-04-01

Optimism and self-affirmation promote adaptive coping, goal achievement, and better health. The aim of this study is to examine the associations of optimism and spontaneous self-affirmation (SSA) with physical, mental, and cognitive health and information seeking among cancer survivors. Cancer survivors (n = 326) completed the Health Information National Trends Survey 2013, a national survey of US adults. Participants reported optimism, SSA, cognitive and physical impairment, affect, health status, and information seeking. Participants higher in optimism reported better health on nearly all indices examined, even when controlling for SSA. Participants higher in SSA reported lower likelihood of cognitive impairment, greater happiness and hopefulness, and greater likelihood of cancer information seeking. SSA remained significantly associated with greater hopefulness and cancer information seeking when controlling for optimism. Optimism and SSA may be associated with beneficial health-related outcomes among cancer survivors. Given the demonstrated malleability of self-affirmation, these findings represent important avenues for future research.

11. The role of the AR/ER ratio in ER-positive breast cancer patients.

Science.gov (United States)

Rangel, Nelson; Rondon-Lagos, Milena; Annaratone, Laura; Osella-Abate, Simona; Metovic, Jasna; Mano, Maria Piera; Bertero, Luca; Cassoni, Paola; Sapino, Anna; Castellano, Isabella

2018-03-01

The significance of androgen receptor (AR) in breast cancer (BC) management is not fully defined, and it is still ambiguous how the level of AR expression influences oestrogen receptor-positive (ER+) tumours. The aim of the present study was to analyse the prognostic impact of AR/ER ratio, evaluated by immunohistochemistry (IHC), correlating this value with clinical, pathological and molecular characteristics. We retrospectively selected a cohort of 402 ER+BC patients. On each tumour, IHC analyses for AR, ER, PgR, HER2 and Ki67 were performed and AR+ cases were used to calculate the AR/ER value. A cut-off of ≥2 was selected using receiver-operating characteristic (ROC) curve analyses. RNA from 19 cases with AR/ER≥2 was extracted and used for Prosigna-PAM50 assays. Tumours with AR/ER≥2 (6%) showed more frequent metastatic lymph nodes, larger size, higher histological grade and lower PgR levels than cases with AR/ERAR/ER≥2 had worse disease-free interval (DFI) and disease-specific survival (DSS) (hazard ratios (HR) = 4.96 for DFI and HR = 8.69 for DSS, both P  ≤ 0.004). According to the Prosigna-PAM50 assay, 63% (12/19) of these cases resulted in intermediate or high risk of recurrence categories. Additionally, although all samples were positive for ER assessed by IHC, the molecular test assigned 47.4% (9/19) of BCs to intrinsic non-luminal subtypes. In conclusion, the AR/ER ratio ≥2 identifies a subgroup of patients with aggressive biological features and may represent an additional independent marker of worse BC prognosis. Moreover, the Prosigna-PAM50 results indicate that a significant number of cases with AR/ER≥2 could be non-luminal tumours. © 2018 Society for Endocrinology.

12. Angle resolved mass spectrometry of positive ions transmitted through high aspect ratio channels in a radio frequency discharge

NARCIS (Netherlands)

Stoffels - Adamowicz, E.; Stoffels, W.W.; Tachibana, K.; Imai, S.

1997-01-01

The behavior of positive ions in high aspect ratio structures, relevant to the reactive ion etching of deep trenches, has been studied by means of energy resolved mass spectrometry. High aspect ratio trenches are simulated by capillary plates with various aspect ratios. Angle resolved measurements

13. Power calculations for likelihood ratio tests for offspring genotype risks, maternal effects, and parent-of-origin (POO) effects in the presence of missing parental genotypes when unaffected siblings are available.

Science.gov (United States)

Rampersaud, E; Morris, R W; Weinberg, C R; Speer, M C; Martin, E R

2007-01-01

Genotype-based likelihood-ratio tests (LRT) of association that examine maternal and parent-of-origin effects have been previously developed in the framework of log-linear and conditional logistic regression models. In the situation where parental genotypes are missing, the expectation-maximization (EM) algorithm has been incorporated in the log-linear approach to allow incomplete triads to contribute to the LRT. We present an extension to this model which we call the Combined_LRT that incorporates additional information from the genotypes of unaffected siblings to improve assignment of incompletely typed families to mating type categories, thereby improving inference of missing parental data. Using simulations involving a realistic array of family structures, we demonstrate the validity of the Combined_LRT under the null hypothesis of no association and provide power comparisons under varying levels of missing data and using sibling genotype data. We demonstrate the improved power of the Combined_LRT compared with the family-based association test (FBAT), another widely used association test. Lastly, we apply the Combined_LRT to a candidate gene analysis in Autism families, some of which have missing parental genotypes. We conclude that the proposed log-linear model will be an important tool for future candidate gene studies, for many complex diseases where unaffected siblings can often be ascertained and where epigenetic factors such as imprinting may play a role in disease etiology.

14. Comparison of likelihood testing procedures for parallel systems with covariances

International Nuclear Information System (INIS)

Ayman Baklizi; Isa Daud; Noor Akma Ibrahim

1998-01-01

In this paper we considered investigating and comparing the behavior of the likelihood ratio, the Rao's and the Wald's statistics for testing hypotheses on the parameters of the simple linear regression model based on parallel systems with covariances. These statistics are asymptotically equivalent (Barndorff-Nielsen and Cox, 1994). However, their relative performances in finite samples are generally known. A Monte Carlo experiment is conducted to stimulate the sizes and the powers of these statistics for complete samples and in the presence of time censoring. Comparisons of the statistics are made according to the attainment of assumed size of the test and their powers at various points in the parameter space. The results show that the likelihood ratio statistics appears to have the best performance in terms of the attainment of the assumed size of the test. Power comparisons show that the Rao statistic has some advantage over the Wald statistic in almost all of the space of alternatives while likelihood ratio statistic occupies either the first or the last position in term of power. Overall, the likelihood ratio statistic appears to be more appropriate to the model under study, especially for small sample sizes

15. Asymptotic Likelihood Distribution for Correlated & Constrained Systems

CERN Document Server

Agarwal, Ujjwal

2016-01-01

It describes my work as summer student at CERN. The report discusses the asymptotic distribution of the likelihood ratio for total no. of parameters being h and 2 out of these being are constrained and correlated.

16. Positive predictive value of albumin: globulin ratio for feline infectious peritonitis in a mid-western referral hospital population.

Science.gov (United States)

Jeffery, Unity; Deitz, Krysta; Hostetter, Shannon

2012-12-01

Low albumin to globulin ratio has been found previously to have a high positive predictive value for feline infectious peritonitis (FIP) in cats with clinical signs highly suggestive of the disease. However, FIP can have a more vague clinical presentation. This retrospective study found that the positive predictive value of an albumin:globulin (A:G) ratio of <0.8 and <0.6 was only 12.5% and 25%, respectively, in a group of 100 cats with one or more clinical signs consistent with FIP. The negative predictive value was 100% and 99% for an A:G ratio of <0.8 and A:G<0.6%, respectively. Therefore, when the prevalence of FIP is low, the A:G ratio is useful to rule out FIP but is not helpful in making a positive diagnosis of FIP.

17. Likelihood of early detection of breast cancer in relation to false-positive risk in life-time mammographic screening: population-based cohort study.

Science.gov (United States)

Otten, J D M; Fracheboud, J; den Heeten, G J; Otto, S J; Holland, R; de Koning, H J; Broeders, M J M; Verbeek, A L M

2013-10-01

Women require balanced, high-quality information when making an informed decision on screening benefits and harms before attending biennial mammographic screening. The cumulative risk of a false-positive recall and/or (small) screen-detected or interval cancer over 13 consecutive screening examinations for women aged 50 from the start of screening were estimated using data from the Nijmegen programme, the Netherlands. Women who underwent 13 successive screens in the period 1975-1976 had a 5.3% cumulative chance of a screen-detected cancer, with a 4.2% risk of at least one false-positive recall. The risk of being diagnosed with interval cancer was 3.7%. Two decades later, these estimates were 6.9%, 7.3% and 2.9%, respectively. The chance of detection of a small, favourable invasive breast cancer, anticipating a normal life-expectancy, rose from 2.3% to 3.7%. Extrapolation to digital screening mammography indicates that the proportion of false-positive results will rise to 16%. Dutch women about to participate in the screening programme can be reassured that the chance of false-positive recall in the Netherlands is relatively low. A new screening policy and improved mammography have increased the detection of an early screening carcinoma and lowering the risk of interval carcinoma.

18. Effect of anesthesia, positioning, time, and feeding on the proventriculus: keel ratio of clinically healthy parrots.

Science.gov (United States)

Dennison, Sophie E; Paul-Murphy, Joanne R; Yandell, Brian S; Adams, William M

2010-01-01

Healthy, adult Hispaniolan Amazon parrots (Amazona ventralis) were imaged on three occasions to determine the effects of anesthesia, patient rotation, feeding, and short/long-term temporal factors on the proventriculus:keel ratio. Increasing rotation up to 15 degrees from right lateral resulted in increased inability to measure the proventriculus in up to 44% of birds, meaning that the proventriculus:keel ratio could not be calculated from those radiographs. There was a significant difference between the proventriculus:keel ratio for individual parrots when quantified 3 weeks apart. Despite this difference, all ratios remained within normal limits. No significant effect was identified due to anesthesia, feeding, fasting, or repeated imaging through an 8-h period. Interobserver agreement for measurability and correlation for the proventriculus:keel ratio values was high. It is recommended that the proventriculus:keel ratio be calculated from anesthetized parrots to attain images in true lateral recumbency. Ratio fluctuations within the normal range between radiographs obtained on different dates may be observed in normal parrots.

19. Logic of likelihood

International Nuclear Information System (INIS)

Wall, M.J.W.

1992-01-01

The notion of open-quotes probabilityclose quotes is generalized to that of open-quotes likelihood,close quotes and a natural logical structure is shown to exist for any physical theory which predicts likelihoods. Two physically based axioms are given for this logical structure to form an orthomodular poset, with an order-determining set of states. The results strengthen the basis of the quantum logic approach to axiomatic quantum theory. 25 refs

20. The phylogenetic likelihood library.

Science.gov (United States)

Flouri, T; Izquierdo-Carrasco, F; Darriba, D; Aberer, A J; Nguyen, L-T; Minh, B Q; Von Haeseler, A; Stamatakis, A

2015-03-01

We introduce the Phylogenetic Likelihood Library (PLL), a highly optimized application programming interface for developing likelihood-based phylogenetic inference and postanalysis software. The PLL implements appropriate data structures and functions that allow users to quickly implement common, error-prone, and labor-intensive tasks, such as likelihood calculations, model parameter as well as branch length optimization, and tree space exploration. The highly optimized and parallelized implementation of the phylogenetic likelihood function and a thorough documentation provide a framework for rapid development of scalable parallel phylogenetic software. By example of two likelihood-based phylogenetic codes we show that the PLL improves the sequential performance of current software by a factor of 2-10 while requiring only 1 month of programming time for integration. We show that, when numerical scaling for preventing floating point underflow is enabled, the double precision likelihood calculations in the PLL are up to 1.9 times faster than those in BEAGLE. On an empirical DNA dataset with 2000 taxa the AVX version of PLL is 4 times faster than BEAGLE (scaling enabled and required). The PLL is available at http://www.libpll.org under the GNU General Public License (GPL). © The Author(s) 2014. Published by Oxford University Press, on behalf of the Society of Systematic Biologists.

1. Traceability of synthetic drugs by position-specific deuterium isotope ratio analysis

Energy Technology Data Exchange (ETDEWEB)

Brenna, Elisabetta [Dipartimento di Chimica, Materiali e di Ingegneria Chimica del Politecnico di Milano and Istituto CNR per la Chimica del Riconoscimento Molecolare, Via Mancinelli 7, Milan I-20131 (Italy)], E-mail: elisabetta.brenna@polimi.it; Fronza, Giovanni [Dipartimento di Chimica, Materiali e di Ingegneria Chimica del Politecnico di Milano (Italy) and Instituto CNR per la Chimica del Riconoscimento Molecolare, Via Mancinelli 7, Milan I-20131 (Italy)], E-mail: giovanni.fronza@polimi.it; Fuganti, Claudio [Dipartimento di Chimica, Materiali e di Ingegneria Chimica del Politecnico di Milano (Italy) and Istituto CNR per la Chimica del Riconoscimento Molecolare, Via Mancinelli 7, Milan I-20131 (Italy)

2007-10-10

Samples of fluoxetine of different origin were submitted to natural abundance {sup 2}H NMR spectroscopy. The deuterium content at the various sites of the molecule was found to depend on its synthetic history. Hints on the synthetic procedure can be obtained by comparison with standard compounds, whose synthesis is known. These preliminary results give an idea of the potential of site-specific isotope ratio analysis in the fight against patent infringement and drug counterfeiting.

2. Traceability of synthetic drugs by position-specific deuterium isotope ratio analysis

International Nuclear Information System (INIS)

Brenna, Elisabetta; Fronza, Giovanni; Fuganti, Claudio

2007-01-01

Samples of fluoxetine of different origin were submitted to natural abundance 2 H NMR spectroscopy. The deuterium content at the various sites of the molecule was found to depend on its synthetic history. Hints on the synthetic procedure can be obtained by comparison with standard compounds, whose synthesis is known. These preliminary results give an idea of the potential of site-specific isotope ratio analysis in the fight against patent infringement and drug counterfeiting

3. The position dependent influence that sensitivity correction processing gives the signal-to-noise ratio measurement in parallel imaging

International Nuclear Information System (INIS)

Murakami, Koichi; Yoshida, Koji; Yanagimoto, Shinichi

2012-01-01

We studied the position dependent influence that sensitivity correction processing gave the signal-to-noise ratio (SNR) measurement of parallel imaging (PI). Sensitivity correction processing that referred to the sensitivity distribution of the body coil improved regional uniformity more than the sensitivity uniformity correction filter with a fixed correction factor. In addition, the position dependent influence to give the SNR measurement in PI was different from the sensitivity correction processing. Therefore, if we divide SNR of the sensitivity correction processing image by SNR of the original image in each pixel and calculate SNR ratio, we can show the position dependent influence that sensitivity correction processing gives the SNR measurement in PI. It is with an index of the sensitivity correction processing precision. (author)

4. Position of the Patella among Emirati Adult Knees. Is Insall-Salvati Ratio Applicable to Middle-Easterners?

Directory of Open Access Journals (Sweden)

Saeed Althani

2016-04-01

Full Text Available Background: Abnormal patellar height is associated with anterior knee pain and several conditions that affect the patellofemoral joint. The aim of this study was to 1 report the incidence of patella alta and patella baja and 2 investigate whether the normal limits of the Insall-Salvati ratio is applicable in adult Middle-Easterners. Methods: A radiographic review of the lateral radiographs of 736 Middle-Eastern knees were performed. Patellar tendon length (TL and the patellar length (TP was digitally measured and the ratios of these measures was used to calculate the Insall-Salvati ratio. Results: The overall mean TL/PL ratio was 1.20±0.17. The Insall-Salvati ratio was higher (p=0.0013 in males (1.22± 0.12 than in females (1.18±0.17. According to our measurement, the recommended levels for defining abnormal patellar position should be 0.86 for patella baja and 1.54 for patella alta. Conclusion: The use of TL/PL ratio demonstrated a higher incidence of patella alta and a higher mean TL/PL ratio compared to other techniques. The normal ranges for the TL/PL differs from western populations and may be attributed to lifestyle differences.

5. Final results of the cadmium and spectral ratios obtained inside of the fuel rod positioned in the central position of the IPEN/MB-01 nuclear reactor

Energy Technology Data Exchange (ETDEWEB)

Bitelli, Ulysses d' Utra; Mura, Luiz Ernesto C.; Santos, Diogo Feliciano dos, E-mail: ubitelli@ipen.br [Instituto de Pesquisas Energeticas e Nucleares (IPEN/CNEN-SP), Sao Paulo, SP (Brazil); Lambiasi, Beatriz G.N. [Centro Tecnologico da Marinha em Sao Paulo (CTMSP), SP (Brazil)

2015-07-01

The spectral ratios are very important to determine some nuclear reactors parameters such as reaction rates, fuel lifetime, etc and some safety operational conditions. This study aims to determine the spectral ratios in 2 (two) spatial positions located inside the core of the Nuclear Reactor IPEN/MB-01. These places are at the central position of the nuclear reactor core in an asymptotic neutron flux region. The experiment consists in inserting different activation foil detectors inside an experimental fuel rod. The experimental rod is assembled at the central position of the reactor core. Activation neutron foil detectors of different elements such as {sup 197}Au, {sup 238}U, {sup 45}Sc, {sup 58}Ni, {sup 24}Mg, {sup 47}Ti and {sup 115m}In were used to cover a large range of neutron spectrum. Saturation activity per target nucleus was obtained by gamma spectrometry using a HPGe system. The experimental cadmium ratios compared with values computed by MCNP-4C code show good agreement. (author)

6. Cognitive ability correlates positively with son birth and predicts cross-cultural variation of the offspring sex ratio

Science.gov (United States)

2013-06-01

Human populations show remarkable variation in the sex ratio at birth which is believed to be related to the parental condition. In the present study, the global variation of sex ratio at birth (SRB, proportion of male offspring born) was analyzed with respect to indirect measure of condition, the intelligence quotient (IQ). IQ correlates strongly with lifespan across nations, which makes it a good indicator of health of the large populations. Relation between three standard measures of average national IQ and SRB was studied using multiple linear regression models. Average national IQ was positively correlated with SRB ( r = 0.54 to 0.57, p difference in general condition of populations.

7. The modified signed likelihood statistic and saddlepoint approximations

DEFF Research Database (Denmark)

Jensen, Jens Ledet

1992-01-01

SUMMARY: For a number of tests in exponential families we show that the use of a normal approximation to the modified signed likelihood ratio statistic r * is equivalent to the use of a saddlepoint approximation. This is also true in a large deviation region where the signed likelihood ratio...... statistic r is of order √ n. © 1992 Biometrika Trust....

8. Earthquake likelihood model testing

Science.gov (United States)

Schorlemmer, D.; Gerstenberger, M.C.; Wiemer, S.; Jackson, D.D.; Rhoades, D.A.

2007-01-01

INTRODUCTIONThe Regional Earthquake Likelihood Models (RELM) project aims to produce and evaluate alternate models of earthquake potential (probability per unit volume, magnitude, and time) for California. Based on differing assumptions, these models are produced to test the validity of their assumptions and to explore which models should be incorporated in seismic hazard and risk evaluation. Tests based on physical and geological criteria are useful but we focus on statistical methods using future earthquake catalog data only. We envision two evaluations: a test of consistency with observed data and a comparison of all pairs of models for relative consistency. Both tests are based on the likelihood method, and both are fully prospective (i.e., the models are not adjusted to fit the test data). To be tested, each model must assign a probability to any possible event within a specified region of space, time, and magnitude. For our tests the models must use a common format: earthquake rates in specified “bins” with location, magnitude, time, and focal mechanism limits.Seismology cannot yet deterministically predict individual earthquakes; however, it should seek the best possible models for forecasting earthquake occurrence. This paper describes the statistical rules of an experiment to examine and test earthquake forecasts. The primary purposes of the tests described below are to evaluate physical models for earthquakes, assure that source models used in seismic hazard and risk studies are consistent with earthquake data, and provide quantitative measures by which models can be assigned weights in a consensus model or be judged as suitable for particular regions.In this paper we develop a statistical method for testing earthquake likelihood models. A companion paper (Schorlemmer and Gerstenberger 2007, this issue) discusses the actual implementation of these tests in the framework of the RELM initiative.Statistical testing of hypotheses is a common task and a

9. Evidence for positive, but not negative, behavioral contrast with wheel-running reinforcement on multiple variable-ratio schedules.

Science.gov (United States)

Belke, Terry W; Pierce, W David

2016-12-01

Rats responded on a multiple variable-ratio (VR) 10 VR 10 schedule of reinforcement in which lever pressing was reinforced by the opportunity to run in a wheel for 30s in both the changed (manipulated) and unchanged components. To generate positive contrast, the schedule of reinforcement in the changed component was shifted to extinction; to generate negative contrast, the schedule was shifted to VR 3. With the shift to extinction in the changed component, wheel-running and local lever-pressing rates increased in the unchanged component, a result supporting positive contrast; however, the shift to a VR 3 schedule in the changed component showed no evidence of negative contrast in the unaltered setting, only wheel running decreased in the unchanged component. Changes in wheel-running rates across components were consistent in showing a compensation effect, depending on whether the schedule manipulation increased or decreased opportunities for wheel running in the changed component. These findings are the first to demonstrate positive behavioral contrast on a multiple schedule with wheel running as reinforcement in both components. Copyright Â© 2016 Elsevier B.V. All rights reserved.

10. Age and axillary lymph node ratio in postmenopausal women with T1-T2 node positive breast cancer.

Science.gov (United States)

Vinh-Hung, Vincent; Joseph, Sue A; Coutty, Nadege; Ly, Bevan Hong; Vlastos, Georges; Nguyen, Nam Phong

2010-01-01

The purpose of this article was to examine the relationship between age and lymph node ratio (LNR, number of positive nodes divided by number of examined nodes), and to determine their effects on breast cancer (BC) and overall mortality. Women aged ≥50 years, diagnosed in 1988-1997 with a unilateral histologically confirmed T1-T2 node positive surgically treated primary nonmetastatic BC, were selected from the National Cancer Institute's Surveillance, Epidemiology, and End Results (SEER). Generalized Additive Models for Location Scale and Shape (GAMLSS) were used to evaluate the age-LNR relationship. Cumulative incidence functions and multivariate competing risks analysis based on model selection by the Bayesian Information Criterion (BIC) were used to examine the effect of age and LNR on mortality. Low LNR was defined as ≤0.20, mid-LNR 0.21-0.65, and high LNR >0.65. GAMLSS showed a nonlinear LNR-age relationship, increasing from mean LNR 0.26-0.28 at age 50-70 years to 0.30 at 80 years and 0.40 at 90 years. Compared with a 9.8% [95% confidence interval (CI) 8.8%-10.8%] risk of BC death at 5 years in women aged 50-59 years with low LNR, the risk in women ≥80 years with low LNR was 12.6% [95% CI 10.1%-15.0%], mid-LNR 18.1% [13.9%-22.1%], high LNR 29.8% [22.7%-36.1%]. Five-years overall risk of death increased from 40.8% [37.5%-43.9%] by low LNR to 67.4% [61.4%-72.4%] by high LNR. The overall mortality hazard ratio for age ≥80 years with high LNR was 7.49 [6.54-8.59], as compared with women aged 50-59 years with low LNR. High LNR combined with older age was associated with a threefold increased risk of BC death and a sevenfold increased hazard ratio of overall mortality.

11. Composite likelihood estimation of demographic parameters

Directory of Open Access Journals (Sweden)

Garrigan Daniel

2009-11-01

Full Text Available Abstract Background Most existing likelihood-based methods for fitting historical demographic models to DNA sequence polymorphism data to do not scale feasibly up to the level of whole-genome data sets. Computational economies can be achieved by incorporating two forms of pseudo-likelihood: composite and approximate likelihood methods. Composite likelihood enables scaling up to large data sets because it takes the product of marginal likelihoods as an estimator of the likelihood of the complete data set. This approach is especially useful when a large number of genomic regions constitutes the data set. Additionally, approximate likelihood methods can reduce the dimensionality of the data by summarizing the information in the original data by either a sufficient statistic, or a set of statistics. Both composite and approximate likelihood methods hold promise for analyzing large data sets or for use in situations where the underlying demographic model is complex and has many parameters. This paper considers a simple demographic model of allopatric divergence between two populations, in which one of the population is hypothesized to have experienced a founder event, or population bottleneck. A large resequencing data set from human populations is summarized by the joint frequency spectrum, which is a matrix of the genomic frequency spectrum of derived base frequencies in two populations. A Bayesian Metropolis-coupled Markov chain Monte Carlo (MCMCMC method for parameter estimation is developed that uses both composite and likelihood methods and is applied to the three different pairwise combinations of the human population resequence data. The accuracy of the method is also tested on data sets sampled from a simulated population model with known parameters. Results The Bayesian MCMCMC method also estimates the ratio of effective population size for the X chromosome versus that of the autosomes. The method is shown to estimate, with reasonable

12. Preoperative Serum Thyrotropin to Thyroglobulin Ratio Is Effective for Thyroid Nodule Evaluation in Euthyroid Patients.

Science.gov (United States)

Wang, Lina; Li, Hao; Yang, Zhongyuan; Guo, Zhuming; Zhang, Quan

2015-07-01

This study was designed to assess the efficiency of the serum thyrotropin to thyroglobulin ratio for thyroid nodule evaluation in euthyroid patients. Cross-sectional study. Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China. Retrospective analysis was performed for 400 previously untreated cases presenting with thyroid nodules. Thyroid function was tested with commercially available radioimmunoassays. The receiver operating characteristic curves were constructed to determine cutoff values. The efficacy of the thyrotropin:thyroglobulin ratio and thyroid-stimulating hormone for thyroid nodule evaluation was evaluated in terms of sensitivity, specificity, positive predictive value, positive likelihood ratio, negative likelihood ratio, and odds ratio. In receiver operating characteristic curve analysis, the area under the curve was 0.746 for the thyrotropin:thyroglobulin ratio and 0.659 for thyroid-stimulating hormone. With a cutoff point value of 24.97 IU/g for the thyrotropin:thyroglobulin ratio, the sensitivity, specificity, positive predictive value, positive likelihood ratio, and negative likelihood ratio were 78.9%, 60.8%, 75.5%, 2.01, and 0.35, respectively. The odds ratio for the thyrotropin:thyroglobulin ratio indicating malignancy was 5.80. With a cutoff point value of 1.525 µIU/mL for thyroid-stimulating hormone, the sensitivity, specificity, positive predictive value, positive likelihood ratio, and negative likelihood ratio were 74.0%, 53.2%, 70.8%, 1.58, and 0.49, respectively. The odds ratio indicating malignancy for thyroid-stimulating hormone was 3.23. Increasing preoperative serum thyrotropin:thyroglobulin ratio is a risk factor for thyroid carcinoma, and the correlation of the thyrotropin:thyroglobulin ratio to malignancy is higher than that for serum thyroid-stimulating hormone. © American Academy of Otolaryngology—Head and Neck Surgery Foundation 2015.

13. Likelihood devices in spatial statistics

NARCIS (Netherlands)

Zwet, E.W. van

1999-01-01

One of the main themes of this thesis is the application to spatial data of modern semi- and nonparametric methods. Another, closely related theme is maximum likelihood estimation from spatial data. Maximum likelihood estimation is not common practice in spatial statistics. The method of moments

14. Approximate maximum parsimony and ancestral maximum likelihood.

Science.gov (United States)

Alon, Noga; Chor, Benny; Pardi, Fabio; Rapoport, Anat

2010-01-01

We explore the maximum parsimony (MP) and ancestral maximum likelihood (AML) criteria in phylogenetic tree reconstruction. Both problems are NP-hard, so we seek approximate solutions. We formulate the two problems as Steiner tree problems under appropriate distances. The gist of our approach is the succinct characterization of Steiner trees for a small number of leaves for the two distances. This enables the use of known Steiner tree approximation algorithms. The approach leads to a 16/9 approximation ratio for AML and asymptotically to a 1.55 approximation ratio for MP.

15. Risk factors and likelihood of Campylobacter colonization in broiler flocks

Directory of Open Access Journals (Sweden)

SL Kuana

2007-09-01

Full Text Available Campylobacter was investigated in cecal droppings, feces, and cloacal swabs of 22 flocks of 3 to 5 week-old broilers. Risk factors and the likelihood of the presence of this agent in these flocks were determined. Management practices, such as cleaning and disinfection, feeding, drinkers, and litter treatments, were assessed. Results were evaluated using Odds Ratio (OR test, and their significance was tested by Fisher's test (p<0.05. A Campylobacter prevalence of 81.8% was found in the broiler flocks (18/22, and within positive flocks, it varied between 85 and 100%. Campylobacter incidence among sample types was homogenous, being 81.8% in cecal droppings, 80.9% in feces, and 80.4% in cloacal swabs (230. Flocks fed by automatic feeding systems presented higher incidence of Campylobacter as compared to those fed by tube feeders. Litter was reused in 63.6% of the farm, and, despite the lack of statistical significance, there was higher likelihood of Campylobacter incidence when litter was reused. Foot bath was not used in 45.5% of the flocks, whereas the use of foot bath associated to deficient lime management increased the number of positive flocks, although with no statiscal significance. The evaluated parameters were not significantly associated with Campylobacter colonization in the assessed broiler flocks.

16. Extended likelihood inference in reliability

International Nuclear Information System (INIS)

Martz, H.F. Jr.; Beckman, R.J.; Waller, R.A.

1978-10-01

Extended likelihood methods of inference are developed in which subjective information in the form of a prior distribution is combined with sampling results by means of an extended likelihood function. The extended likelihood function is standardized for use in obtaining extended likelihood intervals. Extended likelihood intervals are derived for the mean of a normal distribution with known variance, the failure-rate of an exponential distribution, and the parameter of a binomial distribution. Extended second-order likelihood methods are developed and used to solve several prediction problems associated with the exponential and binomial distributions. In particular, such quantities as the next failure-time, the number of failures in a given time period, and the time required to observe a given number of failures are predicted for the exponential model with a gamma prior distribution on the failure-rate. In addition, six types of life testing experiments are considered. For the binomial model with a beta prior distribution on the probability of nonsurvival, methods are obtained for predicting the number of nonsurvivors in a given sample size and for predicting the required sample size for observing a specified number of nonsurvivors. Examples illustrate each of the methods developed. Finally, comparisons are made with Bayesian intervals in those cases where these are known to exist

17. Borrowing strength : a likelihood ratio test for related sparse signals

NARCIS (Netherlands)

Wit, Ernst C.; Bakewell, David J. G.

2012-01-01

Motivation: Cancer biology is a field where the complexity of the phenomena battles against the availability of data. Often only a few observations per signal source, i.e. genes, are available. Such scenarios are becoming increasingly more relevant as modern sensing technologies generally have no

18. Oestrus Detection in Dairy Cows Using Likelihood Ratio Tests

DEFF Research Database (Denmark)

Jónsson, Ragnar Ingi; Björgvinssin, Trausti; Blanke, Mogens

2008-01-01

This paper addresses detection of oestrus in dairy cows using methods from statistical change detection. The activity of the cows was measured by a necklace attached sensor. Statistical properties of the activity measure were investigated. Using data sets from 17 cows, diurnal activity variations...

19. Multi-Channel Maximum Likelihood Pitch Estimation

DEFF Research Database (Denmark)

2012-01-01

In this paper, a method for multi-channel pitch estimation is proposed. The method is a maximum likelihood estimator and is based on a parametric model where the signals in the various channels share the same fundamental frequency but can have different amplitudes, phases, and noise characteristics....... This essentially means that the model allows for different conditions in the various channels, like different signal-to-noise ratios, microphone characteristics and reverberation. Moreover, the method does not assume that a certain array structure is used but rather relies on a more general model and is hence...

20. Factors Associated with Young Adults’ Pregnancy Likelihood

Science.gov (United States)

Kitsantas, Panagiota; Lindley, Lisa L.; Wu, Huichuan

2014-01-01

1. Positive selection of deleterious alleles through interaction with a sex-ratio suppressor gene in African Buffalo: a plausible new mechanism for a high frequency anomaly.

Directory of Open Access Journals (Sweden)

Pim van Hooft

Full Text Available Although generally rare, deleterious alleles can become common through genetic drift, hitchhiking or reductions in selective constraints. Here we present a possible new mechanism that explains the attainment of high frequencies of deleterious alleles in the African buffalo (Syncerus caffer population of Kruger National Park, through positive selection of these alleles that is ultimately driven by a sex-ratio suppressor. We have previously shown that one in four Kruger buffalo has a Y-chromosome profile that, despite being associated with low body condition, appears to impart a relative reproductive advantage, and which is stably maintained through a sex-ratio suppressor. Apparently, this sex-ratio suppressor prevents fertility reduction that generally accompanies sex-ratio distortion. We hypothesize that this body-condition-associated reproductive advantage increases the fitness of alleles that negatively affect male body condition, causing genome-wide positive selection of these alleles. To investigate this we genotyped 459 buffalo using 17 autosomal microsatellites. By correlating heterozygosity with body condition (heterozygosity-fitness correlations, we found that most microsatellites were associated with one of two gene types: one with elevated frequencies of deleterious alleles that have a negative effect on body condition, irrespective of sex; the other with elevated frequencies of sexually antagonistic alleles that are negative for male body condition but positive for female body condition. Positive selection and a direct association with a Y-chromosomal sex-ratio suppressor are indicated, respectively, by allele clines and by relatively high numbers of homozygous deleterious alleles among sex-ratio suppressor carriers. This study, which employs novel statistical techniques to analyse heterozygosity-fitness correlations, is the first to demonstrate the abundance of sexually-antagonistic genes in a natural mammal population. It also has

2. An Increased Ratio of Glycated Albumin to HbA1c Is Associated with the Degree of Liver Fibrosis in Hepatitis B Virus-Positive Patients

Directory of Open Access Journals (Sweden)

Hirayuki Enomoto

2014-01-01

Full Text Available Background. In hepatitis B virus- (HBV- positive patients, the relationship between the metabolic variables and histological degree of liver fibrosis has been poorly investigated. Methods. A total of 176 HBV-positive patients were assessed in whom the ratios of glycated albumin-to-glycated hemoglobin (GA/HbA1c were calculated in order to investigate the relationship with the degree of liver fibrosis. Results. The GA/HbA1c ratio increased in association with the severity of fibrosis (METAVIR scores: F0-1: 2.61 ± 0.24, F2: 2.65 ± 0.24, F3: 2.74 ± 0.38, and F4: 2.91 ± 0.63. The GA/HbA1c ratios were inversely correlated with four variables of liver function: the prothrombin time (PT percentage (P<0.0001, platelet count (P<0.0001, albumin value (P<0.0001, and cholinesterase value (P<0.0001. The GA/HbA1c ratio was positively correlated with two well-known markers of liver fibrosis, FIB-4 (P<0.0001 and the AST-to-platelet ratio index (APRI (P<0.0001. Furthermore, the GA/HbA1c showed better correlations with two variables of liver function (PT percentage and cholinesterase value than did FIB-4 and with all four variables than did the APRI. Conclusion. The GA/HbA1c ratio is associated with the degree of liver fibrosis in HBV-positive patients.

3. Cracked rocks with positive and negative Poisson's ratio: real-crack properties extracted from pressure dependence of elastic-wave velocities

Science.gov (United States)

Zaitsev, Vladimir Y.; Radostin, Andrey V.; Dyskin, Arcady V.; Pasternak, Elena

2017-04-01

We report results of analysis of literature data on P- and S-wave velocities of rocks subjected to variable hydrostatic pressure. Out of about 90 examined samples, in more than 40% of the samples the reconstructed Poisson's ratios are negative for lowest confining pressure with gradual transition to the conventional positive values at higher pressure. The portion of rocks exhibiting negative Poisson's ratio appeared to be unexpectedly high. To understand the mechanism of negative Poisson's ratio, pressure dependences of P- and S-wave velocities were analyzed using the effective medium model in which the reduction in the elastic moduli due to cracks is described in terms of compliances with respect to shear and normal loading that are imparted to the rock by the presence of cracks. This is in contrast to widely used descriptions of effective cracked medium based on a specific crack model (e.g., penny-shape crack) in which the ratio between normal and shear compliances of such a crack is strictly predetermined. The analysis of pressure-dependences of the elastic wave velocities makes it possible to reveal the ratio between pure normal and shear compliances (called q-ratio below) for real defects and quantify their integral content in the rock. The examination performed demonstrates that a significant portion (over 50%) of cracks exhibit q-ratio several times higher than that assumed for the conventional penny-shape cracks. This leads to faster reduction of the Poisson's ratio with increasing the crack concentration. Samples with negative Poisson's ratio are characterized by elevated q-ratio and simultaneously crack concentration. Our results clearly indicate that the traditional crack model is not adequate for a significant portion of rocks and that the interaction between the opposite crack faces leading to domination of the normal compliance and reduced shear displacement discontinuity can play an important role in the mechanical behavior of rocks.

4. Fractionation in position-specific isotope composition during vaporization of environmental pollutants measured with isotope ratio monitoring by 13C nuclear magnetic resonance spectrometry

International Nuclear Information System (INIS)

Julien, Maxime; Parinet, Julien; Nun, Pierrick; Bayle, Kevin; Höhener, Patrick; Robins, Richard J.; Remaud, Gérald S.

2015-01-01

Isotopic fractionation of pollutants in terrestrial or aqueous environments is a well-recognized means by which to track different processes during remediation. As a complement to the common practice of measuring the change in isotope ratio for the whole molecule using isotope ratio monitoring by mass spectrometry (irm-MS), position-specific isotope analysis (PSIA) can provide further information that can be exploited to investigate source and remediation of soil and water pollutants. Position-specific fractionation originates from either degradative or partitioning processes. We show that isotope ratio monitoring by 13 C NMR (irm- 13 C NMR) spectrometry can be effectively applied to methyl tert-butylether, toluene, ethanol and trichloroethene to obtain this position-specific data for partitioning. It is found that each compound exhibits characteristic position-specific isotope fractionation patterns, and that these are modulated by the type of evaporative process occurring. Such data should help refine models of how remediation is taking place, hence back-tracking to identify pollutant sources. - Highlights: • Position-Specific Isotope Analysis (PSIA) by 13 C NMR spectrometry. • PSIA on isotope fractionation during several vaporization processes. • PSIA for isotope profiling in environment pollutants. • Intramolecular 13 C reveal normal and inverse effects, bulk values being unchanged. - PSIA in pollutants during evaporation processes shows more detailed information for discerning the nature of the process involved than does bulk isotope measurements

5. Ego involvement increases doping likelihood.

Science.gov (United States)

Ring, Christopher; Kavussanu, Maria

2018-08-01

Achievement goal theory provides a framework to help understand how individuals behave in achievement contexts, such as sport. Evidence concerning the role of motivation in the decision to use banned performance enhancing substances (i.e., doping) is equivocal on this issue. The extant literature shows that dispositional goal orientation has been weakly and inconsistently associated with doping intention and use. It is possible that goal involvement, which describes the situational motivational state, is a stronger determinant of doping intention. Accordingly, the current study used an experimental design to examine the effects of goal involvement, manipulated using direct instructions and reflective writing, on doping likelihood in hypothetical situations in college athletes. The ego-involving goal increased doping likelihood compared to no goal and a task-involving goal. The present findings provide the first evidence that ego involvement can sway the decision to use doping to improve athletic performance.

6. Study of Branching Ratio And Polarization Fraction in Neutral B Meson Decays to Negative Rho Meson Positive Kaon Resonance

Energy Technology Data Exchange (ETDEWEB)

Cheng, Baosen; /Wisconsin U., Madison

2006-03-07

We present the preliminary results on the search for B{sup 0} {yields} {rho}{sup -}K*{sup +}. The data sample comprises 122.7 million B{bar B} pairs in the e{sup +}e{sup -} annihilation through the {Upsilon}(4S) resonance collected during 1999-2003 with the BABAR detector at the PEP-II asymmetric-energy collider at Stanford Linear Accelerator Center (SLAC). We obtain an upper limit of the branching ratio at 90% confidence level as {Beta}(B{sup 0} {yields} {rho}{sup -}K*{sup +}) < 17.2 x 10{sup -6}. The fitted result on the polarization fraction shows no evidence that the decay is longitudinally dominated as predicted by various theoretical models.

7. Legume finishing provides beef with positive human dietary fatty acid ratios and consumer preference comparable with grain-finished beef.

Science.gov (United States)

Chail, A; Legako, J F; Pitcher, L R; Griggs, T C; Ward, R E; Martini, S; MacAdam, J W

2016-05-01

Consumer liking, proximate composition, pH, Warner-Bratzler shear force, fatty acid composition, and volatile compounds were determined from the LM (longissimus thoracis) of cattle ( = 6 per diet) finished on conventional feedlot (USUGrain), legume, and grass forage diets. Forage diets included a condensed tannin-containing perennial legume, birdsfoot trefoil (; USUBFT), and a grass, meadow brome ( Rehmann; USUGrass). Moreover, representative retail forage (USDA Certified Organic Grass-fed [OrgGrass]) and conventional beef (USDA Choice, Grain-fed; ChGrain) were investigated ( = 6 per retail type). The ChGrain had the greatest ( 0.05) to that of both USUGrain and USUGrass. Both grain-finished beef treatments were rated greater ( Consumer liking of USUBFT beef tenderness, fattiness, and overall liking were comparable ( > 0.05) with that of USUGrain and ChGrain. Flavor liking was rated greatest ( 0.05) to those of ChGrain, USUGrass, and OrgGrass. Cumulative SFA and MUFA concentrations were greatest ( 0.05) to those of USUGrain and USUGrass. Each forage-finished beef treatment, USUGrass, OrgGrass, and USUBFT, had lower ( < 0.001) ratios of -6:-3 fatty acids. Hexanal was the most numerically abundant volatile compound. The concentration of hexanal increased with increasing concentrations of total PUFA. Among all the lipid degradation products (aldehydes, alcohols, furans, carboxylic acids, and ketones) measured in this study, there was an overall trend toward greater quantities in grain-finished products, lower quantities in USUGrass and OrgGrass, and intermediate quantities in USUBFT. This trend was in agreement with IMF content, fatty acid concentrations, and sensory attributes. These results suggest an opportunity for a birdsfoot trefoil finishing program, which results in beef comparable in sensory quality with grain-finished beef but with reduced -6 and SFA, similar to grass-finished beef.

8. Likelihood estimators for multivariate extremes

KAUST Repository

Huser, Raphaë l; Davison, Anthony C.; Genton, Marc G.

2015-01-01

The main approach to inference for multivariate extremes consists in approximating the joint upper tail of the observations by a parametric family arising in the limit for extreme events. The latter may be expressed in terms of componentwise maxima, high threshold exceedances or point processes, yielding different but related asymptotic characterizations and estimators. The present paper clarifies the connections between the main likelihood estimators, and assesses their practical performance. We investigate their ability to estimate the extremal dependence structure and to predict future extremes, using exact calculations and simulation, in the case of the logistic model.

9. Likelihood estimators for multivariate extremes

KAUST Repository

Huser, Raphaël

2015-11-17

The main approach to inference for multivariate extremes consists in approximating the joint upper tail of the observations by a parametric family arising in the limit for extreme events. The latter may be expressed in terms of componentwise maxima, high threshold exceedances or point processes, yielding different but related asymptotic characterizations and estimators. The present paper clarifies the connections between the main likelihood estimators, and assesses their practical performance. We investigate their ability to estimate the extremal dependence structure and to predict future extremes, using exact calculations and simulation, in the case of the logistic model.

10. Meta-analysis of amino acid stable nitrogen isotope ratios for estimating trophic position in marine organisms.

Science.gov (United States)

Nielsen, Jens M; Popp, Brian N; Winder, Monika

2015-07-01

Estimating trophic structures is a common approach used to retrieve information regarding energy pathways, predation, and competition in complex ecosystems. The application of amino acid (AA) compound-specific nitrogen (N) isotope analysis (CSIA) is a relatively new method used to estimate trophic position (TP) and feeding relationships in diverse organisms. Here, we conducted the first meta-analysis of δ(15)N AA values from measurements of 359 marine species covering four trophic levels, and compared TP estimates from AA-CSIA to literature values derived from food items, gut or stomach content analysis. We tested whether the AA trophic enrichment factor (TEF), or the (15)N enrichment among different individual AAs is constant across trophic levels and whether inclusion of δ(15)N values from multiple AAs improves TP estimation. For the TEF of glutamic acid relative to phenylalanine (Phe) we found an average value of 6.6‰ across all taxa, which is significantly lower than the commonly applied 7.6‰. We found that organism feeding ecology influences TEF values of several trophic AAs relative to Phe, with significantly higher TEF values for herbivores compared to omnivores and carnivores, while TEF values were also significantly lower for animals excreting urea compared to ammonium. Based on the comparison of multiple model structures using the metadata of δ(15)N AA values we show that increasing the number of AAs in principle improves precision in TP estimation. This meta-analysis clarifies the advantages and limitations of using individual δ(15)N AA values as tools in trophic ecology and provides a guideline for the future application of AA-CSIA to food web studies.

11. Maximum likelihood of phylogenetic networks.

Science.gov (United States)

Jin, Guohua; Nakhleh, Luay; Snir, Sagi; Tuller, Tamir

2006-11-01

Horizontal gene transfer (HGT) is believed to be ubiquitous among bacteria, and plays a major role in their genome diversification as well as their ability to develop resistance to antibiotics. In light of its evolutionary significance and implications for human health, developing accurate and efficient methods for detecting and reconstructing HGT is imperative. In this article we provide a new HGT-oriented likelihood framework for many problems that involve phylogeny-based HGT detection and reconstruction. Beside the formulation of various likelihood criteria, we show that most of these problems are NP-hard, and offer heuristics for efficient and accurate reconstruction of HGT under these criteria. We implemented our heuristics and used them to analyze biological as well as synthetic data. In both cases, our criteria and heuristics exhibited very good performance with respect to identifying the correct number of HGT events as well as inferring their correct location on the species tree. Implementation of the criteria as well as heuristics and hardness proofs are available from the authors upon request. Hardness proofs can also be downloaded at http://www.cs.tau.ac.il/~tamirtul/MLNET/Supp-ML.pdf

12. The scrape-off layer in a finite-aspect-ratio Torus: The influence of limiter position

International Nuclear Information System (INIS)

Harbour, P.J.; Loarte, A.

1995-01-01

The effect on the scrape-off layer (SOL) of changing the position of a tokamak limiter from the low field side (LFS) to the high field side (HFS) of the plasma is considered. Conservation of magnetics flux in the SOL requires that the area, A parallel , for flow of particles and energy parallel to the magnetic field, be smaller on the HFS. The effect that this reduction in A parallel has on the SOL is that, when the limiter is on the HFS, as compared with the LFS, then the characteristic scrape-off thickness is increased and the plasma particle and energy densities are higher. Such increases are described. They were observed in an experiment in JET in which the plasma was limited first on the LFS and then, during the same discharge, on the HFS. The magnitude of the effect was larger than expected, with the line integrated density of the SOL, nλ n , having an average increase by a factor of 5.8 ± 1.1, and an increase by a factor of 7 at the respective limiter. Moreover, there was an increase in the characteristic scrape-off thickness for power flow, λ p , by a factor of 2.5 ± 0.4 on average, and a corresponding increase by a factor of 3.0 at the respective limiter. Such an increase in nλ n can substantially improve the screening of recycled particles and can increase wall pumping, although not strongly in the experiments described, which were in helium to avoid these effects. The large increase in λ p measured in JET is useful in spreading power over a larger area of a limiter on the HFS. Evidence is presented which suggests that either the observed enhancement of the effect on the SOL above expectations is due to a reduction in the Mach number at the limiter sheath when on the HFS or there are fundamental problems with the modelling procedure usually used for the SOL. Related experiments are discussed. (author). 29 refs, 3 figs, 3 tabs

13. A simulation study of likelihood inference procedures in rayleigh distribution with censored data

International Nuclear Information System (INIS)

Baklizi, S. A.; Baker, H. M.

2001-01-01

Inference procedures based on the likelihood function are considered for the one parameter Rayleigh distribution with type1 and type 2 censored data. Using simulation techniques, the finite sample performances of the maximum likelihood estimator and the large sample likelihood interval estimation procedures based on the Wald, the Rao, and the likelihood ratio statistics are investigated. It appears that the maximum likelihood estimator is unbiased. The approximate variance estimates obtained from the asymptotic normal distribution of the maximum likelihood estimator are accurate under type 2 censored data while they tend to be smaller than the actual variances when considering type1 censored data of small size. It appears also that interval estimation based on the Wald and Rao statistics need much more sample size than interval estimation based on the likelihood ratio statistic to attain reasonable accuracy. (authors). 15 refs., 4 tabs

14. The influence of the negative-positive ratio and screening database size on the performance of machine learning-based virtual screening.

Science.gov (United States)

Kurczab, Rafał; Bojarski, Andrzej J

2017-01-01

The machine learning-based virtual screening of molecular databases is a commonly used approach to identify hits. However, many aspects associated with training predictive models can influence the final performance and, consequently, the number of hits found. Thus, we performed a systematic study of the simultaneous influence of the proportion of negatives to positives in the testing set, the size of screening databases and the type of molecular representations on the effectiveness of classification. The results obtained for eight protein targets, five machine learning algorithms (SMO, Naïve Bayes, Ibk, J48 and Random Forest), two types of molecular fingerprints (MACCS and CDK FP) and eight screening databases with different numbers of molecules confirmed our previous findings that increases in the ratio of negative to positive training instances greatly influenced most of the investigated parameters of the ML methods in simulated virtual screening experiments. However, the performance of screening was shown to also be highly dependent on the molecular library dimension. Generally, with the increasing size of the screened database, the optimal training ratio also increased, and this ratio can be rationalized using the proposed cost-effectiveness threshold approach. To increase the performance of machine learning-based virtual screening, the training set should be constructed in a way that considers the size of the screening database.

15. Phylogeny, ecology, and heart position in snakes

DEFF Research Database (Denmark)

Gartner, Gabriel E.A.; Hicks, James W.; Manzani, Paulo R.

2010-01-01

both conventional and phylogenetically based statistical methods. General linear models regressing log10 snout‐heart position on log10 snout‐vent length (SVL), as well as dummy variables coding for habitat and/or clade, were compared using likelihood ratio tests and the Akaike Information Criterion...

16. Constraint likelihood analysis for a network of gravitational wave detectors

International Nuclear Information System (INIS)

Klimenko, S.; Rakhmanov, M.; Mitselmakher, G.; Mohanty, S.

2005-01-01

We propose a coherent method for detection and reconstruction of gravitational wave signals with a network of interferometric detectors. The method is derived by using the likelihood ratio functional for unknown signal waveforms. In the likelihood analysis, the global maximum of the likelihood ratio over the space of waveforms is used as the detection statistic. We identify a problem with this approach. In the case of an aligned pair of detectors, the detection statistic depends on the cross correlation between the detectors as expected, but this dependence disappears even for infinitesimally small misalignments. We solve the problem by applying constraints on the likelihood functional and obtain a new class of statistics. The resulting method can be applied to data from a network consisting of any number of detectors with arbitrary detector orientations. The method allows us reconstruction of the source coordinates and the waveforms of two polarization components of a gravitational wave. We study the performance of the method with numerical simulations and find the reconstruction of the source coordinates to be more accurate than in the standard likelihood method

17. Improved protection system for phase faults on marine vessels based on ratio between negative sequence and positive sequence of the fault current

DEFF Research Database (Denmark)

Ciontea, Catalin-Iosif; Hong, Qiteng; Booth, Campbell

2018-01-01

algorithm is implemented in a programmable digital relay embedded in a hardware-in-the-loop (HIL) test set-up that emulates a typical maritime feeder using a real-time digital simulator. The HIL set-up allows testing of the new protection method under a wide range of faults and network conditions......This study presents a new method to protect the radial feeders on marine vessels. The proposed protection method is effective against phase–phase (PP) faults and is based on evaluation of the ratio between the negative sequence and positive sequence of the fault currents. It is shown...... that the magnitude of the introduced ratio increases significantly during the PP fault, hence indicating the fault presence in an electric network. Here, the theoretical background of the new method of protection is firstly discussed, based on which the new protection algorithm is described afterwards. The proposed...

18. Identifying Malignant Pleural Effusion by A Cancer Ratio (Serum LDH: Pleural Fluid ADA Ratio).

Science.gov (United States)

Verma, Akash; Abisheganaden, John; Light, R W

2016-02-01

We studied the diagnostic potential of serum lactate dehydrogenase (LDH) in malignant pleural effusion. Retrospective analysis of patients hospitalized with exudative pleural effusion in 2013. Serum LDH and serum LDH: pleural fluid ADA ratio was significantly higher in cancer patients presenting with exudative pleural effusion. In multivariate logistic regression analysis, pleural fluid ADA was negatively correlated 0.62 (0.45-0.85, p = 0.003) with malignancy, whereas serum LDH 1.02 (1.0-1.03, p = 0.004) and serum LDH: pleural fluid ADA ratio 0.94 (0.99-1.0, p = 0.04) was correlated positively with malignant pleural effusion. For serum LDH: pleural fluid ADA ratio, a cut-off level of >20 showed sensitivity, specificity of 0.98 (95 % CI 0.92-0.99) and 0.94 (95 % CI 0.83-0.98), respectively. The positive likelihood ratio was 32.6 (95 % CI 10.7-99.6), while the negative likelihood ratio at this cut-off was 0.03 (95 % CI 0.01-0.15). Higher serum LDH and serum LDH: pleural fluid ADA ratio in patients presenting with exudative pleural effusion can distinguish between malignant and non-malignant effusion on the first day of hospitalization. The cut-off level for serum LDH: pleural fluid ADA ratio of >20 is highly predictive of malignancy in patients with exudative pleural effusion (whether lymphocytic or neutrophilic) with high sensitivity and specificity.

19. LDR: A Package for Likelihood-Based Sufficient Dimension Reduction

Directory of Open Access Journals (Sweden)

R. Dennis Cook

2011-03-01

Full Text Available We introduce a new mlab software package that implements several recently proposed likelihood-based methods for sufficient dimension reduction. Current capabilities include estimation of reduced subspaces with a fixed dimension d, as well as estimation of d by use of likelihood-ratio testing, permutation testing and information criteria. The methods are suitable for preprocessing data for both regression and classification. Implementations of related estimators are also available. Although the software is more oriented to command-line operation, a graphical user interface is also provided for prototype computations.

20. Evaluation of the generalized likelihood ratio test of hypotheses on the genetic control of continuous traits/ Avaliação da razão de verossimilhança generalizada em teste de hipóteses sobre o controle genético de características contínuas

Directory of Open Access Journals (Sweden)

2004-05-01

1. Generalized empirical likelihood methods for analyzing longitudinal data

KAUST Repository

Wang, S.

2010-02-16

Efficient estimation of parameters is a major objective in analyzing longitudinal data. We propose two generalized empirical likelihood based methods that take into consideration within-subject correlations. A nonparametric version of the Wilks theorem for the limiting distributions of the empirical likelihood ratios is derived. It is shown that one of the proposed methods is locally efficient among a class of within-subject variance-covariance matrices. A simulation study is conducted to investigate the finite sample properties of the proposed methods and compare them with the block empirical likelihood method by You et al. (2006) and the normal approximation with a correctly estimated variance-covariance. The results suggest that the proposed methods are generally more efficient than existing methods which ignore the correlation structure, and better in coverage compared to the normal approximation with correctly specified within-subject correlation. An application illustrating our methods and supporting the simulation study results is also presented.

2. The MRC1/CD68 ratio is positively associated with adipose tissue lipogenesis and with muscle mitochondrial gene expression in humans.

Directory of Open Access Journals (Sweden)

José María Moreno-Navarrete

Full Text Available BACKGROUND: Alternative macrophages (M2 express the cluster differentiation (CD 206 (MCR1 at high levels. Decreased M2 in adipose tissue is known to be associated with obesity and inflammation-related metabolic disturbances. Here we aimed to investigate MCR1 relative to CD68 (total macrophages gene expression in association with adipogenic and mitochondrial genes, which were measured in human visceral [VWAT, n = 147] and subcutaneous adipose tissue [SWAT, n = 76] and in rectus abdominis muscle (n = 23. The effects of surgery-induced weight loss were also longitudinally evaluated (n = 6. RESULTS: MCR1 and CD68 gene expression levels were similar in VWAT and SWAT. A higher proportion of CD206 relative to total CD68 was present in subjects with less body fat and lower fasting glucose concentrations. The ratio MCR1/CD68was positively associated with IRS1gene expression and with the expression of lipogenic genes such as ACACA, FASN and THRSP, even after adjusting for BMI. The ratio MCR1/CD68 in SWAT increased significantly after the surgery-induced weight loss (+44.7%; p = 0.005 in parallel to the expression of adipogenic genes. In addition, SWAT MCR1/CD68ratio was significantly associated with muscle mitochondrial gene expression (PPARGC1A, TFAM and MT-CO3. AT CD206 was confirmed by immunohistochemistry to be specific of macrophages, especially abundant in crown-like structures. CONCLUSION: A decreased ratio MCR1/CD68 is linked to adipose tissue and muscle mitochondrial dysfunction at least at the level of expression of adipogenic and mitochondrial genes.

3. Smoking increases the likelihood of Helicobacter pylori treatment failure.

Science.gov (United States)

Itskoviz, David; Boltin, Doron; Leibovitzh, Haim; Tsadok Perets, Tsachi; Comaneshter, Doron; Cohen, Arnon; Niv, Yaron; Levi, Zohar

2017-07-01

4. MXLKID: a maximum likelihood parameter identifier

International Nuclear Information System (INIS)

Gavel, D.T.

1980-07-01

MXLKID (MaXimum LiKelihood IDentifier) is a computer program designed to identify unknown parameters in a nonlinear dynamic system. Using noisy measurement data from the system, the maximum likelihood identifier computes a likelihood function (LF). Identification of system parameters is accomplished by maximizing the LF with respect to the parameters. The main body of this report briefly summarizes the maximum likelihood technique and gives instructions and examples for running the MXLKID program. MXLKID is implemented LRLTRAN on the CDC7600 computer at LLNL. A detailed mathematical description of the algorithm is given in the appendices. 24 figures, 6 tables

5. Testing for HPV as an objective measure for quality assurance in gynecologic cytology: positive rates in equivocal and abnormal specimens and comparison with the ASCUS to SIL ratio.

Science.gov (United States)

Ko, Vincent; Nanji, Shabin; Tambouret, Rosemary H; Wilbur, David C

2007-04-25

Inappropriate use of the category of atypical squamous cells of undetermined significance (ASCUS) can result in overtreatment or undertreatment of patients, which may decrease the cost effectiveness of screening. Quality assurance tools, such as the ASCUS to squamous intraepithelial lesion ratio (ASCUS:SIL) and case review, are imperfect. High-risk HPV (hrHPV) testing is an objective test for a known viral carcinogen, and hrHPV may be more useful in monitoring the quality of ASCUS interpretations. hrHPV rates for cytologic diagnoses and patient age groups were calculated for a 2-year period. All hrHPV results for ASCUS and SIL over a 17-month period were analyzed by patient age group, over time, and by individual cytopathologist to compare hrHPV rates with the corresponding ASCUS:SIL. The hrHPV positive rate for SIL was >90%, and it was 32.6% for ASCUS. Stratification by patient age showed that approximately 50% of patients younger than 30 years and older than 70 years of age were hrHPV positive, whereas other patients had a lower rate ranging from 14% to 34%. The overall ASCUS:SIL was 1.42, and the overall hrHPV positive rate was 39.9%. Over time and by individual cytopathologist, the hrHPV rate performed similarly to the ASCUS:SIL. The analysis by patient age showed a high statistical correlation (R(2) = 0.9772) between the 2 methods. Despite differences between these techniques, the hrHPV rate closely recapitulates the ASCUS:SIL. When used together, the 2 methods can complement each other. The desirable hrHPV-positive range appears to be 40% to 50%; however, this may vary based on the patient population. The hrHPV rate is as quick and cost effective as determining the ASCUS:SIL. (c) 2007 American Cancer Society.

6. Likelihood Analysis of Supersymmetric SU(5) GUTs

CERN Document Server

Bagnaschi, E.

2017-01-01

We perform a likelihood analysis of the constraints from accelerator experiments and astrophysical observations on supersymmetric (SUSY) models with SU(5) boundary conditions on soft SUSY-breaking parameters at the GUT scale. The parameter space of the models studied has 7 parameters: a universal gaugino mass $m_{1/2}$, distinct masses for the scalar partners of matter fermions in five- and ten-dimensional representations of SU(5), $m_5$ and $m_{10}$, and for the $\\mathbf{5}$ and $\\mathbf{\\bar 5}$ Higgs representations $m_{H_u}$ and $m_{H_d}$, a universal trilinear soft SUSY-breaking parameter $A_0$, and the ratio of Higgs vevs $\\tan \\beta$. In addition to previous constraints from direct sparticle searches, low-energy and flavour observables, we incorporate constraints based on preliminary results from 13 TeV LHC searches for jets + MET events and long-lived particles, as well as the latest PandaX-II and LUX searches for direct Dark Matter detection. In addition to previously-identified mechanisms for bringi...

7. Essays on empirical likelihood in economics

NARCIS (Netherlands)

Gao, Z.

2012-01-01

This thesis intends to exploit the roots of empirical likelihood and its related methods in mathematical programming and computation. The roots will be connected and the connections will induce new solutions for the problems of estimation, computation, and generalization of empirical likelihood.

8. Cases in which ancestral maximum likelihood will be confusingly misleading.

Science.gov (United States)

Handelman, Tomer; Chor, Benny

2017-05-07

Ancestral maximum likelihood (AML) is a phylogenetic tree reconstruction criteria that "lies between" maximum parsimony (MP) and maximum likelihood (ML). ML has long been known to be statistically consistent. On the other hand, Felsenstein (1978) showed that MP is statistically inconsistent, and even positively misleading: There are cases where the parsimony criteria, applied to data generated according to one tree topology, will be optimized on a different tree topology. The question of weather AML is statistically consistent or not has been open for a long time. Mossel et al. (2009) have shown that AML can "shrink" short tree edges, resulting in a star tree with no internal resolution, which yields a better AML score than the original (resolved) model. This result implies that AML is statistically inconsistent, but not that it is positively misleading, because the star tree is compatible with any other topology. We show that AML is confusingly misleading: For some simple, four taxa (resolved) tree, the ancestral likelihood optimization criteria is maximized on an incorrect (resolved) tree topology, as well as on a star tree (both with specific edge lengths), while the tree with the original, correct topology, has strictly lower ancestral likelihood. Interestingly, the two short edges in the incorrect, resolved tree topology are of length zero, and are not adjacent, so this resolved tree is in fact a simple path. While for MP, the underlying phenomenon can be described as long edge attraction, it turns out that here we have long edge repulsion. Copyright © 2017. Published by Elsevier Ltd.

9. Bilirubin concentration is positively associated with haemoglobin concentration and inversely associated with albumin to creatinine ratio among Indigenous Australians: eGFR Study.

Science.gov (United States)

Hughes, J T; Barzi, F; Hoy, W E; Jones, G R D; Rathnayake, G; Majoni, S W; Thomas, M A B; Sinha, A; Cass, A; MacIsaac, R J; O'Dea, K; Maple-Brown, L J

2017-12-01

Low serum bilirubin concentrations are reported to be strongly associated with cardio-metabolic disease, but this relationship has not been reported among Indigenous Australian people who are known to be at high risk for diabetes and chronic kidney disease (CKD). serum bilirubin will be negatively associated with markers of chronic disease, including CKD and anaemia among Indigenous Australians. A cross-sectional analysis of 594 adult Aboriginal and Torres Strait Islander (TSI) people in good health or with diabetes and markers of CKD. Measures included urine albumin: creatinine ratio (ACR), estimated glomerular filtration rate (eGFR), haemoglobin (Hb) and glycated haemoglobin (HbA1c). Diabetes was defined by medical history, medications or HbA1c≥6.5% or ≥48mmol/mol. Anaemia was defined as Hbbilirubin was performed. Participants mean (SD) age was 45.1 (14.5) years, and included 62.5% females, 71.7% Aboriginal, 41.1% with diabetes, 16.7% with anaemia, 41% with ACR>3mg/mmol and 18.2% with eGFRbilirubin concentration was lower in females than males (6 v 8μmol/L, pbilirubin; Hb and cholesterol (both positively related) and ACR, triglycerides, Aboriginal ethnicity and female gender (all inversely related). Serum bilirubin concentrations were positively associated with Hb and total cholesterol, and inversely associated with ACR. Further research to determine reasons explaining lower bilirubin concentrations among Aboriginal compared with TSI participants are needed. Copyright © 2017 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.

10. Enhanced forensic discrimination of pollutants by position-specific isotope analysis using isotope ratio monitoring by (13)C nuclear magnetic resonance spectrometry.

Science.gov (United States)

Julien, Maxime; Nun, Pierrick; Höhener, Patrick; Parinet, Julien; Robins, Richard J; Remaud, Gérald S

2016-01-15

In forensic environmental investigations the main issue concerns the inference of the original source of the pollutant for determining the liable party. Isotope measurements in geochemistry, combined with complimentary techniques for contaminant identification, have contributed significantly to source determination at polluted sites. In this work we have determined the intramolecular (13)C profiles of several molecules well-known as pollutants. By giving additional analytical parameters, position-specific isotope analysis performed by isotope ratio monitoring by (13)C nuclear magnetic resonance (irm-(13)C NMR) spectrometry gives new information to help in answering the major question: what is the origin of the detected contaminant? We have shown that isotope profiling of the core of a molecule reveals both the raw materials and the process used in its manufacture. It also can reveal processes occurring between the contamination site 'source' and the sampling site. Thus, irm-(13)C NMR is shown to be a very good complement to compound-specific isotope analysis currently performed by mass spectrometry for assessing polluted sites involving substantial spills of pollutant. Copyright © 2015 Elsevier B.V. All rights reserved.

11. Rate of transformation and normal range about cardiac size and cardiothoracic ratio according to patient position and age at chest radiography of Korean adult man

Energy Technology Data Exchange (ETDEWEB)

Joo, Young Cheol [Dept. of Radiology, Samsung Medical Center, Seoul (Korea, Republic of); Lim, Cheong Hwan; Jung, Hong Ryang [Dept. of Radiological Science, Hanseo University, Seosan (Korea, Republic of); Kim, Yun Min [Dept. of Radiotechnology, Wonkwang Health Science University, Iksan (Korea, Republic of); Hong, Dong Hee [Dept. of Radiological Science, Far East University, Eumseong (Korea, Republic of)

2017-06-15

Purpose of this study is present the normal range of cardiac size and cardiothoracic ratio according to patient position(chest PA and AP) and age of Korean adult male on digital chest X - ray, And to propose a mutually compatible conversion rate. 1,024 males were eligible for this study, among 1,300 normal chest patients who underwent chest PA and low-dose CT examinations on the same day at the 'S' Hospital Health Examination Center in Seoul From January to December 2014. CS and CTR were measured by Danzer (1919). The mean difference between CS and CTR was statistically significant (p<0.01) in Chest PA (CS 135.48 mm, CTR 43.99%) and Chest AP image (CS 155.96 mm, CTR 51.75%). There was no statistically significant difference between left and right heart in chest PA and AP images(p>0.05). CS showed statistically significant difference between Chest PA (p>0. 05) and Chest AP (p<0.05). The thorax size and CTR were statistically significant (p<0.01) in both age and chest PA and AP. Result of this study, On Chest AP image CS was magnified 15%, CTR was magnified 17% compare with Chest PA image. CS and CTR were about 10% difference by changing posture at all ages.

12. Likelihood analysis of supersymmetric SU(5) GUTs

Energy Technology Data Exchange (ETDEWEB)

Bagnaschi, E.; Weiglein, G. [DESY, Hamburg (Germany); Costa, J.C.; Buchmueller, O.; Citron, M.; Richards, A.; De Vries, K.J. [Imperial College, High Energy Physics Group, Blackett Laboratory, London (United Kingdom); Sakurai, K. [University of Durham, Science Laboratories, Department of Physics, Institute for Particle Physics Phenomenology, Durham (United Kingdom); University of Warsaw, Faculty of Physics, Institute of Theoretical Physics, Warsaw (Poland); Borsato, M.; Chobanova, V.; Lucio, M.; Martinez Santos, D. [Universidade de Santiago de Compostela, Santiago de Compostela (Spain); Cavanaugh, R. [Fermi National Accelerator Laboratory, Batavia, IL (United States); University of Illinois at Chicago, Physics Department, Chicago, IL (United States); Roeck, A. de [CERN, Experimental Physics Department, Geneva (Switzerland); Antwerp University, Wilrijk (Belgium); Dolan, M.J. [University of Melbourne, ARC Centre of Excellence for Particle Physics at the Terascale, School of Physics, Parkville (Australia); Ellis, J.R. [King' s College London, Theoretical Particle Physics and Cosmology Group, Department of Physics, London (United Kingdom); Theoretical Physics Department, CERN, Geneva 23 (Switzerland); Flaecher, H. [University of Bristol, H.H. Wills Physics Laboratory, Bristol (United Kingdom); Heinemeyer, S. [Campus of International Excellence UAM+CSIC, Cantoblanco, Madrid (Spain); Instituto de Fisica Teorica UAM-CSIC, Madrid (Spain); Instituto de Fisica de Cantabria (CSIC-UC), Santander (Spain); Isidori, G. [Universitaet Zuerich, Physik-Institut, Zurich (Switzerland); Olive, K.A. [University of Minnesota, William I. Fine Theoretical Physics Institute, School of Physics and Astronomy, Minneapolis, MN (United States)

2017-02-15

We perform a likelihood analysis of the constraints from accelerator experiments and astrophysical observations on supersymmetric (SUSY) models with SU(5) boundary conditions on soft SUSY-breaking parameters at the GUT scale. The parameter space of the models studied has seven parameters: a universal gaugino mass m{sub 1/2}, distinct masses for the scalar partners of matter fermions in five- and ten-dimensional representations of SU(5), m{sub 5} and m{sub 10}, and for the 5 and anti 5 Higgs representations m{sub H{sub u}} and m{sub H{sub d}}, a universal trilinear soft SUSY-breaking parameter A{sub 0}, and the ratio of Higgs vevs tan β. In addition to previous constraints from direct sparticle searches, low-energy and flavour observables, we incorporate constraints based on preliminary results from 13 TeV LHC searches for jets + E{sub T} events and long-lived particles, as well as the latest PandaX-II and LUX searches for direct Dark Matter detection. In addition to previously identified mechanisms for bringing the supersymmetric relic density into the range allowed by cosmology, we identify a novel u{sub R}/c{sub R} - χ{sup 0}{sub 1} coannihilation mechanism that appears in the supersymmetric SU(5) GUT model and discuss the role of ν{sub τ} coannihilation. We find complementarity between the prospects for direct Dark Matter detection and SUSY searches at the LHC. (orig.)

13. Likelihood analysis of supersymmetric SU(5) GUTs

Energy Technology Data Exchange (ETDEWEB)

Bagnaschi, E. [DESY, Hamburg (Germany); Costa, J.C. [Imperial College, London (United Kingdom). Blackett Lab.; Sakurai, K. [Durham Univ. (United Kingdom). Inst. for Particle Physics Phenomonology; Warsaw Univ. (Poland). Inst. of Theoretical Physics; Collaboration: MasterCode Collaboration; and others

2016-10-15

We perform a likelihood analysis of the constraints from accelerator experiments and astrophysical observations on supersymmetric (SUSY) models with SU(5) boundary conditions on soft SUSY-breaking parameters at the GUT scale. The parameter space of the models studied has 7 parameters: a universal gaugino mass m{sub 1/2}, distinct masses for the scalar partners of matter fermions in five- and ten-dimensional representations of SU(5), m{sub 5} and m{sub 10}, and for the 5 and anti 5 Higgs representations m{sub H{sub u}} and m{sub H{sub d}}, a universal trilinear soft SUSY-breaking parameter A{sub 0}, and the ratio of Higgs vevs tan β. In addition to previous constraints from direct sparticle searches, low-energy and avour observables, we incorporate constraints based on preliminary results from 13 TeV LHC searches for jets+E{sub T} events and long-lived particles, as well as the latest PandaX-II and LUX searches for direct Dark Matter detection. In addition to previously-identified mechanisms for bringing the supersymmetric relic density into the range allowed by cosmology, we identify a novel u{sub R}/c{sub R}-χ{sup 0}{sub 1} coannihilation mechanism that appears in the supersymmetric SU(5) GUT model and discuss the role of ν{sub T} coannihilation. We find complementarity between the prospects for direct Dark Matter detection and SUSY searches at the LHC.

14. Maximum-Likelihood Detection Of Noncoherent CPM

Science.gov (United States)

Divsalar, Dariush; Simon, Marvin K.

1993-01-01

Simplified detectors proposed for use in maximum-likelihood-sequence detection of symbols in alphabet of size M transmitted by uncoded, full-response continuous phase modulation over radio channel with additive white Gaussian noise. Structures of receivers derived from particular interpretation of maximum-likelihood metrics. Receivers include front ends, structures of which depends only on M, analogous to those in receivers of coherent CPM. Parts of receivers following front ends have structures, complexity of which would depend on N.

15. Total homocysteine is positively correlated with body mass index, waist-to-hip ratio, and fat mass among overweight reproductive women: A cross-sectional study.

Science.gov (United States)

Al-Bayyari, Nahla; Hamadneh, Jehan; Hailat, Rae'd; Hamadneh, Shereen

2017-12-01

Conflicting associations between total homocysteine (tHcy), body mass index (BMI) lean body mass, and fat mass in the general population have been reported. We investigated the hypothesis that elevated tHcy levels are associated with increased BMI, waist-to-hip ratio (WHR), and body fat mass percent. In Jordan, obesity and overweight are prevalent among reproductive women and hyperhomocysteinemia, along with obesity and overweight, are independent risk factors for cardiovascular diseases. The participants used in this cross-sectional study were 325 overweight Jordanian women aged between 18 and 49 years old. The main outcome measures were tHcy, BMI, WHR, fat mass, fat-free mass, and total body water. Serum tHcy was analyzed using a liquid chromatography tandem mass spectrophotometry (LC-MS/MS) complete kit. The body compositions were measured using a bioelectrical impedance analyzer. Study participants were stratified according to their tHcy level into two groups, ≤10 μmol/L and >10 μmol/L, and the difference between mean values of body compositions was evaluated. The tHcy was significantly and negatively correlated with age, fat-free mass, and total body water, and significantly and positively correlated with BMI, hip circumference, WHR, fat mass, and dry lean weight. The chi-square and the independent sample t-tests showed statistically significant (P ≤ .05) differences between tHcy and BMI, WHR, fat and fat-free mass, and total body water percentages. In conclusion, BMI, WHR and body fat mass were found to be associated with elevated tHcy levels among overweight reproductive women, and they might be used as independent predictors of the tHcy level. Copyright © 2017 Elsevier Inc. All rights reserved.

16. Use of (D, MUF) and maximum-likelihood methods for detecting falsification and diversion in data-verification problems

International Nuclear Information System (INIS)

Goldman, A.S.; Beedgen, R.

1982-01-01

The investigation of data falsification and/or diversion is of major concern in nuclear materials accounting procedures used in international safeguards. In this paper, two procedures, denoted by (D,MUF) and LR (Likelihood Ratio), are discussed and compared when testing the hypothesis that neither diversion nor falsification has taken place versus the one-sided alternative that at least one of these parameters is positive. Critical regions and detection probabilities are given for both tests. It is shown that the LR method outperforms (D,MUF) when diversion and falsification take place

17. The fine-tuning cost of the likelihood in SUSY models

CERN Document Server

Ghilencea, D M

2013-01-01

In SUSY models, the fine tuning of the electroweak (EW) scale with respect to their parameters gamma_i={m_0, m_{1/2}, mu_0, A_0, B_0,...} and the maximal likelihood L to fit the experimental data are usually regarded as two different problems. We show that, if one regards the EW minimum conditions as constraints that fix the EW scale, this commonly held view is not correct and that the likelihood contains all the information about fine-tuning. In this case we show that the corrected likelihood is equal to the ratio L/Delta of the usual likelihood L and the traditional fine tuning measure Delta of the EW scale. A similar result is obtained for the integrated likelihood over the set {gamma_i}, that can be written as a surface integral of the ratio L/Delta, with the surface in gamma_i space determined by the EW minimum constraints. As a result, a large likelihood actually demands a large ratio L/Delta or equivalently, a small chi^2_{new}=chi^2_{old}+2*ln(Delta). This shows the fine-tuning cost to the likelihood ...

18. Maximum Likelihood and Restricted Likelihood Solutions in Multiple-Method Studies.

Science.gov (United States)

Rukhin, Andrew L

2011-01-01

A formulation of the problem of combining data from several sources is discussed in terms of random effects models. The unknown measurement precision is assumed not to be the same for all methods. We investigate maximum likelihood solutions in this model. By representing the likelihood equations as simultaneous polynomial equations, the exact form of the Groebner basis for their stationary points is derived when there are two methods. A parametrization of these solutions which allows their comparison is suggested. A numerical method for solving likelihood equations is outlined, and an alternative to the maximum likelihood method, the restricted maximum likelihood, is studied. In the situation when methods variances are considered to be known an upper bound on the between-method variance is obtained. The relationship between likelihood equations and moment-type equations is also discussed.

19. Likelihood inference for unions of interacting discs

DEFF Research Database (Denmark)

Møller, Jesper; Helisová, Katarina

To the best of our knowledge, this is the first paper which discusses likelihood inference or a random set using a germ-grain model, where the individual grains are unobservable edge effects occur, and other complications appear. We consider the case where the grains form a disc process modelled...... is specified with respect to a given marked Poisson model (i.e. a Boolean model). We show how edge effects and other complications can be handled by considering a certain conditional likelihood. Our methodology is illustrated by analyzing Peter Diggle's heather dataset, where we discuss the results...... of simulation-based maximum likelihood inference and the effect of specifying different reference Poisson models....

20. Maximum likelihood estimation for integrated diffusion processes

DEFF Research Database (Denmark)

Baltazar-Larios, Fernando; Sørensen, Michael

We propose a method for obtaining maximum likelihood estimates of parameters in diffusion models when the data is a discrete time sample of the integral of the process, while no direct observations of the process itself are available. The data are, moreover, assumed to be contaminated...... EM-algorithm to obtain maximum likelihood estimates of the parameters in the diffusion model. As part of the algorithm, we use a recent simple method for approximate simulation of diffusion bridges. In simulation studies for the Ornstein-Uhlenbeck process and the CIR process the proposed method works...... by measurement errors. Integrated volatility is an example of this type of observations. Another example is ice-core data on oxygen isotopes used to investigate paleo-temperatures. The data can be viewed as incomplete observations of a model with a tractable likelihood function. Therefore we propose a simulated...

1. Maintaining symmetry of simulated likelihood functions

DEFF Research Database (Denmark)

Andersen, Laura Mørch

This paper suggests solutions to two different types of simulation errors related to Quasi-Monte Carlo integration. Likelihood functions which depend on standard deviations of mixed parameters are symmetric in nature. This paper shows that antithetic draws preserve this symmetry and thereby...... improves precision substantially. Another source of error is that models testing away mixing dimensions must replicate the relevant dimensions of the quasi-random draws in the simulation of the restricted likelihood. These simulation errors are ignored in the standard estimation procedures used today...

2. Likelihood inference for unions of interacting discs

DEFF Research Database (Denmark)

Møller, Jesper; Helisova, K.

2010-01-01

This is probably the first paper which discusses likelihood inference for a random set using a germ-grain model, where the individual grains are unobservable, edge effects occur and other complications appear. We consider the case where the grains form a disc process modelled by a marked point...... process, where the germs are the centres and the marks are the associated radii of the discs. We propose to use a recent parametric class of interacting disc process models, where the minimal sufficient statistic depends on various geometric properties of the random set, and the density is specified......-based maximum likelihood inference and the effect of specifying different reference Poisson models....

3. Maximum Likelihood Compton Polarimetry with the Compton Spectrometer and Imager

Energy Technology Data Exchange (ETDEWEB)

Lowell, A. W.; Boggs, S. E; Chiu, C. L.; Kierans, C. A.; Sleator, C.; Tomsick, J. A.; Zoglauer, A. C. [Space Sciences Laboratory, University of California, Berkeley (United States); Chang, H.-K.; Tseng, C.-H.; Yang, C.-Y. [Institute of Astronomy, National Tsing Hua University, Taiwan (China); Jean, P.; Ballmoos, P. von [IRAP Toulouse (France); Lin, C.-H. [Institute of Physics, Academia Sinica, Taiwan (China); Amman, M. [Lawrence Berkeley National Laboratory (United States)

2017-10-20

Astrophysical polarization measurements in the soft gamma-ray band are becoming more feasible as detectors with high position and energy resolution are deployed. Previous work has shown that the minimum detectable polarization (MDP) of an ideal Compton polarimeter can be improved by ∼21% when an unbinned, maximum likelihood method (MLM) is used instead of the standard approach of fitting a sinusoid to a histogram of azimuthal scattering angles. Here we outline a procedure for implementing this maximum likelihood approach for real, nonideal polarimeters. As an example, we use the recent observation of GRB 160530A with the Compton Spectrometer and Imager. We find that the MDP for this observation is reduced by 20% when the MLM is used instead of the standard method.

4. Role of the Aspartate Transaminase and Platelet Ratio Index in Assessing Hepatic Fibrosis and Liver Inflammation in Adolescent Patients with HBeAg-Positive Chronic Hepatitis B

Directory of Open Access Journals (Sweden)

Yu Zhijian

2015-01-01

Full Text Available This study described an index of aspartate aminotransferase-to-platelet ratio index (APRI to assess hepatic fibrosis with limited expense and widespread availability compared to the liver biopsy in adolescent patients with CHB.

5. Approximate maximum likelihood estimation for population genetic inference.

Science.gov (United States)

Bertl, Johanna; Ewing, Gregory; Kosiol, Carolin; Futschik, Andreas

2017-11-27

In many population genetic problems, parameter estimation is obstructed by an intractable likelihood function. Therefore, approximate estimation methods have been developed, and with growing computational power, sampling-based methods became popular. However, these methods such as Approximate Bayesian Computation (ABC) can be inefficient in high-dimensional problems. This led to the development of more sophisticated iterative estimation methods like particle filters. Here, we propose an alternative approach that is based on stochastic approximation. By moving along a simulated gradient or ascent direction, the algorithm produces a sequence of estimates that eventually converges to the maximum likelihood estimate, given a set of observed summary statistics. This strategy does not sample much from low-likelihood regions of the parameter space, and is fast, even when many summary statistics are involved. We put considerable efforts into providing tuning guidelines that improve the robustness and lead to good performance on problems with high-dimensional summary statistics and a low signal-to-noise ratio. We then investigate the performance of our resulting approach and study its properties in simulations. Finally, we re-estimate parameters describing the demographic history of Bornean and Sumatran orang-utans.

6. Improved Likelihood Function in Particle-based IR Eye Tracking

DEFF Research Database (Denmark)

Satria, R.; Sorensen, J.; Hammoud, R.

2005-01-01

In this paper we propose a log likelihood-ratio function of foreground and background models used in a particle filter to track the eye region in dark-bright pupil image sequences. This model fuses information from both dark and bright pupil images and their difference image into one model. Our...... enhanced tracker overcomes the issues of prior selection of static thresholds during the detection of feature observations in the bright-dark difference images. The auto-initialization process is performed using cascaded classifier trained using adaboost and adapted to IR eye images. Experiments show good...

7. Maximum Likelihood Joint Tracking and Association in Strong Clutter

Directory of Open Access Journals (Sweden)

Leonid I. Perlovsky

2013-01-01

Full Text Available We have developed a maximum likelihood formulation for a joint detection, tracking and association problem. An efficient non-combinatorial algorithm for this problem is developed in case of strong clutter for radar data. By using an iterative procedure of the dynamic logic process “from vague-to-crisp” explained in the paper, the new tracker overcomes the combinatorial complexity of tracking in highly-cluttered scenarios and results in an orders-of-magnitude improvement in signal-to-clutter ratio.

8. Efficient Bit-to-Symbol Likelihood Mappings

Science.gov (United States)

Moision, Bruce E.; Nakashima, Michael A.

2010-01-01

This innovation is an efficient algorithm designed to perform bit-to-symbol and symbol-to-bit likelihood mappings that represent a significant portion of the complexity of an error-correction code decoder for high-order constellations. Recent implementation of the algorithm in hardware has yielded an 8- percent reduction in overall area relative to the prior design.

9. Phylogenetic analysis using parsimony and likelihood methods.

Science.gov (United States)

Yang, Z

1996-02-01

The assumptions underlying the maximum-parsimony (MP) method of phylogenetic tree reconstruction were intuitively examined by studying the way the method works. Computer simulations were performed to corroborate the intuitive examination. Parsimony appears to involve very stringent assumptions concerning the process of sequence evolution, such as constancy of substitution rates between nucleotides, constancy of rates across nucleotide sites, and equal branch lengths in the tree. For practical data analysis, the requirement of equal branch lengths means similar substitution rates among lineages (the existence of an approximate molecular clock), relatively long interior branches, and also few species in the data. However, a small amount of evolution is neither a necessary nor a sufficient requirement of the method. The difficulties involved in the application of current statistical estimation theory to tree reconstruction were discussed, and it was suggested that the approach proposed by Felsenstein (1981, J. Mol. Evol. 17: 368-376) for topology estimation, as well as its many variations and extensions, differs fundamentally from the maximum likelihood estimation of a conventional statistical parameter. Evidence was presented showing that the Felsenstein approach does not share the asymptotic efficiency of the maximum likelihood estimator of a statistical parameter. Computer simulations were performed to study the probability that MP recovers the true tree under a hierarchy of models of nucleotide substitution; its performance relative to the likelihood method was especially noted. The results appeared to support the intuitive examination of the assumptions underlying MP. When a simple model of nucleotide substitution was assumed to generate data, the probability that MP recovers the true topology could be as high as, or even higher than, that for the likelihood method. When the assumed model became more complex and realistic, e.g., when substitution rates were

10. Early infant adipose deposition is positively associated with the n-6 to n-3 fatty acid ratio in human milk independent of maternal BMI.

Science.gov (United States)

Rudolph, M C; Young, B E; Lemas, D J; Palmer, C E; Hernandez, T L; Barbour, L A; Friedman, J E; Krebs, N F; MacLean, P S

2017-04-01

Excessive infant weight gain in the first 6-month of life is a powerful predictor of childhood obesity and related health risks. In mice, omega-6 fatty acids (FAs) serve as potent ligands driving adipogenesis during early development. The ratio of omega-6 relative to omega-3 (n-6/n-3) FA in human milk (HM) has increased threefold over the last 30 years, but the impact of this shift on infant adipose development remains undetermined. This study investigated how maternal obesity and maternal dietary FA (as reflected in maternal red blood cells (RBCs) composition) influenced HM n-6 and n-3 FAs, and whether the HM n-6/n-3 ratio was associated with changes in infant adipose deposition between 2 weeks and 4 months postpartum. Forty-eight infants from normal weight (NW), overweight (OW) and obese (OB) mothers were exclusively or predominantly breastfed over the first 4 months of lactation. Mid-feed HM and maternal RBC were collected at either transitional (2 weeks) or established (4 months) lactation, along with infant body composition assessed using air-displacement plethysmography. The FA composition of HM and maternal RBC was measured quantitatively by lipid mass spectrometry. In transitional and established HM, docosahexaenoic acid (DHA) was lower (P=0.008; 0.005) and the arachidonic acid (AA)/DHA+eicosapentaenoic acid (EPA) ratio was higher (P=0.05; 0.02) in the OB relative to the NW group. Maternal prepregnancy body mass index (BMI) and AA/DHA+EPA ratios in transitional and established HM were moderately correlated (P=0.018; 0.001). Total infant fat mass was increased in the upper AA/DHA+EPA tertile of established HM relative to the lower tertile (P=0.019). The amount of changes in infant fat mass and percentage of body fat were predicted by AA/EPA+DHA ratios in established HM (P=0.038; 0.010). Perinatal infant exposures to a high AA/EPA+DHA ratio during the first 4 months of life, which is primarily reflective of maternal dietary FA, may significantly contribute to

11. [Waist-to-height ratio is an indicator of metabolic risk in children].

Science.gov (United States)

Valle-Leal, Jaime; Abundis-Castro, Leticia; Hernández-Escareño, Juan; Flores-Rubio, Salvador

2016-01-01

Abdominal fat, particularly visceral, is associated with a high risk of metabolic complications. The waist-height ratio (WHtR) is used to assess abdominal fat in individuals of all ages. To determine the ability of the waist-to-height ratio to detect metabolic risk in mexican schoolchildren. A study was conducted on children between 6 and 12 years. Obesity was diagnosed as a body mass index (BMI) ≥ 85th percentile, and an ICE ≥0.5 was considered abdominal obesity. Blood levels of glucose, cholesterol and triglycerides were measured. The sensitivity, specificity, positive predictive and negative value, area under curve, the positive likelihood ratio and negative likelihood ratio of the WHtR and BMI were calculated in order to identify metabolic alterations. WHtR and BMI were compared to determine which had the best diagnostic efficiency. Of the 223 children included in the study, 51 had hypertriglyceridaemia, 27 with hypercholesterolaemia, and 9 with hyperglycaemia. On comparing the diagnostic efficiency of WHtR with that of BMI, there was a sensitivity of 100% vs. 56% for hyperglycaemia, 93 vs. 70% for cholesterol, and 76 vs. 59% for hypertriglyceridaemia. The specificity, negative predictive value, positive predictive value, positive likelihood ratio, negative likelihood ratio, and area under curve were also higher for WHtR. The WHtR is a more efficient indicator than BMI in identifying metabolic risk in mexican school-age. Copyright © 2015 Sociedad Chilena de Pediatría. Publicado por Elsevier España, S.L.U. All rights reserved.

12. Review of Elaboration Likelihood Model of persuasion

OpenAIRE

藤原, 武弘; 神山, 貴弥

1989-01-01

This article mainly introduces Elaboration Likelihood Model (ELM), proposed by Petty & Cacioppo, that is, a general attitude change theory. ELM posturates two routes to persuasion; central and peripheral route. Attitude change by central route is viewed as resulting from a diligent consideration of the issue-relevant informations presented. On the other hand, attitude change by peripheral route is viewed as resulting from peripheral cues in the persuasion context. Secondly we compare these tw...

13. Unbinned likelihood analysis of EGRET observations

International Nuclear Information System (INIS)

Digel, Seth W.

2000-01-01

We present a newly-developed likelihood analysis method for EGRET data that defines the likelihood function without binning the photon data or averaging the instrumental response functions. The standard likelihood analysis applied to EGRET data requires the photons to be binned spatially and in energy, and the point-spread functions to be averaged over energy and inclination angle. The full-width half maximum of the point-spread function increases by about 40% from on-axis to 30 degree sign inclination, and depending on the binning in energy can vary by more than that in a single energy bin. The new unbinned method avoids the loss of information that binning and averaging cause and can properly analyze regions where EGRET viewing periods overlap and photons with different inclination angles would otherwise be combined in the same bin. In the poster, we describe the unbinned analysis method and compare its sensitivity with binned analysis for detecting point sources in EGRET data

14. The association between sequences of sexual initiation and the likelihood of teenage pregnancy.

Science.gov (United States)

Reese, Bianka M; Haydon, Abigail A; Herring, Amy H; Halpern, Carolyn T

2013-02-01

15. Chemical analysis of isolated cell walls of Gram-positive bacteria and the determination of the cell wall to cell mass ratio.

NARCIS (Netherlands)

Wal, van der A.; Norde, W.; Bendinger, B.; Zehnder, A.J.B.; Lyklema, J.

1997-01-01

Cell walls of five Gram-positive bacterial strains, including four coryneforms and a Bacillus brevis strain were isolated and subsequently chemically analysed. The wall contribution to the total cell mass is calculated from a comparison of D-Lactate concentrations in hydrolysates of whole cells and

16. Golden Ratio

Indian Academy of Sciences (India)

Our attraction to another body increases if the body is symmetricaland in proportion. If a face or a structure is in proportion,we are more likely to notice it and find it beautiful.The universal ratio of beauty is the 'Golden Ratio', found inmany structures. This ratio comes from Fibonacci numbers.In this article, we explore this ...

17. Golden Ratio

Indian Academy of Sciences (India)

Keywords. Fibonacci numbers, golden ratio, Sanskrit prosody, solar panel. Abstract. Our attraction to another body increases if the body is symmetricaland in proportion. If a face or a structure is in proportion,we are more likely to notice it and find it beautiful.The universal ratio of beauty is the 'Golden Ratio', found inmany ...

18. Golden Ratio

Indian Academy of Sciences (India)

Our attraction to another body increases if the body is sym- metrical and in proportion. If a face or a structure is in pro- portion, we are more likely to notice it and find it beautiful. The universal ratio of beauty is the 'Golden Ratio', found in many structures. This ratio comes from Fibonacci numbers. In this article, we explore this ...

19. The value of N staging with the positive lymph node ratio, and splenectomy, for remnant gastric cancer: A multicenter retrospective study.

Science.gov (United States)

Son, Sang-Yong; Kong, Seong-Ho; Ahn, Hye Seong; Park, Young Suk; Ahn, Sang-Hoon; Suh, Yun-Suhk; Park, Do Joong; Lee, Hyuk-Joon; Kim, Hyung-Ho; Yang, Han-Kwang

2017-12-01

Surgery for remnant gastric cancer (RGC) frequently fails to obtain the >15 lymph nodes necessary for tumor-node-metastasis (TNM) staging. We aimed to evaluate the utility of the recently developed tumor-ratio-metastasis (TRM) staging system. We also examined the pattern of lymph node metastasis and the role of prophylactic splenectomy in RGC. Between May 2003 and December 2012, data from 170 patients who underwent surgery for RGC were retrospectively analyzed. RGC arising after previous benign disease (n = 46) was associated with retrieval of more lymph nodes (27.3 vs 10.0; P splenectomy and non-splenectomy groups at each stage (P = 0.751, 0.723, 0.151, and 0.706 for stage I, II, III, and IV, respectively). The analyses did not identify a survival benefit from prophylactic splenectomy or show an improvement in staging with the TRM system for RGC. © 2017 Wiley Periodicals, Inc.

20. Dimension-Independent Likelihood-Informed MCMC

KAUST Repository

Cui, Tiangang; Law, Kody; Marzouk, Youssef

2015-01-01

Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters, which in principle can be described as functions. By exploiting low-dimensional structure in the change from prior to posterior [distributions], we introduce a suite of MCMC samplers that can adapt to the complex structure of the posterior distribution, yet are well-defined on function space. Posterior sampling in nonlinear inverse problems arising from various partial di erential equations and also a stochastic differential equation are used to demonstrate the e ciency of these dimension-independent likelihood-informed samplers.

1. Dimension-Independent Likelihood-Informed MCMC

KAUST Repository

Cui, Tiangang

2015-01-07

Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters, which in principle can be described as functions. By exploiting low-dimensional structure in the change from prior to posterior [distributions], we introduce a suite of MCMC samplers that can adapt to the complex structure of the posterior distribution, yet are well-defined on function space. Posterior sampling in nonlinear inverse problems arising from various partial di erential equations and also a stochastic differential equation are used to demonstrate the e ciency of these dimension-independent likelihood-informed samplers.

2. The performance of blood pressure-to-height ratio as a screening measure for identifying children and adolescents with hypertension: a meta-analysis.

Science.gov (United States)

Ma, Chunming; Liu, Yue; Lu, Qiang; Lu, Na; Liu, Xiaoli; Tian, Yiming; Wang, Rui; Yin, Fuzai

2016-02-01

The blood pressure-to-height ratio (BPHR) has been shown to be an accurate index for screening hypertension in children and adolescents. The aim of the present study was to perform a meta-analysis to assess the performance of BPHR for the assessment of hypertension. Electronic and manual searches were performed to identify studies of the BPHR. After methodological quality assessment and data extraction, pooled estimates of the sensitivity, specificity, positive likelihood ratio, negative likelihood ratio, diagnostic odds ratio, area under the receiver operating characteristic curve and summary receiver operating characteristics were assessed systematically. The extent of heterogeneity for it was assessed. Six studies were identified for analysis. The pooled sensitivity, specificity, positive likelihood ratio, negative likelihood ratio and diagnostic odds ratio values of BPHR, for assessment of hypertension, were 96% [95% confidence interval (CI)=0.95-0.97], 90% (95% CI=0.90-0.91), 10.68 (95% CI=8.03-14.21), 0.04 (95% CI=0.03-0.07) and 247.82 (95% CI=114.50-536.34), respectively. The area under the receiver operating characteristic curve was 0.9472. The BPHR had higher diagnostic accuracies for identifying hypertension in children and adolescents.

3. The fine-tuning cost of the likelihood in SUSY models

International Nuclear Information System (INIS)

Ghilencea, D.M.; Ross, G.G.

2013-01-01

In SUSY models, the fine-tuning of the electroweak (EW) scale with respect to their parameters γ i ={m 0 ,m 1/2 ,μ 0 ,A 0 ,B 0 ,…} and the maximal likelihood L to fit the experimental data are usually regarded as two different problems. We show that, if one regards the EW minimum conditions as constraints that fix the EW scale, this commonly held view is not correct and that the likelihood contains all the information about fine-tuning. In this case we show that the corrected likelihood is equal to the ratio L/Δ of the usual likelihood L and the traditional fine-tuning measure Δ of the EW scale. A similar result is obtained for the integrated likelihood over the set {γ i }, that can be written as a surface integral of the ratio L/Δ, with the surface in γ i space determined by the EW minimum constraints. As a result, a large likelihood actually demands a large ratio L/Δ or equivalently, a small χ new 2 =χ old 2 +2lnΔ. This shows the fine-tuning cost to the likelihood (χ new 2 ) of the EW scale stability enforced by SUSY, that is ignored in data fits. A good χ new 2 /d.o.f.≈1 thus demands SUSY models have a fine-tuning amount Δ≪exp(d.o.f./2), which provides a model-independent criterion for acceptable fine-tuning. If this criterion is not met, one can thus rule out SUSY models without a further χ 2 /d.o.f. analysis. Numerical methods to fit the data can easily be adapted to account for this effect.

4. Sex ratios

OpenAIRE

West, Stuart A; Reece, S E; Sheldon, Ben C

2002-01-01

Sex ratio theory attempts to explain variation at all levels (species, population, individual, brood) in the proportion of offspring that are male (the sex ratio). In many cases this work has been extremely successful, providing qualitative and even quantitative explanations of sex ratio variation. However, this is not always the situation, and one of the greatest remaining problems is explaining broad taxonomic patterns. Specifically, why do different organisms show so ...

5. A Non-standard Empirical Likelihood for Time Series

DEFF Research Database (Denmark)

Nordman, Daniel J.; Bunzel, Helle; Lahiri, Soumendra N.

Standard blockwise empirical likelihood (BEL) for stationary, weakly dependent time series requires specifying a fixed block length as a tuning parameter for setting confidence regions. This aspect can be difficult and impacts coverage accuracy. As an alternative, this paper proposes a new version...... of BEL based on a simple, though non-standard, data-blocking rule which uses a data block of every possible length. Consequently, the method involves no block selection and is also anticipated to exhibit better coverage performance. Its non-standard blocking scheme, however, induces non......-standard asymptotics and requires a significantly different development compared to standard BEL. We establish the large-sample distribution of log-ratio statistics from the new BEL method for calibrating confidence regions for mean or smooth function parameters of time series. This limit law is not the usual chi...

6. A Predictive Likelihood Approach to Bayesian Averaging

Directory of Open Access Journals (Sweden)

Tomáš Jeřábek

2015-01-01

Full Text Available Multivariate time series forecasting is applied in a wide range of economic activities related to regional competitiveness and is the basis of almost all macroeconomic analysis. In this paper we combine multivariate density forecasts of GDP growth, inflation and real interest rates from four various models, two type of Bayesian vector autoregression (BVAR models, a New Keynesian dynamic stochastic general equilibrium (DSGE model of small open economy and DSGE-VAR model. The performance of models is identified using historical dates including domestic economy and foreign economy, which is represented by countries of the Eurozone. Because forecast accuracy of observed models are different, the weighting scheme based on the predictive likelihood, the trace of past MSE matrix, model ranks are used to combine the models. The equal-weight scheme is used as a simple combination scheme. The results show that optimally combined densities are comparable to the best individual models.

7. Maximum Likelihood Reconstruction for Magnetic Resonance Fingerprinting.

Science.gov (United States)

Zhao, Bo; Setsompop, Kawin; Ye, Huihui; Cauley, Stephen F; Wald, Lawrence L

2016-08-01

This paper introduces a statistical estimation framework for magnetic resonance (MR) fingerprinting, a recently proposed quantitative imaging paradigm. Within this framework, we present a maximum likelihood (ML) formalism to estimate multiple MR tissue parameter maps directly from highly undersampled, noisy k-space data. A novel algorithm, based on variable splitting, the alternating direction method of multipliers, and the variable projection method, is developed to solve the resulting optimization problem. Representative results from both simulations and in vivo experiments demonstrate that the proposed approach yields significantly improved accuracy in parameter estimation, compared to the conventional MR fingerprinting reconstruction. Moreover, the proposed framework provides new theoretical insights into the conventional approach. We show analytically that the conventional approach is an approximation to the ML reconstruction; more precisely, it is exactly equivalent to the first iteration of the proposed algorithm for the ML reconstruction, provided that a gridding reconstruction is used as an initialization.

8. Subtracting and Fitting Histograms using Profile Likelihood

CERN Document Server

D'Almeida, F M L

2008-01-01

It is known that many interesting signals expected at LHC are of unknown shape and strongly contaminated by background events. These signals will be dif cult to detect during the rst years of LHC operation due to the initial low luminosity. In this work, one presents a method of subtracting histograms based on the pro le likelihood function when the background is previously estimated by Monte Carlo events and one has low statistics. Estimators for the signal in each bin of the histogram difference are calculated so as limits for the signals with 68.3% of Con dence Level in a low statistics case when one has a exponential background and a Gaussian signal. The method can also be used to t histograms when the signal shape is known. Our results show a good performance and avoid the problem of negative values when subtracting histograms.

9. Maximum likelihood estimation of ancestral codon usage bias parameters in Drosophila

DEFF Research Database (Denmark)

Nielsen, Rasmus; Bauer DuMont, Vanessa L; Hubisz, Melissa J

2007-01-01

: the selection coefficient for optimal codon usage (S), allowing joint maximum likelihood estimation of S and the dN/dS ratio. We apply the method to previously published data from Drosophila melanogaster, Drosophila simulans, and Drosophila yakuba and show, in accordance with previous results, that the D...

10. Experimental demonstration of the maximum likelihood-based chromatic dispersion estimator for coherent receivers

DEFF Research Database (Denmark)

Borkowski, Robert; Johannisson, Pontus; Wymeersch, Henk

2014-01-01

We perform an experimental investigation of a maximum likelihood-based (ML-based) algorithm for bulk chromatic dispersion estimation for digital coherent receivers operating in uncompensated optical networks. We demonstrate the robustness of the method at low optical signal-to-noise ratio (OSNR...

11. A maximum likelihood framework for protein design

Directory of Open Access Journals (Sweden)

Philippe Hervé

2006-06-01

Full Text Available Abstract Background The aim of protein design is to predict amino-acid sequences compatible with a given target structure. Traditionally envisioned as a purely thermodynamic question, this problem can also be understood in a wider context, where additional constraints are captured by learning the sequence patterns displayed by natural proteins of known conformation. In this latter perspective, however, we still need a theoretical formalization of the question, leading to general and efficient learning methods, and allowing for the selection of fast and accurate objective functions quantifying sequence/structure compatibility. Results We propose a formulation of the protein design problem in terms of model-based statistical inference. Our framework uses the maximum likelihood principle to optimize the unknown parameters of a statistical potential, which we call an inverse potential to contrast with classical potentials used for structure prediction. We propose an implementation based on Markov chain Monte Carlo, in which the likelihood is maximized by gradient descent and is numerically estimated by thermodynamic integration. The fit of the models is evaluated by cross-validation. We apply this to a simple pairwise contact potential, supplemented with a solvent-accessibility term, and show that the resulting models have a better predictive power than currently available pairwise potentials. Furthermore, the model comparison method presented here allows one to measure the relative contribution of each component of the potential, and to choose the optimal number of accessibility classes, which turns out to be much higher than classically considered. Conclusion Altogether, this reformulation makes it possible to test a wide diversity of models, using different forms of potentials, or accounting for other factors than just the constraint of thermodynamic stability. Ultimately, such model-based statistical analyses may help to understand the forces

12. Modelling maximum likelihood estimation of availability

International Nuclear Information System (INIS)

Waller, R.A.; Tietjen, G.L.; Rock, G.W.

1975-01-01

Suppose the performance of a nuclear powered electrical generating power plant is continuously monitored to record the sequence of failure and repairs during sustained operation. The purpose of this study is to assess one method of estimating the performance of the power plant when the measure of performance is availability. That is, we determine the probability that the plant is operational at time t. To study the availability of a power plant, we first assume statistical models for the variables, X and Y, which denote the time-to-failure and the time-to-repair variables, respectively. Once those statistical models are specified, the availability, A(t), can be expressed as a function of some or all of their parameters. Usually those parameters are unknown in practice and so A(t) is unknown. This paper discusses the maximum likelihood estimator of A(t) when the time-to-failure model for X is an exponential density with parameter, lambda, and the time-to-repair model for Y is an exponential density with parameter, theta. Under the assumption of exponential models for X and Y, it follows that the instantaneous availability at time t is A(t)=lambda/(lambda+theta)+theta/(lambda+theta)exp[-[(1/lambda)+(1/theta)]t] with t>0. Also, the steady-state availability is A(infinity)=lambda/(lambda+theta). We use the observations from n failure-repair cycles of the power plant, say X 1 , X 2 , ..., Xsub(n), Y 1 , Y 2 , ..., Ysub(n) to present the maximum likelihood estimators of A(t) and A(infinity). The exact sampling distributions for those estimators and some statistical properties are discussed before a simulation model is used to determine 95% simulation intervals for A(t). The methodology is applied to two examples which approximate the operating history of two nuclear power plants. (author)

13. Corporate brand extensions based on the purchase likelihood: governance implications

Directory of Open Access Journals (Sweden)

Spyridon Goumas

2018-03-01

Full Text Available This paper is examining the purchase likelihood of hypothetical service brand extensions from product companies focusing on consumer electronics based on sector categorization and perceptions of fit between the existing product category and image of the company. Prior research has recognized that levels of brand knowledge eases the transference of associations and affect to the new products. Similarity to the existing products of the parent company and perceived image also influence the success of brand extensions. However, sector categorization may interfere with this relationship. The purpose of this study is to examine Greek consumers’ attitudes towards hypothetical brand extensions, and how these are affected by consumers’ existing knowledge about the brand, sector categorization and perceptions of image and category fit of cross-sector extensions. This aim is examined in the context of technological categories, where less-known companies exhibited significance in purchase likelihood, and contradictory with the existing literature, service companies did not perform as positively as expected. Additional insights to the existing literature about sector categorization are provided. The effect of both image and category fit is also examined and predictions regarding the effect of each are made.

14. Approximate likelihood approaches for detecting the influence of primordial gravitational waves in cosmic microwave background polarization

Science.gov (United States)

Pan, Zhen; Anderes, Ethan; Knox, Lloyd

2018-05-01

One of the major targets for next-generation cosmic microwave background (CMB) experiments is the detection of the primordial B-mode signal. Planning is under way for Stage-IV experiments that are projected to have instrumental noise small enough to make lensing and foregrounds the dominant source of uncertainty for estimating the tensor-to-scalar ratio r from polarization maps. This makes delensing a crucial part of future CMB polarization science. In this paper we present a likelihood method for estimating the tensor-to-scalar ratio r from CMB polarization observations, which combines the benefits of a full-scale likelihood approach with the tractability of the quadratic delensing technique. This method is a pixel space, all order likelihood analysis of the quadratic delensed B modes, and it essentially builds upon the quadratic delenser by taking into account all order lensing and pixel space anomalies. Its tractability relies on a crucial factorization of the pixel space covariance matrix of the polarization observations which allows one to compute the full Gaussian approximate likelihood profile, as a function of r , at the same computational cost of a single likelihood evaluation.

15. Preliminary application of maximum likelihood method in HL-2A Thomson scattering system

International Nuclear Information System (INIS)

Yao Ke; Huang Yuan; Feng Zhen; Liu Chunhua; Li Enping; Nie Lin

2010-01-01

Maximum likelihood method to process the data of HL-2A Thomson scattering system is presented. Using mathematical statistics, this method maximizes the possibility of the likeness between the theoretical data and the observed data, so that we could get more accurate result. It has been proved to be applicable in comparison with that of the ratios method, and some of the drawbacks in ratios method do not exist in this new one. (authors)

16. Neutron spectra unfolding with maximum entropy and maximum likelihood

International Nuclear Information System (INIS)

Itoh, Shikoh; Tsunoda, Toshiharu

1989-01-01

A new unfolding theory has been established on the basis of the maximum entropy principle and the maximum likelihood method. This theory correctly embodies the Poisson statistics of neutron detection, and always brings a positive solution over the whole energy range. Moreover, the theory unifies both problems of overdetermined and of underdetermined. For the latter, the ambiguity in assigning a prior probability, i.e. the initial guess in the Bayesian sense, has become extinct by virtue of the principle. An approximate expression of the covariance matrix for the resultant spectra is also presented. An efficient algorithm to solve the nonlinear system, which appears in the present study, has been established. Results of computer simulation showed the effectiveness of the present theory. (author)

17. Likelihood analysis of the minimal AMSB model

Energy Technology Data Exchange (ETDEWEB)

Bagnaschi, E.; Weiglein, G. [DESY, Hamburg (Germany); Borsato, M.; Chobanova, V.; Lucio, M.; Santos, D.M. [Universidade de Santiago de Compostela, Santiago de Compostela (Spain); Sakurai, K. [Institute for Particle Physics Phenomenology, University of Durham, Science Laboratories, Department of Physics, Durham (United Kingdom); University of Warsaw, Faculty of Physics, Institute of Theoretical Physics, Warsaw (Poland); Buchmueller, O.; Citron, M.; Costa, J.C.; Richards, A. [Imperial College, High Energy Physics Group, Blackett Laboratory, London (United Kingdom); Cavanaugh, R. [Fermi National Accelerator Laboratory, Batavia, IL (United States); University of Illinois at Chicago, Physics Department, Chicago, IL (United States); De Roeck, A. [Experimental Physics Department, CERN, Geneva (Switzerland); Antwerp University, Wilrijk (Belgium); Dolan, M.J. [School of Physics, University of Melbourne, ARC Centre of Excellence for Particle Physics at the Terascale, Melbourne (Australia); Ellis, J.R. [King' s College London, Theoretical Particle Physics and Cosmology Group, Department of Physics, London (United Kingdom); CERN, Theoretical Physics Department, Geneva (Switzerland); Flaecher, H. [University of Bristol, H.H. Wills Physics Laboratory, Bristol (United Kingdom); Heinemeyer, S. [Campus of International Excellence UAM+CSIC, Madrid (Spain); Instituto de Fisica Teorica UAM-CSIC, Madrid (Spain); Instituto de Fisica de Cantabria (CSIC-UC), Cantabria (Spain); Isidori, G. [Physik-Institut, Universitaet Zuerich, Zurich (Switzerland); Luo, F. [Kavli IPMU (WPI), UTIAS, The University of Tokyo, Kashiwa, Chiba (Japan); Olive, K.A. [School of Physics and Astronomy, University of Minnesota, William I. Fine Theoretical Physics Institute, Minneapolis, MN (United States)

2017-04-15

We perform a likelihood analysis of the minimal anomaly-mediated supersymmetry-breaking (mAMSB) model using constraints from cosmology and accelerator experiments. We find that either a wino-like or a Higgsino-like neutralino LSP, χ{sup 0}{sub 1}, may provide the cold dark matter (DM), both with similar likelihoods. The upper limit on the DM density from Planck and other experiments enforces m{sub χ{sup 0}{sub 1}} 0) but the scalar mass m{sub 0} is poorly constrained. In the wino-LSP case, m{sub 3/2} is constrained to about 900 TeV and m{sub χ{sup 0}{sub 1}} to 2.9 ± 0.1 TeV, whereas in the Higgsino-LSP case m{sub 3/2} has just a lower limit >or similar 650 TeV (>or similar 480 TeV) and m{sub χ{sup 0}{sub 1}} is constrained to 1.12 (1.13) ± 0.02 TeV in the μ > 0 (μ < 0) scenario. In neither case can the anomalous magnetic moment of the muon, (g-2){sub μ}, be improved significantly relative to its Standard Model (SM) value, nor do flavour measurements constrain the model significantly, and there are poor prospects for discovering supersymmetric particles at the LHC, though there are some prospects for direct DM detection. On the other hand, if the χ{sup 0}{sub 1} contributes only a fraction of the cold DM density, future LHC E{sub T}-based searches for gluinos, squarks and heavier chargino and neutralino states as well as disappearing track searches in the wino-like LSP region will be relevant, and interference effects enable BR(B{sub s,d} → μ{sup +}μ{sup -}) to agree with the data better than in the SM in the case of wino-like DM with μ > 0. (orig.)

18. Dimension-independent likelihood-informed MCMC

KAUST Repository

Cui, Tiangang

2015-10-08

Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters that represent the discretization of an underlying function. This work introduces a family of Markov chain Monte Carlo (MCMC) samplers that can adapt to the particular structure of a posterior distribution over functions. Two distinct lines of research intersect in the methods developed here. First, we introduce a general class of operator-weighted proposal distributions that are well defined on function space, such that the performance of the resulting MCMC samplers is independent of the discretization of the function. Second, by exploiting local Hessian information and any associated low-dimensional structure in the change from prior to posterior distributions, we develop an inhomogeneous discretization scheme for the Langevin stochastic differential equation that yields operator-weighted proposals adapted to the non-Gaussian structure of the posterior. The resulting dimension-independent and likelihood-informed (DILI) MCMC samplers may be useful for a large class of high-dimensional problems where the target probability measure has a density with respect to a Gaussian reference measure. Two nonlinear inverse problems are used to demonstrate the efficiency of these DILI samplers: an elliptic PDE coefficient inverse problem and path reconstruction in a conditioned diffusion.

19. Reducing the likelihood of long tennis matches.

Science.gov (United States)

Barnett, Tristan; Alan, Brown; Pollard, Graham

2006-01-01

Long matches can cause problems for tournaments. For example, the starting times of subsequent matches can be substantially delayed causing inconvenience to players, spectators, officials and television scheduling. They can even be seen as unfair in the tournament setting when the winner of a very long match, who may have negative aftereffects from such a match, plays the winner of an average or shorter length match in the next round. Long matches can also lead to injuries to the participating players. One factor that can lead to long matches is the use of the advantage set as the fifth set, as in the Australian Open, the French Open and Wimbledon. Another factor is long rallies and a greater than average number of points per game. This tends to occur more frequently on the slower surfaces such as at the French Open. The mathematical method of generating functions is used to show that the likelihood of long matches can be substantially reduced by using the tiebreak game in the fifth set, or more effectively by using a new type of game, the 50-40 game, throughout the match. Key PointsThe cumulant generating function has nice properties for calculating the parameters of distributions in a tennis matchA final tiebreaker set reduces the length of matches as currently being used in the US OpenA new 50-40 game reduces the length of matches whilst maintaining comparable probabilities for the better player to win the match.

20. Dimension-independent likelihood-informed MCMC

KAUST Repository

Cui, Tiangang; Law, Kody; Marzouk, Youssef M.

2015-01-01

Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters that represent the discretization of an underlying function. This work introduces a family of Markov chain Monte Carlo (MCMC) samplers that can adapt to the particular structure of a posterior distribution over functions. Two distinct lines of research intersect in the methods developed here. First, we introduce a general class of operator-weighted proposal distributions that are well defined on function space, such that the performance of the resulting MCMC samplers is independent of the discretization of the function. Second, by exploiting local Hessian information and any associated low-dimensional structure in the change from prior to posterior distributions, we develop an inhomogeneous discretization scheme for the Langevin stochastic differential equation that yields operator-weighted proposals adapted to the non-Gaussian structure of the posterior. The resulting dimension-independent and likelihood-informed (DILI) MCMC samplers may be useful for a large class of high-dimensional problems where the target probability measure has a density with respect to a Gaussian reference measure. Two nonlinear inverse problems are used to demonstrate the efficiency of these DILI samplers: an elliptic PDE coefficient inverse problem and path reconstruction in a conditioned diffusion.

1. Maximum likelihood window for time delay estimation

International Nuclear Information System (INIS)

Lee, Young Sup; Yoon, Dong Jin; Kim, Chi Yup

2004-01-01

Time delay estimation for the detection of leak location in underground pipelines is critically important. Because the exact leak location depends upon the precision of the time delay between sensor signals due to leak noise and the speed of elastic waves, the research on the estimation of time delay has been one of the key issues in leak lovating with the time arrival difference method. In this study, an optimal Maximum Likelihood window is considered to obtain a better estimation of the time delay. This method has been proved in experiments, which can provide much clearer and more precise peaks in cross-correlation functions of leak signals. The leak location error has been less than 1 % of the distance between sensors, for example the error was not greater than 3 m for 300 m long underground pipelines. Apart from the experiment, an intensive theoretical analysis in terms of signal processing has been described. The improved leak locating with the suggested method is due to the windowing effect in frequency domain, which offers a weighting in significant frequencies.

2. Optimized Large-scale CMB Likelihood and Quadratic Maximum Likelihood Power Spectrum Estimation

Science.gov (United States)

Gjerløw, E.; Colombo, L. P. L.; Eriksen, H. K.; Górski, K. M.; Gruppuso, A.; Jewell, J. B.; Plaszczynski, S.; Wehus, I. K.

2015-11-01

We revisit the problem of exact cosmic microwave background (CMB) likelihood and power spectrum estimation with the goal of minimizing computational costs through linear compression. This idea was originally proposed for CMB purposes by Tegmark et al., and here we develop it into a fully functioning computational framework for large-scale polarization analysis, adopting WMAP as a working example. We compare five different linear bases (pixel space, harmonic space, noise covariance eigenvectors, signal-to-noise covariance eigenvectors, and signal-plus-noise covariance eigenvectors) in terms of compression efficiency, and find that the computationally most efficient basis is the signal-to-noise eigenvector basis, which is closely related to the Karhunen-Loeve and Principal Component transforms, in agreement with previous suggestions. For this basis, the information in 6836 unmasked WMAP sky map pixels can be compressed into a smaller set of 3102 modes, with a maximum error increase of any single multipole of 3.8% at ℓ ≤ 32 and a maximum shift in the mean values of a joint distribution of an amplitude-tilt model of 0.006σ. This compression reduces the computational cost of a single likelihood evaluation by a factor of 5, from 38 to 7.5 CPU seconds, and it also results in a more robust likelihood by implicitly regularizing nearly degenerate modes. Finally, we use the same compression framework to formulate a numerically stable and computationally efficient variation of the Quadratic Maximum Likelihood implementation, which requires less than 3 GB of memory and 2 CPU minutes per iteration for ℓ ≤ 32, rendering low-ℓ QML CMB power spectrum analysis fully tractable on a standard laptop.

3. Maximum likelihood versus likelihood-free quantum system identification in the atom maser

International Nuclear Information System (INIS)

Catana, Catalin; Kypraios, Theodore; Guţă, Mădălin

2014-01-01

We consider the problem of estimating a dynamical parameter of a Markovian quantum open system (the atom maser), by performing continuous time measurements in the system's output (outgoing atoms). Two estimation methods are investigated and compared. Firstly, the maximum likelihood estimator (MLE) takes into account the full measurement data and is asymptotically optimal in terms of its mean square error. Secondly, the ‘likelihood-free’ method of approximate Bayesian computation (ABC) produces an approximation of the posterior distribution for a given set of summary statistics, by sampling trajectories at different parameter values and comparing them with the measurement data via chosen statistics. Building on previous results which showed that atom counts are poor statistics for certain values of the Rabi angle, we apply MLE to the full measurement data and estimate its Fisher information. We then select several correlation statistics such as waiting times, distribution of successive identical detections, and use them as input of the ABC algorithm. The resulting posterior distribution follows closely the data likelihood, showing that the selected statistics capture ‘most’ statistical information about the Rabi angle. (paper)

4. Person fit for test speededness: normal curvatures, likelihood ratio tests and empirical Bayes estimates

NARCIS (Netherlands)

Goegebeur, Y.; de Boeck, P.; Molenberghs, G.

2010-01-01

The local influence diagnostics, proposed by Cook (1986), provide a flexible way to assess the impact of minor model perturbations on key model parameters’ estimates. In this paper, we apply the local influence idea to the detection of test speededness in a model describing nonresponse in test data,

5. Table 1 Log likelihood ratios for the respective traits under different ...

African Journals Online (AJOL)

Schalk Cloete

content, digestion, gut fill and passage rate in ruminants (West, 1998). ... Before the start of each sampling cycle, the plots were machine-cut at the early ..... Hay quality and marketing in the Rocky Mountain Front Range and High Plains.

6. Diffuse Prior Monotonic Likelihood Ratio Test for Evaluation of Fused Image Quality Measures

Science.gov (United States)

2011-02-01

pathological case 3As discussed in [33], any pmf of over can be generated by choosing a specific pdf to generate . WEI et al.: DIFFUSE PRIOR...fused via six different algorithms: 1) contrast pyramid A (CONA), 2) contrast pyramid B (CONB) [34], 3) dis- crete wavelet transform (DWTT) [1], [35...36] 4) color discrete wavelet transform (CDWT), 5) color averaging (CLAV), and 6) color multiscale transform (CLMT) [37]. The first three al

7. Lay Understanding of Forensic Statistics: Evaluation of Random Match Probabilities, Likelihood Ratios, and Verbal Equivalents

OpenAIRE

Thompson, WC; Newman, EJ

2015-01-01

Forensic scientists have come under increasing pressure to quantify the strength of their evidence, but it is not clear which of several possible formats for presenting quantitative conclusions will be easiest for lay people, such as jurors, to understand. This experiment examined the way that people recruited from Amazon’s Mechanical Turk (n = 541) responded to 2 types of forensic evidence—a DNA comparison and a shoeprint comparison—when an expert explained the strength of this evidence 3 di...

8. Table 1 Log likelihood ratios for the respective traits under different ...

African Journals Online (AJOL)

Schalk Cloete

2003). The live weight of lactating cows is an important measure as it reflects feeding costs related to maintenance. Although genetic parameters for yield traits have been estimated in the major South African ... km east of Cape Town in the winter rainfall region of South Africa at an altitude of 177 m, longitude 18° 51'.

9. Testing Measurement Invariance Using MIMIC: Likelihood Ratio Test with a Critical Value Adjustment

Science.gov (United States)

Kim, Eun Sook; Yoon, Myeongsun; Lee, Taehun

2012-01-01

Multiple-indicators multiple-causes (MIMIC) modeling is often used to test a latent group mean difference while assuming the equivalence of factor loadings and intercepts over groups. However, this study demonstrated that MIMIC was insensitive to the presence of factor loading noninvariance, which implies that factor loading invariance should be…

10. Maximum mutual information vector quantization of log-likelihood ratios for memory efficient HARQ implementations

DEFF Research Database (Denmark)

Danieli, Matteo; Forchhammer, Søren; Andersen, Jakob Dahl

2010-01-01

analysis leads to using maximum mutual information (MMI) as optimality criterion and in turn Kullback-Leibler (KL) divergence as distortion measure. Simulations run based on an LTE-like system have proven that VQ can be implemented in a computationally simple way at low rates of 2-3 bits per LLR value......Modern mobile telecommunication systems, such as 3GPP LTE, make use of Hybrid Automatic Repeat reQuest (HARQ) for efficient and reliable communication between base stations and mobile terminals. To this purpose, marginal posterior probabilities of the received bits are stored in the form of log...

11. Predictors for benign paroxysmal positional vertigo with positive Dix–Hallpike test

Directory of Open Access Journals (Sweden)

Noda K

2011-12-01

Full Text Available Kazutaka Noda, Masatomi Ikusaka, Yoshiyuki Ohira, Toshihiko Takada, Tomoko TsukamotoDepartment of General Medicine, Chiba University Hospital, Chiba, JapanObjective: Patient medical history is important for making a diagnosis of causes of dizziness, but there have been no studies on the diagnostic value of individual items in the history. This study was performed to identify and validate useful questions for suspecting a diagnosis of benign paroxysmal positional vertigo (BPPV.Methods: Construction and validation of a disease prediction model was performed at the outpatient clinic in the Department of General Medicine of Chiba University Hospital. Patients with dizziness were enrolled (145 patients for construction of the disease prediction model and 61 patients for its validation. This study targeted BPPV of the posterior semicircular canals only with a positive Dix–Hallpike test (DHT + BPPV to avoid diagnostic ambiguity. Binomial logistic regression analysis was performed to identify the items that were useful for diagnosis or exclusion of DHT + BPPV.Results: Twelve patients from the derivation set and six patients from the validation set had DHT + BPPV. Binomial logistic regression analysis selected a "duration of dizziness ≤15 seconds" and "onset when turning over in bed" as independent predictors of DHT + BPPV with an odds ratio (95% confidence interval of 4.36 (1.18–16.19 and 10.17 (2.49–41.63, respectively. Affirmative answers to both questions yielded a likelihood ratio of 6.81 (5.11–9.10 for diagnosis of DHT + BPPV, while negative answers to both had a likelihood ratio of 0.19 (0.08–0.47.Conclusion: A "duration of dizziness ≤15 seconds" and "onset when turning over in bed" were the two most important questions among various historical features of BPPV.Keywords: benign paroxysmal positional vertigo, likelihood ratio, diagnosis, screening, prediction rules

12. How groups contest depends on group power and the likelihood that power determines victory and defeat

NARCIS (Netherlands)

Kamans, Elanor; Otten, Sabine; Gordijn, Ernestine H.; Spears, Russell

2010-01-01

The aim of the current study was to show that the type of conflict behavior (constructive vs. unconstructive) groups use in conflicts depends on their power position as well as the likelihood that power determines victory and defeat. In an alleged online debate, we created a conflict between two

13. Targeted maximum likelihood estimation for a binary treatment: A tutorial.

Science.gov (United States)

Luque-Fernandez, Miguel Angel; Schomaker, Michael; Rachet, Bernard; Schnitzer, Mireille E

2018-04-23

When estimating the average effect of a binary treatment (or exposure) on an outcome, methods that incorporate propensity scores, the G-formula, or targeted maximum likelihood estimation (TMLE) are preferred over naïve regression approaches, which are biased under misspecification of a parametric outcome model. In contrast propensity score methods require the correct specification of an exposure model. Double-robust methods only require correct specification of either the outcome or the exposure model. Targeted maximum likelihood estimation is a semiparametric double-robust method that improves the chances of correct model specification by allowing for flexible estimation using (nonparametric) machine-learning methods. It therefore requires weaker assumptions than its competitors. We provide a step-by-step guided implementation of TMLE and illustrate it in a realistic scenario based on cancer epidemiology where assumptions about correct model specification and positivity (ie, when a study participant had 0 probability of receiving the treatment) are nearly violated. This article provides a concise and reproducible educational introduction to TMLE for a binary outcome and exposure. The reader should gain sufficient understanding of TMLE from this introductory tutorial to be able to apply the method in practice. Extensive R-code is provided in easy-to-read boxes throughout the article for replicability. Stata users will find a testing implementation of TMLE and additional material in the Appendix S1 and at the following GitHub repository: https://github.com/migariane/SIM-TMLE-tutorial. © 2018 The Authors. Statistics in Medicine published by John Wiley & Sons Ltd.

14. A multifactorial likelihood model for MMR gene variant classification incorporating probabilities based on sequence bioinformatics and tumor characteristics: a report from the Colon Cancer Family Registry.

Science.gov (United States)

Thompson, Bryony A; Goldgar, David E; Paterson, Carol; Clendenning, Mark; Walters, Rhiannon; Arnold, Sven; Parsons, Michael T; Michael D, Walsh; Gallinger, Steven; Haile, Robert W; Hopper, John L; Jenkins, Mark A; Lemarchand, Loic; Lindor, Noralane M; Newcomb, Polly A; Thibodeau, Stephen N; Young, Joanne P; Buchanan, Daniel D; Tavtigian, Sean V; Spurdle, Amanda B

2013-01-01

Mismatch repair (MMR) gene sequence variants of uncertain clinical significance are often identified in suspected Lynch syndrome families, and this constitutes a challenge for both researchers and clinicians. Multifactorial likelihood model approaches provide a quantitative measure of MMR variant pathogenicity, but first require input of likelihood ratios (LRs) for different MMR variation-associated characteristics from appropriate, well-characterized reference datasets. Microsatellite instability (MSI) and somatic BRAF tumor data for unselected colorectal cancer probands of known pathogenic variant status were used to derive LRs for tumor characteristics using the Colon Cancer Family Registry (CFR) resource. These tumor LRs were combined with variant segregation within families, and estimates of prior probability of pathogenicity based on sequence conservation and position, to analyze 44 unclassified variants identified initially in Australasian Colon CFR families. In addition, in vitro splicing analyses were conducted on the subset of variants based on bioinformatic splicing predictions. The LR in favor of pathogenicity was estimated to be ~12-fold for a colorectal tumor with a BRAF mutation-negative MSI-H phenotype. For 31 of the 44 variants, the posterior probabilities of pathogenicity were such that altered clinical management would be indicated. Our findings provide a working multifactorial likelihood model for classification that carefully considers mode of ascertainment for gene testing. © 2012 Wiley Periodicals, Inc.

15. Penalized Maximum Likelihood Estimation for univariate normal mixture distributions

International Nuclear Information System (INIS)

Ridolfi, A.; Idier, J.

2001-01-01

Due to singularities of the likelihood function, the maximum likelihood approach for the estimation of the parameters of normal mixture models is an acknowledged ill posed optimization problem. Ill posedness is solved by penalizing the likelihood function. In the Bayesian framework, it amounts to incorporating an inverted gamma prior in the likelihood function. A penalized version of the EM algorithm is derived, which is still explicit and which intrinsically assures that the estimates are not singular. Numerical evidence of the latter property is put forward with a test

16. Number of evaluated lymph nodes and positive lymph nodes, lymph node ratio, and log odds evaluation in early-stage pancreatic ductal adenocarcinoma: numerology or valid indicators of patient outcome?

Science.gov (United States)

Lahat, G; Lubezky, N; Gerstenhaber, F; Nizri, E; Gysi, M; Rozenek, M; Goichman, Y; Nachmany, I; Nakache, R; Wolf, I; Klausner, J M

2016-09-29

We evaluated the prognostic significance and universal validity of the total number of evaluated lymph nodes (ELN), number of positive lymph nodes (PLN), lymph node ratio (LNR), and log odds of positive lymph nodes (LODDS) in a relatively large and homogenous cohort of surgically treated pancreatic ductal adenocarcinoma (PDAC) patients. Prospectively accrued data were retrospectively analyzed for 282 PDAC patients who had pancreaticoduodenectomy (PD) at our institution. Long-term survival was analyzed according to the ELN, PLN, LNR, and LODDS. Of these patients, 168 patients (59.5 %) had LN metastasis (N1). Mean ELN and PLN were 13.5 and 1.6, respectively. LN positivity correlated with a greater number of evaluated lymph nodes; positive lymph nodes were identified in 61.4 % of the patients with ELN ≥ 13 compared with 44.9 % of the patients with ELN < 13 (p = 0.014). Median overall survival (OS) and 5-year OS rate were higher in N0 than in N1 patients, 22.4 vs. 18.7 months and 35 vs. 11 %, respectively (p = 0.008). Mean LNR was 0.12; 91 patients (54.1 %) had LNR < 0.3. Among the N1 patients, median OS was comparable in those with LNR ≥ 0.3 vs. LNR < 0.3 (16.7 vs. 14.1 months, p = 0.950). Neither LODDS nor various ELN and PLN cutoff values provided more discriminative information within the group of N1 patients. Our data confirms that lymph node positivity strongly reflects PDAC biology and thus patient outcome. While a higher number of evaluated lymph nodes may provide a more accurate nodal staging, it does not have any prognostic value among N1 patients. Similarly, PLN, LNR, and LODDS had limited prognostic relevance.

17. Calibration of two complex ecosystem models with different likelihood functions

Science.gov (United States)

Hidy, Dóra; Haszpra, László; Pintér, Krisztina; Nagy, Zoltán; Barcza, Zoltán

2014-05-01

The biosphere is a sensitive carbon reservoir. Terrestrial ecosystems were approximately carbon neutral during the past centuries, but they became net carbon sinks due to climate change induced environmental change and associated CO2 fertilization effect of the atmosphere. Model studies and measurements indicate that the biospheric carbon sink can saturate in the future due to ongoing climate change which can act as a positive feedback. Robustness of carbon cycle models is a key issue when trying to choose the appropriate model for decision support. The input parameters of the process-based models are decisive regarding the model output. At the same time there are several input parameters for which accurate values are hard to obtain directly from experiments or no local measurements are available. Due to the uncertainty associated with the unknown model parameters significant bias can be experienced if the model is used to simulate the carbon and nitrogen cycle components of different ecosystems. In order to improve model performance the unknown model parameters has to be estimated. We developed a multi-objective, two-step calibration method based on Bayesian approach in order to estimate the unknown parameters of PaSim and Biome-BGC models. Biome-BGC and PaSim are a widely used biogeochemical models that simulate the storage and flux of water, carbon, and nitrogen between the ecosystem and the atmosphere, and within the components of the terrestrial ecosystems (in this research the developed version of Biome-BGC is used which is referred as BBGC MuSo). Both models were calibrated regardless the simulated processes and type of model parameters. The calibration procedure is based on the comparison of measured data with simulated results via calculating a likelihood function (degree of goodness-of-fit between simulated and measured data). In our research different likelihood function formulations were used in order to examine the effect of the different model

18. Variability of signal-to-noise ratio and the network analysis of gravitational wave burst signals

International Nuclear Information System (INIS)

Mohanty, S D; Rakhmanov, M; Klimenko, S; Mitselmakher, G

2006-01-01

The detection and estimation of gravitational wave burst signals, with a priori unknown polarization waveforms, requires the use of data from a network of detectors. Maximizing the network likelihood functional over all waveforms and sky positions yields point estimates for them as well as a detection statistic. However, the transformation from the data to estimates can become ill-conditioned over parts of the sky, resulting in significant errors in estimation. We modify the likelihood procedure by introducing a penalty functional which suppresses candidate solutions that display large signal-to-noise ratio (SNR) variability as the source is displaced on the sky. Simulations show that the resulting network analysis method performs significantly better in estimating the sky position of a source. Further, this method can be applied to any network, irrespective of the number or mutual alignment of detectors

19. Evaluation of robustness of maximum likelihood cone-beam CT reconstruction with total variation regularization

International Nuclear Information System (INIS)

Stsepankou, D; Arns, A; Hesser, J; Ng, S K; Zygmanski, P

2012-01-01

The objective of this paper is to evaluate an iterative maximum likelihood (ML) cone–beam computed tomography (CBCT) reconstruction with total variation (TV) regularization with respect to the robustness of the algorithm due to data inconsistencies. Three different and (for clinical application) typical classes of errors are considered for simulated phantom and measured projection data: quantum noise, defect detector pixels and projection matrix errors. To quantify those errors we apply error measures like mean square error, signal-to-noise ratio, contrast-to-noise ratio and streak indicator. These measures are derived from linear signal theory and generalized and applied for nonlinear signal reconstruction. For quality check, we focus on resolution and CT-number linearity based on a Catphan phantom. All comparisons are made versus the clinical standard, the filtered backprojection algorithm (FBP). In our results, we confirm and substantially extend previous results on iterative reconstruction such as massive undersampling of the number of projections. Errors of projection matrix parameters of up to 1° projection angle deviations are still in the tolerance level. Single defect pixels exhibit ring artifacts for each method. However using defect pixel compensation, allows up to 40% of defect pixels for passing the standard clinical quality check. Further, the iterative algorithm is extraordinarily robust in the low photon regime (down to 0.05 mAs) when compared to FPB, allowing for extremely low-dose image acquisitions, a substantial issue when considering daily CBCT imaging for position correction in radiotherapy. We conclude that the ML method studied herein is robust under clinical quality assurance conditions. Consequently, low-dose regime imaging, especially for daily patient localization in radiation therapy is possible without change of the current hardware of the imaging system. (paper)

20. Efficient Detection of Repeating Sites to Accelerate Phylogenetic Likelihood Calculations.

Science.gov (United States)

Kobert, K; Stamatakis, A; Flouri, T

2017-03-01

The phylogenetic likelihood function (PLF) is the major computational bottleneck in several applications of evolutionary biology such as phylogenetic inference, species delimitation, model selection, and divergence times estimation. Given the alignment, a tree and the evolutionary model parameters, the likelihood function computes the conditional likelihood vectors for every node of the tree. Vector entries for which all input data are identical result in redundant likelihood operations which, in turn, yield identical conditional values. Such operations can be omitted for improving run-time and, using appropriate data structures, reducing memory usage. We present a fast, novel method for identifying and omitting such redundant operations in phylogenetic likelihood calculations, and assess the performance improvement and memory savings attained by our method. Using empirical and simulated data sets, we show that a prototype implementation of our method yields up to 12-fold speedups and uses up to 78% less memory than one of the fastest and most highly tuned implementations of the PLF currently available. Our method is generic and can seamlessly be integrated into any phylogenetic likelihood implementation. [Algorithms; maximum likelihood; phylogenetic likelihood function; phylogenetics]. © The Author(s) 2016. Published by Oxford University Press, on behalf of the Society of Systematic Biologists.

1. Planck intermediate results: XVI. Profile likelihoods for cosmological parameters

DEFF Research Database (Denmark)

Bartlett, J.G.; Cardoso, J.-F.; Delabrouille, J.

2014-01-01

We explore the 2013 Planck likelihood function with a high-precision multi-dimensional minimizer (Minuit). This allows a refinement of the CDM best-fit solution with respect to previously-released results, and the construction of frequentist confidence intervals using profile likelihoods. The agr...

2. Planck 2013 results. XV. CMB power spectra and likelihood

DEFF Research Database (Denmark)

Tauber, Jan; Bartlett, J.G.; Bucher, M.

2014-01-01

This paper presents the Planck 2013 likelihood, a complete statistical description of the two-point correlation function of the CMB temperature fluctuations that accounts for all known relevant uncertainties, both instrumental and astrophysical in nature. We use this likelihood to derive our best...

3. Likelihood analysis of parity violation in the compound nucleus

International Nuclear Information System (INIS)

Bowman, D.; Sharapov, E.

1993-01-01

We discuss the determination of the root mean-squared matrix element of the parity-violating interaction between compound-nuclear states using likelihood analysis. We briefly review the relevant features of the statistical model of the compound nucleus and the formalism of likelihood analysis. We then discuss the application of likelihood analysis to data on panty-violating longitudinal asymmetries. The reliability of the extracted value of the matrix element and errors assigned to the matrix element is stressed. We treat the situations where the spins of the p-wave resonances are not known and known using experimental data and Monte Carlo techniques. We conclude that likelihood analysis provides a reliable way to determine M and its confidence interval. We briefly discuss some problems associated with the normalization of the likelihood function

4. Constructing valid density matrices on an NMR quantum information processor via maximum likelihood estimation

Energy Technology Data Exchange (ETDEWEB)

Singh, Harpreet; Arvind; Dorai, Kavita, E-mail: kavita@iisermohali.ac.in

2016-09-07

Estimation of quantum states is an important step in any quantum information processing experiment. A naive reconstruction of the density matrix from experimental measurements can often give density matrices which are not positive, and hence not physically acceptable. How do we ensure that at all stages of reconstruction, we keep the density matrix positive? Recently a method has been suggested based on maximum likelihood estimation, wherein the density matrix is guaranteed to be positive definite. We experimentally implement this protocol on an NMR quantum information processor. We discuss several examples and compare with the standard method of state estimation. - Highlights: • State estimation using maximum likelihood method was performed on an NMR quantum information processor. • Physically valid density matrices were obtained every time in contrast to standard quantum state tomography. • Density matrices of several different entangled and separable states were reconstructed for two and three qubits.

5. Supervisor Autonomy and Considerate Leadership Style are Associated with Supervisors' Likelihood to Accommodate Back Injured Workers.

Science.gov (United States)

McGuire, Connor; Kristman, Vicki L; Shaw, William; Williams-Whitt, Kelly; Reguly, Paula; Soklaridis, Sophie

2015-09-01

To determine the association between supervisors' leadership style and autonomy and supervisors' likelihood of supporting job accommodations for back-injured workers. A cross-sectional study of supervisors from Canadian and US employers was conducted using a web-based, self-report questionnaire that included a case vignette of a back-injured worker. Autonomy and two dimensions of leadership style (considerate and initiating structure) were included as exposures. The outcome, supervisors' likeliness to support job accommodation, was measured with the Job Accommodation Scale (JAS). We conducted univariate analyses of all variables and bivariate analyses of the JAS score with each exposure and potential confounding factor. We used multivariable generalized linear models to control for confounding factors. A total of 796 supervisors participated. Considerate leadership style (β = .012; 95% CI .009-.016) and autonomy (β = .066; 95% CI .025-.11) were positively associated with supervisors' likelihood to accommodate after adjusting for appropriate confounding factors. An initiating structure leadership style was not significantly associated with supervisors' likelihood to accommodate (β = .0018; 95% CI -.0026 to .0061) after adjusting for appropriate confounders. Autonomy and a considerate leadership style were positively associated with supervisors' likelihood to accommodate a back-injured worker. Providing supervisors with more autonomy over decisions of accommodation and developing their considerate leadership style may aid in increasing work accommodation for back-injured workers and preventing prolonged work disability.

6. Supervisor Autonomy and Considerate Leadership Style are Associated with Supervisors’ Likelihood to Accommodate Back Injured Workers

Science.gov (United States)

McGuire, Connor; Kristman, Vicki L; Williams-Whitt, Kelly; Reguly, Paula; Shaw, William; Soklaridis, Sophie

2015-01-01

PURPOSE To determine the association between supervisors’ leadership style and autonomy and supervisors’ likelihood of supporting job accommodations for back-injured workers. METHODS A cross-sectional study of supervisors from Canadian and US employers was conducted using a web-based, self-report questionnaire that included a case vignette of a back-injured worker. Autonomy and two dimensions of leadership style (considerate and initiating structure) were included as exposures. The outcome, supervisors’ likeliness to support job accommodation, was measured with the Job Accommodation Scale. We conducted univariate analyses of all variables and bivariate analyses of the JAS score with each exposure and potential confounding factor. We used multivariable generalized linear models to control for confounding factors. RESULTS A total of 796 supervisors participated. Considerate leadership style (β= .012; 95% CI: .009–.016) and autonomy (β= .066; 95% CI: .025–.11) were positively associated with supervisors’ likelihood to accommodate after adjusting for appropriate confounding factors. An initiating structure leadership style was not significantly associated with supervisors’ likelihood to accommodate (β = .0018; 95% CI: −.0026–.0061) after adjusting for appropriate confounders. CONCLUSIONS Autonomy and a considerate leadership style were positively associated with supervisors’ likelihood to accommodate a back-injured worker. Providing supervisors with more autonomy over decisions of accommodation and developing their considerate leadership style may aid in increasing work accommodation for back-injured workers and preventing prolonged work disability. PMID:25595332

7. High p95HER2/HER2 Ratio Associated With Poor Outcome in Trastuzumab-Treated HER2-Positive Metastatic Breast Cancer NCCTG N0337 and NCCTG 98-32-52 (Alliance).

Science.gov (United States)

Chumsri, Saranya; Sperinde, Jeff; Liu, Heshan; Gligorov, Joseph; Spano, Jean-Philippe; Antoine, Martine; Moreno Aspitia, Alvaro; Tan, Winston; Winslow, John; Petropoulos, Christos J; Chenna, Ahmed; Bates, Michael; Weidler, Jodi Marie; Huang, Weidong; Dueck, Amylou; Perez, Edith A

2018-03-12

Purpose: p95HER2 is a truncated form of HER2 that confers resistance to trastuzumab in vitro , but clinical results have been conflicting to date. Given that p95HER2 levels correlate with total HER2 expression levels, which confer better outcomes, we sought to evaluate the p95HER2/HER2 ratio in the North Central Cancer Treatment Group N0337 and N98-32-52 trials. Experimental Design: The HERmark assay and VeraTag technology (Monogram Biosciences) were used to measure total HER2 and p95HER2 expression levels in 91 patient samples. Results: In the multivariate model, increasing total HER2 level was significantly associated with longer (OS; HR, 0.33; P = 0.002) and decreasing p95HER2 level was significantly associated with longer OS (HR, 4.2; P = 0.01). Total HER2 expression level was significantly associated with longer progression-free survival (PFS) (HR, 0.57; P = 0.04), whereas p95HER2 level was not (HR, 1.7; P = 0.25). However, there was a positive association between p95HER2 and total HER2 expression levels ( R 2 = 0.48; P < 0.001). Consistent with our hypothesis, the ratio of p95HER2/HER2 was significantly associated with worsening PFS (HR, 1.7; P = 0.04) and OS (HR, 2.8; P = 0.002). Patients with the highest tertile of p95HER2/HER2 values had significantly less favorable PFS (HR, 1.8; P = 0.06) and OS (HR, 2.3; P = 0.02). Conclusions: A high p95HER2/HER2 ratio identified patients with metastatic breast cancer with poor outcomes on trastuzumab-based therapies. Further investigation of the p95HER2/HER2 ratio as a potential prognostic or predictive biomarker for HER2-targeted therapy is warranted. Clin Cancer Res; 1-6. ©2018 AACR. ©2018 American Association for Cancer Research.

8. Likelihood for transcriptions in a genetic regulatory system under asymmetric stable Lévy noise.

Science.gov (United States)

Wang, Hui; Cheng, Xiujun; Duan, Jinqiao; Kurths, Jürgen; Li, Xiaofan

2018-01-01

This work is devoted to investigating the evolution of concentration in a genetic regulation system, when the synthesis reaction rate is under additive and multiplicative asymmetric stable Lévy fluctuations. By focusing on the impact of skewness (i.e., non-symmetry) in the probability distributions of noise, we find that via examining the mean first exit time (MFET) and the first escape probability (FEP), the asymmetric fluctuations, interacting with nonlinearity in the system, lead to peculiar likelihood for transcription. This includes, in the additive noise case, realizing higher likelihood of transcription for larger positive skewness (i.e., asymmetry) index β, causing a stochastic bifurcation at the non-Gaussianity index value α = 1 (i.e., it is a separating point or line for the likelihood for transcription), and achieving a turning point at the threshold value β≈-0.5 (i.e., beyond which the likelihood for transcription suddenly reversed for α values). The stochastic bifurcation and turning point phenomena do not occur in the symmetric noise case (β = 0). While in the multiplicative noise case, non-Gaussianity index value α = 1 is a separating point or line for both the MFET and the FEP. We also investigate the noise enhanced stability phenomenon. Additionally, we are able to specify the regions in the whole parameter space for the asymmetric noise, in which we attain desired likelihood for transcription. We have conducted a series of numerical experiments in "regulating" the likelihood of gene transcription by tuning asymmetric stable Lévy noise indexes. This work offers insights for possible ways of achieving gene regulation in experimental research.

9. Obstetric History and Likelihood of Preterm Birth of Twins.

Science.gov (United States)

Easter, Sarah Rae; Little, Sarah E; Robinson, Julian N; Mendez-Figueroa, Hector; Chauhan, Suneet P

2018-01-05

The objective of this study was to investigate the relationship between preterm birth in a prior pregnancy and preterm birth in a twin pregnancy.  We performed a secondary analysis of a randomized controlled trial evaluating 17-α-hydroxyprogesterone caproate in twins. Women were classified as nulliparous, multiparous with a prior term birth, or multiparous with a prior preterm birth. We used logistic regression to examine the odds of spontaneous preterm birth of twins before 35 weeks according to past obstetric history.  Of the 653 women analyzed, 294 were nulliparas, 310 had a prior term birth, and 49 had a prior preterm birth. Prior preterm birth increased the likelihood of spontaneous delivery before 35 weeks (adjusted odds ratio [aOR]: 2.44, 95% confidence interval [CI]: 1.28-4.66), whereas prior term delivery decreased these odds (aOR: 0.55, 95% CI: 0.38-0.78) in the current twin pregnancy compared with the nulliparous reference group. This translated into a lower odds of composite neonatal morbidity (aOR: 0.38, 95% CI: 0.27-0.53) for women with a prior term delivery.  For women carrying twins, a history of preterm birth increases the odds of spontaneous preterm birth, whereas a prior term birth decreases odds of spontaneous preterm birth and neonatal morbidity for the current twin pregnancy. These results offer risk stratification and reassurance for clinicians. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

10. The Influence of Total Nodes Examined, Number of Positive Nodes, and Lymph Node Ratio on Survival After Surgical Resection and Adjuvant Chemoradiation for Pancreatic Cancer: A Secondary Analysis of RTOG 9704

Energy Technology Data Exchange (ETDEWEB)

2011-12-01

Purpose: Lymph node status is an important predictor of survival in pancreatic cancer. We performed a secondary analysis of Radiation Therapy Oncology Group (RTOG) 9704, an adjuvant chemotherapy and chemoradiation trial, to determine the influence of lymph node factors-number of positive nodes (NPN), total nodes examined (TNE), and lymph node ratio (LNR ratio of NPN to TNE)-on OS and disease-free survival (DFS). Patient and Methods: Eligible patients from RTOG 9704 form the basis of this secondary analysis of lymph node parameters. Actuarial estimates for OS and DFS were calculated using Kaplan-Meier methods. Cox proportional hazards models were performed to evaluate associations of NPN, TNE, and LNR with OS and DFS. Multivariate Cox proportional hazards models were also performed. Results: There were 538 patients enrolled in the RTOG 9704 trial. Of these, 445 patients were eligible with lymph nodes removed. Overall median NPN was 1 (min-max, 0-18). Increased NPN was associated with worse OS (HR = 1.06, p = 0.001) and DFS (HR = 1.05, p = 0.01). In multivariate analyses, both NPN and TNE were associated with OS and DFS. TNE > 12, and >15 were associated with increased OS for all patients, but not for node-negative patients (n = 142). Increased LNR was associated with worse OS (HR = 1.01, p < 0.0001) and DFS (HR = 1.006, p = 0.002). Conclusion: In patients who undergo surgical resection followed by adjuvant chemoradiation, TNE, NPN, and LNR are associated with OS and DFS. This secondary analysis of a prospective, cooperative group trial supports the influence of these lymph node parameters on outcomes after surgery and adjuvant therapy using contemporary techniques.

11. Practical likelihood analysis for spatial generalized linear mixed models

DEFF Research Database (Denmark)

Bonat, W. H.; Ribeiro, Paulo Justiniano

2016-01-01

We investigate an algorithm for maximum likelihood estimation of spatial generalized linear mixed models based on the Laplace approximation. We compare our algorithm with a set of alternative approaches for two datasets from the literature. The Rhizoctonia root rot and the Rongelap are......, respectively, examples of binomial and count datasets modeled by spatial generalized linear mixed models. Our results show that the Laplace approximation provides similar estimates to Markov Chain Monte Carlo likelihood, Monte Carlo expectation maximization, and modified Laplace approximation. Some advantages...... of Laplace approximation include the computation of the maximized log-likelihood value, which can be used for model selection and tests, and the possibility to obtain realistic confidence intervals for model parameters based on profile likelihoods. The Laplace approximation also avoids the tuning...

12. Algorithms of maximum likelihood data clustering with applications

Science.gov (United States)

Giada, Lorenzo; Marsili, Matteo

2002-12-01

We address the problem of data clustering by introducing an unsupervised, parameter-free approach based on maximum likelihood principle. Starting from the observation that data sets belonging to the same cluster share a common information, we construct an expression for the likelihood of any possible cluster structure. The likelihood in turn depends only on the Pearson's coefficient of the data. We discuss clustering algorithms that provide a fast and reliable approximation to maximum likelihood configurations. Compared to standard clustering methods, our approach has the advantages that (i) it is parameter free, (ii) the number of clusters need not be fixed in advance and (iii) the interpretation of the results is transparent. In order to test our approach and compare it with standard clustering algorithms, we analyze two very different data sets: time series of financial market returns and gene expression data. We find that different maximization algorithms produce similar cluster structures whereas the outcome of standard algorithms has a much wider variability.

13. Generalized empirical likelihood methods for analyzing longitudinal data

KAUST Repository

Wang, S.; Qian, L.; Carroll, R. J.

2010-01-01

Efficient estimation of parameters is a major objective in analyzing longitudinal data. We propose two generalized empirical likelihood based methods that take into consideration within-subject correlations. A nonparametric version of the Wilks

14. Maximum likelihood estimation of finite mixture model for economic data

Science.gov (United States)

Phoong, Seuk-Yen; Ismail, Mohd Tahir

2014-06-01

Finite mixture model is a mixture model with finite-dimension. This models are provides a natural representation of heterogeneity in a finite number of latent classes. In addition, finite mixture models also known as latent class models or unsupervised learning models. Recently, maximum likelihood estimation fitted finite mixture models has greatly drawn statistician's attention. The main reason is because maximum likelihood estimation is a powerful statistical method which provides consistent findings as the sample sizes increases to infinity. Thus, the application of maximum likelihood estimation is used to fit finite mixture model in the present paper in order to explore the relationship between nonlinear economic data. In this paper, a two-component normal mixture model is fitted by maximum likelihood estimation in order to investigate the relationship among stock market price and rubber price for sampled countries. Results described that there is a negative effect among rubber price and stock market price for Malaysia, Thailand, Philippines and Indonesia.

15. Attitude towards, and likelihood of, complaining in the banking ...

African Journals Online (AJOL)

aims to determine customers' attitudes towards complaining as well as their likelihood of voicing a .... is particularly powerful and impacts greatly on customer satisfaction and retention. ...... 'Cross-national analysis of hotel customers' attitudes ...

16. Narrow band interference cancelation in OFDM: Astructured maximum likelihood approach

KAUST Repository

2012-01-01

This paper presents a maximum likelihood (ML) approach to mitigate the effect of narrow band interference (NBI) in a zero padded orthogonal frequency division multiplexing (ZP-OFDM) system. The NBI is assumed to be time variant and asynchronous

17. On the likelihood function of Gaussian max-stable processes

KAUST Repository

Genton, M. G.; Ma, Y.; Sang, H.

2011-01-01

We derive a closed form expression for the likelihood function of a Gaussian max-stable process indexed by ℝd at p≤d+1 sites, d≥1. We demonstrate the gain in efficiency in the maximum composite likelihood estimators of the covariance matrix from p=2 to p=3 sites in ℝ2 by means of a Monte Carlo simulation study. © 2011 Biometrika Trust.

18. Incorporating Nuisance Parameters in Likelihoods for Multisource Spectra

CERN Document Server

Conway, J.S.

2011-01-01

We describe here the general mathematical approach to constructing likelihoods for fitting observed spectra in one or more dimensions with multiple sources, including the effects of systematic uncertainties represented as nuisance parameters, when the likelihood is to be maximized with respect to these parameters. We consider three types of nuisance parameters: simple multiplicative factors, source spectra "morphing" parameters, and parameters representing statistical uncertainties in the predicted source spectra.

19. On the likelihood function of Gaussian max-stable processes

KAUST Repository

Genton, M. G.

2011-05-24

We derive a closed form expression for the likelihood function of a Gaussian max-stable process indexed by ℝd at p≤d+1 sites, d≥1. We demonstrate the gain in efficiency in the maximum composite likelihood estimators of the covariance matrix from p=2 to p=3 sites in ℝ2 by means of a Monte Carlo simulation study. © 2011 Biometrika Trust.

20. Tapered composite likelihood for spatial max-stable models

KAUST Repository

Sang, Huiyan

2014-05-01

Spatial extreme value analysis is useful to environmental studies, in which extreme value phenomena are of interest and meaningful spatial patterns can be discerned. Max-stable process models are able to describe such phenomena. This class of models is asymptotically justified to characterize the spatial dependence among extremes. However, likelihood inference is challenging for such models because their corresponding joint likelihood is unavailable and only bivariate or trivariate distributions are known. In this paper, we propose a tapered composite likelihood approach by utilizing lower dimensional marginal likelihoods for inference on parameters of various max-stable process models. We consider a weighting strategy based on a "taper range" to exclude distant pairs or triples. The "optimal taper range" is selected to maximize various measures of the Godambe information associated with the tapered composite likelihood function. This method substantially reduces the computational cost and improves the efficiency over equally weighted composite likelihood estimators. We illustrate its utility with simulation experiments and an analysis of rainfall data in Switzerland.

1. Dissociating response conflict and error likelihood in anterior cingulate cortex.

Science.gov (United States)

Yeung, Nick; Nieuwenhuis, Sander

2009-11-18

Neuroimaging studies consistently report activity in anterior cingulate cortex (ACC) in conditions of high cognitive demand, leading to the view that ACC plays a crucial role in the control of cognitive processes. According to one prominent theory, the sensitivity of ACC to task difficulty reflects its role in monitoring for the occurrence of competition, or "conflict," between responses to signal the need for increased cognitive control. However, a contrasting theory proposes that ACC is the recipient rather than source of monitoring signals, and that ACC activity observed in relation to task demand reflects the role of this region in learning about the likelihood of errors. Response conflict and error likelihood are typically confounded, making the theories difficult to distinguish empirically. The present research therefore used detailed computational simulations to derive contrasting predictions regarding ACC activity and error rate as a function of response speed. The simulations demonstrated a clear dissociation between conflict and error likelihood: fast response trials are associated with low conflict but high error likelihood, whereas slow response trials show the opposite pattern. Using the N2 component as an index of ACC activity, an EEG study demonstrated that when conflict and error likelihood are dissociated in this way, ACC activity tracks conflict and is negatively correlated with error likelihood. These findings support the conflict-monitoring theory and suggest that, in speeded decision tasks, ACC activity reflects current task demands rather than the retrospective coding of past performance.

2. Tapered composite likelihood for spatial max-stable models

KAUST Repository

Sang, Huiyan; Genton, Marc G.

2014-01-01

Spatial extreme value analysis is useful to environmental studies, in which extreme value phenomena are of interest and meaningful spatial patterns can be discerned. Max-stable process models are able to describe such phenomena. This class of models is asymptotically justified to characterize the spatial dependence among extremes. However, likelihood inference is challenging for such models because their corresponding joint likelihood is unavailable and only bivariate or trivariate distributions are known. In this paper, we propose a tapered composite likelihood approach by utilizing lower dimensional marginal likelihoods for inference on parameters of various max-stable process models. We consider a weighting strategy based on a "taper range" to exclude distant pairs or triples. The "optimal taper range" is selected to maximize various measures of the Godambe information associated with the tapered composite likelihood function. This method substantially reduces the computational cost and improves the efficiency over equally weighted composite likelihood estimators. We illustrate its utility with simulation experiments and an analysis of rainfall data in Switzerland.

3. Behavior of the maximum likelihood in quantum state tomography

Science.gov (United States)

Scholten, Travis L.; Blume-Kohout, Robin

2018-02-01

Quantum state tomography on a d-dimensional system demands resources that grow rapidly with d. They may be reduced by using model selection to tailor the number of parameters in the model (i.e., the size of the density matrix). Most model selection methods typically rely on a test statistic and a null theory that describes its behavior when two models are equally good. Here, we consider the loglikelihood ratio. Because of the positivity constraint ρ ≥ 0, quantum state space does not generally satisfy local asymptotic normality (LAN), meaning the classical null theory for the loglikelihood ratio (the Wilks theorem) should not be used. Thus, understanding and quantifying how positivity affects the null behavior of this test statistic is necessary for its use in model selection for state tomography. We define a new generalization of LAN, metric-projected LAN, show that quantum state space satisfies it, and derive a replacement for the Wilks theorem. In addition to enabling reliable model selection, our results shed more light on the qualitative effects of the positivity constraint on state tomography.

4. Behavior of the maximum likelihood in quantum state tomography

Energy Technology Data Exchange (ETDEWEB)

Blume-Kohout, Robin J [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Univ. of New Mexico, Albuquerque, NM (United States); Scholten, Travis L. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Univ. of New Mexico, Albuquerque, NM (United States)

2018-02-22

Quantum state tomography on a d-dimensional system demands resources that grow rapidly with d. They may be reduced by using model selection to tailor the number of parameters in the model (i.e., the size of the density matrix). Most model selection methods typically rely on a test statistic and a null theory that describes its behavior when two models are equally good. Here, we consider the loglikelihood ratio. Because of the positivity constraint ρ ≥ 0, quantum state space does not generally satisfy local asymptotic normality (LAN), meaning the classical null theory for the loglikelihood ratio (the Wilks theorem) should not be used. Thus, understanding and quantifying how positivity affects the null behavior of this test statistic is necessary for its use in model selection for state tomography. We define a new generalization of LAN, metric-projected LAN, show that quantum state space satisfies it, and derive a replacement for the Wilks theorem. In addition to enabling reliable model selection, our results shed more light on the qualitative effects of the positivity constraint on state tomography.

5. Framing the frame: How task goals determine the likelihood and direction of framing effects

OpenAIRE

Todd McElroy; John J. Seta

2007-01-01

We examined how the goal of a decision task influences the perceived positive, negative valence of the alternatives and thereby the likelihood and direction of framing effects. In Study 1 we manipulated the goal to increase, decrease or maintain the commodity in question and found that when the goal of the task was to increase the commodity, a framing effect consistent with those typically observed in the literature was found. When the goal was to decrease, a framing effect opposite to the ty...

6. Image properties of list mode likelihood reconstruction for a rectangular positron emission mammography with DOI measurements

International Nuclear Information System (INIS)

Qi, Jinyi; Klein, Gregory J.; Huesman, Ronald H.

2000-01-01

A positron emission mammography scanner is under development at our Laboratory. The tomograph has a rectangular geometry consisting of four banks of detector modules. For each detector, the system can measure the depth of interaction information inside the crystal. The rectangular geometry leads to irregular radial and angular sampling and spatially variant sensitivity that are different from conventional PET systems. Therefore, it is of importance to study the image properties of the reconstructions. We adapted the theoretical analysis that we had developed for conventional PET systems to the list mode likelihood reconstruction for this tomograph. The local impulse response and covariance of the reconstruction can be easily computed using FFT. These theoretical results are also used with computer observer models to compute the signal-to-noise ratio for lesion detection. The analysis reveals the spatially variant resolution and noise properties of the list mode likelihood reconstruction. The theoretical predictions are in good agreement with Monte Carlo results

7. Clarification of the use of chi-square and likelihood functions in fits to histograms

International Nuclear Information System (INIS)

Baker, S.; Cousins, R.D.

1984-01-01

We consider the problem of fitting curves to histograms in which the data obey multinomial or Poisson statistics. Techniques commonly used by physicists are examined in light of standard results found in the statistics literature. We review the relationship between multinomial and Poisson distributions, and clarify a sufficient condition for equality of the area under the fitted curve and the number of events on the histogram. Following the statisticians, we use the likelihood ratio test to construct a general Z 2 statistic, Zsub(lambda) 2 , which yields parameter and error estimates identical to those of the method of maximum likelihood. The Zsub(lambda) 2 statistic is further useful for testing goodness-of-fit since the value of its minimum asymptotically obeys a classical chi-square distribution. One should be aware, however, of the potential for statistical bias, especially when the number of events is small. (orig.)

8. Cosmic shear measurement with maximum likelihood and maximum a posteriori inference

Science.gov (United States)

Hall, Alex; Taylor, Andy

2017-06-01

We investigate the problem of noise bias in maximum likelihood and maximum a posteriori estimators for cosmic shear. We derive the leading and next-to-leading order biases and compute them in the context of galaxy ellipticity measurements, extending previous work on maximum likelihood inference for weak lensing. We show that a large part of the bias on these point estimators can be removed using information already contained in the likelihood when a galaxy model is specified, without the need for external calibration. We test these bias-corrected estimators on simulated galaxy images similar to those expected from planned space-based weak lensing surveys, with promising results. We find that the introduction of an intrinsic shape prior can help with mitigation of noise bias, such that the maximum a posteriori estimate can be made less biased than the maximum likelihood estimate. Second-order terms offer a check on the convergence of the estimators, but are largely subdominant. We show how biases propagate to shear estimates, demonstrating in our simple set-up that shear biases can be reduced by orders of magnitude and potentially to within the requirements of planned space-based surveys at mild signal-to-noise ratio. We find that second-order terms can exhibit significant cancellations at low signal-to-noise ratio when Gaussian noise is assumed, which has implications for inferring the performance of shear-measurement algorithms from simplified simulations. We discuss the viability of our point estimators as tools for lensing inference, arguing that they allow for the robust measurement of ellipticity and shear.

9. Profile-likelihood Confidence Intervals in Item Response Theory Models.

Science.gov (United States)

Chalmers, R Philip; Pek, Jolynn; Liu, Yang

2017-01-01

Confidence intervals (CIs) are fundamental inferential devices which quantify the sampling variability of parameter estimates. In item response theory, CIs have been primarily obtained from large-sample Wald-type approaches based on standard error estimates, derived from the observed or expected information matrix, after parameters have been estimated via maximum likelihood. An alternative approach to constructing CIs is to quantify sampling variability directly from the likelihood function with a technique known as profile-likelihood confidence intervals (PL CIs). In this article, we introduce PL CIs for item response theory models, compare PL CIs to classical large-sample Wald-type CIs, and demonstrate important distinctions among these CIs. CIs are then constructed for parameters directly estimated in the specified model and for transformed parameters which are often obtained post-estimation. Monte Carlo simulation results suggest that PL CIs perform consistently better than Wald-type CIs for both non-transformed and transformed parameters.

10. Maximum Likelihood-Based Methods for Target Velocity Estimation with Distributed MIMO Radar

Directory of Open Access Journals (Sweden)

Zhenxin Cao

2018-02-01

Full Text Available The estimation problem for target velocity is addressed in this in the scenario with a distributed multi-input multi-out (MIMO radar system. A maximum likelihood (ML-based estimation method is derived with the knowledge of target position. Then, in the scenario without the knowledge of target position, an iterative method is proposed to estimate the target velocity by updating the position information iteratively. Moreover, the Carmér-Rao Lower Bounds (CRLBs for both scenarios are derived, and the performance degradation of velocity estimation without the position information is also expressed. Simulation results show that the proposed estimation methods can approach the CRLBs, and the velocity estimation performance can be further improved by increasing either the number of radar antennas or the information accuracy of the target position. Furthermore, compared with the existing methods, a better estimation performance can be achieved.

11. Unbinned likelihood maximisation framework for neutrino clustering in Python

Energy Technology Data Exchange (ETDEWEB)

Coenders, Stefan [Technische Universitaet Muenchen, Boltzmannstr. 2, 85748 Garching (Germany)

2016-07-01

Albeit having detected an astrophysical neutrino flux with IceCube, sources of astrophysical neutrinos remain hidden up to now. A detection of a neutrino point source is a smoking gun for hadronic processes and acceleration of cosmic rays. The search for neutrino sources has many degrees of freedom, for example steady versus transient, point-like versus extended sources, et cetera. Here, we introduce a Python framework designed for unbinned likelihood maximisations as used in searches for neutrino point sources by IceCube. Implementing source scenarios in a modular way, likelihood searches on various kinds can be implemented in a user-friendly way, without sacrificing speed and memory management.

12. A note on estimating errors from the likelihood function

International Nuclear Information System (INIS)

Barlow, Roger

2005-01-01

The points at which the log likelihood falls by 12 from its maximum value are often used to give the 'errors' on a result, i.e. the 68% central confidence interval. The validity of this is examined for two simple cases: a lifetime measurement and a Poisson measurement. Results are compared with the exact Neyman construction and with the simple Bartlett approximation. It is shown that the accuracy of the log likelihood method is poor, and the Bartlett construction explains why it is flawed

13. Maximum Likelihood DOA Estimation of Multiple Wideband Sources in the Presence of Nonuniform Sensor Noise

Directory of Open Access Journals (Sweden)

K. Yao

2007-12-01

Full Text Available We investigate the maximum likelihood (ML direction-of-arrival (DOA estimation of multiple wideband sources in the presence of unknown nonuniform sensor noise. New closed-form expression for the direction estimation CramÃƒÂ©r-Rao-Bound (CRB has been derived. The performance of the conventional wideband uniform ML estimator under nonuniform noise has been studied. In order to mitigate the performance degradation caused by the nonuniformity of the noise, a new deterministic wideband nonuniform ML DOA estimator is derived and two associated processing algorithms are proposed. The first algorithm is based on an iterative procedure which stepwise concentrates the log-likelihood function with respect to the DOAs and the noise nuisance parameters, while the second is a noniterative algorithm that maximizes the derived approximately concentrated log-likelihood function. The performance of the proposed algorithms is tested through extensive computer simulations. Simulation results show the stepwise-concentrated ML algorithm (SC-ML requires only a few iterations to converge and both the SC-ML and the approximately-concentrated ML algorithm (AC-ML attain a solution close to the derived CRB at high signal-to-noise ratio.

14. A New Maximum-Likelihood Change Estimator for Two-Pass SAR Coherent Change Detection.

Energy Technology Data Exchange (ETDEWEB)

Wahl, Daniel E.; Yocky, David A.; Jakowatz, Charles V,

2014-09-01

In this paper, we derive a new optimal change metric to be used in synthetic aperture RADAR (SAR) coherent change detection (CCD). Previous CCD methods tend to produce false alarm states (showing change when there is none) in areas of the image that have a low clutter-to-noise power ratio (CNR). The new estimator does not suffer from this shortcoming. It is a surprisingly simple expression, easy to implement, and is optimal in the maximum-likelihood (ML) sense. The estimator produces very impressive results on the CCD collects that we have tested.

15. Maximum likelihood estimation of the attenuated ultrasound pulse

DEFF Research Database (Denmark)

Rasmussen, Klaus Bolding

1994-01-01

The attenuated ultrasound pulse is divided into two parts: a stationary basic pulse and a nonstationary attenuation pulse. A standard ARMA model is used for the basic pulse, and a nonstandard ARMA model is derived for the attenuation pulse. The maximum likelihood estimator of the attenuated...

16. Planck 2013 results. XV. CMB power spectra and likelihood

CERN Document Server

Ade, P.A.R.; Armitage-Caplan, C.; Arnaud, M.; Ashdown, M.; Atrio-Barandela, F.; Aumont, J.; Baccigalupi, C.; Banday, A.J.; Barreiro, R.B.; Bartlett, J.G.; Battaner, E.; Benabed, K.; Benoit, A.; Benoit-Levy, A.; Bernard, J.P.; Bersanelli, M.; Bielewicz, P.; Bobin, J.; Bock, J.J.; Bonaldi, A.; Bonavera, L.; Bond, J.R.; Borrill, J.; Bouchet, F.R.; Boulanger, F.; Bridges, M.; Bucher, M.; Burigana, C.; Butler, R.C.; Calabrese, E.; Cardoso, J.F.; Catalano, A.; Challinor, A.; Chamballu, A.; Chiang, L.Y.; Chiang, H.C.; Christensen, P.R.; Church, S.; Clements, D.L.; Colombi, S.; Colombo, L.P.L.; Combet, C.; Couchot, F.; Coulais, A.; Crill, B.P.; Curto, A.; Cuttaia, F.; Danese, L.; Davies, R.D.; Davis, R.J.; de Bernardis, P.; de Rosa, A.; de Zotti, G.; Delabrouille, J.; Delouis, J.M.; Desert, F.X.; Dickinson, C.; Diego, J.M.; Dole, H.; Donzelli, S.; Dore, O.; Douspis, M.; Dunkley, J.; Dupac, X.; Efstathiou, G.; Elsner, F.; Ensslin, T.A.; Eriksen, H.K.; Finelli, F.; Forni, O.; Frailis, M.; Fraisse, A.A.; Franceschi, E.; Gaier, T.C.; Galeotta, S.; Galli, S.; Ganga, K.; Giard, M.; Giardino, G.; Giraud-Heraud, Y.; Gjerlow, E.; Gonzalez-Nuevo, J.; Gorski, K.M.; Gratton, S.; Gregorio, A.; Gruppuso, A.; Gudmundsson, J.E.; Hansen, F.K.; Hanson, D.; Harrison, D.; Helou, G.; Henrot-Versille, S.; Hernandez-Monteagudo, C.; Herranz, D.; Hildebrandt, S.R.; Hivon, E.; Hobson, M.; Holmes, W.A.; Hornstrup, A.; Hovest, W.; Huffenberger, K.M.; Hurier, G.; Jaffe, T.R.; Jaffe, A.H.; Jewell, J.; Jones, W.C.; Juvela, M.; Keihanen, E.; Keskitalo, R.; Kiiveri, K.; Kisner, T.S.; Kneissl, R.; Knoche, J.; Knox, L.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lahteenmaki, A.; Lamarre, J.M.; Lasenby, A.; Lattanzi, M.; Laureijs, R.J.; Lawrence, C.R.; Le Jeune, M.; Leach, S.; Leahy, J.P.; Leonardi, R.; Leon-Tavares, J.; Lesgourgues, J.; Liguori, M.; Lilje, P.B.; Lindholm, V.; Linden-Vornle, M.; Lopez-Caniego, M.; Lubin, P.M.; Macias-Perez, J.F.; Maffei, B.; Maino, D.; Mandolesi, N.; Marinucci, D.; Maris, M.; Marshall, D.J.; Martin, P.G.; Martinez-Gonzalez, E.; Masi, S.; Matarrese, S.; Matthai, F.; Mazzotta, P.; Meinhold, P.R.; Melchiorri, A.; Mendes, L.; Menegoni, E.; Mennella, A.; Migliaccio, M.; Millea, M.; Mitra, S.; Miville-Deschenes, M.A.; Molinari, D.; Moneti, A.; Montier, L.; Morgante, G.; Mortlock, D.; Moss, A.; Munshi, D.; Naselsky, P.; Nati, F.; Natoli, P.; Netterfield, C.B.; Norgaard-Nielsen, H.U.; Noviello, F.; Novikov, D.; Novikov, I.; O'Dwyer, I.J.; Orieux, F.; Osborne, S.; Oxborrow, C.A.; Paci, F.; Pagano, L.; Pajot, F.; Paladini, R.; Paoletti, D.; Partridge, B.; Pasian, F.; Patanchon, G.; Paykari, P.; Perdereau, O.; Perotto, L.; Perrotta, F.; Piacentini, F.; Piat, M.; Pierpaoli, E.; Pietrobon, D.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Ponthieu, N.; Popa, L.; Poutanen, T.; Pratt, G.W.; Prezeau, G.; Prunet, S.; Puget, J.L.; Rachen, J.P.; Rahlin, A.; Rebolo, R.; Reinecke, M.; Remazeilles, M.; Renault, C.; Ricciardi, S.; Riller, T.; Ringeval, C.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Roudier, G.; Rowan-Robinson, M.; Rubino-Martin, J.A.; Rusholme, B.; Sandri, M.; Sanselme, L.; Santos, D.; Savini, G.; Scott, D.; Seiffert, M.D.; Shellard, E.P.S.; Spencer, L.D.; Starck, J.L.; Stolyarov, V.; Stompor, R.; Sudiwala, R.; Sureau, F.; Sutton, D.; Suur-Uski, A.S.; Sygnet, J.F.; Tauber, J.A.; Tavagnacco, D.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tristram, M.; Tucci, M.; Tuovinen, J.; Turler, M.; Valenziano, L.; Valiviita, J.; Van Tent, B.; Varis, J.; Vielva, P.; Villa, F.; Vittorio, N.; Wade, L.A.; Wandelt, B.D.; Wehus, I.K.; White, M.; White, S.D.M.; Yvon, D.; Zacchei, A.; Zonca, A.

2014-01-01

We present the Planck likelihood, a complete statistical description of the two-point correlation function of the CMB temperature fluctuations. We use this likelihood to derive the Planck CMB power spectrum over three decades in l, covering 2 = 50, we employ a correlated Gaussian likelihood approximation based on angular cross-spectra derived from the 100, 143 and 217 GHz channels. We validate our likelihood through an extensive suite of consistency tests, and assess the impact of residual foreground and instrumental uncertainties on cosmological parameters. We find good internal agreement among the high-l cross-spectra with residuals of a few uK^2 at l <= 1000. We compare our results with foreground-cleaned CMB maps, and with cross-spectra derived from the 70 GHz Planck map, and find broad agreement in terms of spectrum residuals and cosmological parameters. The best-fit LCDM cosmology is in excellent agreement with preliminary Planck polarisation spectra. The standard LCDM cosmology is well constrained b...

17. Robust Gaussian Process Regression with a Student-t Likelihood

NARCIS (Netherlands)

Jylänki, P.P.; Vanhatalo, J.; Vehtari, A.

2011-01-01

This paper considers the robust and efficient implementation of Gaussian process regression with a Student-t observation model, which has a non-log-concave likelihood. The challenge with the Student-t model is the analytically intractable inference which is why several approximative methods have

18. MAXIMUM-LIKELIHOOD-ESTIMATION OF THE ENTROPY OF AN ATTRACTOR

NARCIS (Netherlands)

SCHOUTEN, JC; TAKENS, F; VANDENBLEEK, CM

In this paper, a maximum-likelihood estimate of the (Kolmogorov) entropy of an attractor is proposed that can be obtained directly from a time series. Also, the relative standard deviation of the entropy estimate is derived; it is dependent on the entropy and on the number of samples used in the

19. Adaptive Unscented Kalman Filter using Maximum Likelihood Estimation

DEFF Research Database (Denmark)

Mahmoudi, Zeinab; Poulsen, Niels Kjølstad; Madsen, Henrik

2017-01-01

The purpose of this study is to develop an adaptive unscented Kalman filter (UKF) by tuning the measurement noise covariance. We use the maximum likelihood estimation (MLE) and the covariance matching (CM) method to estimate the noise covariance. The multi-step prediction errors generated...

20. LIKELIHOOD ESTIMATION OF PARAMETERS USING SIMULTANEOUSLY MONITORED PROCESSES

DEFF Research Database (Denmark)

Friis-Hansen, Peter; Ditlevsen, Ove Dalager

2004-01-01

The topic is maximum likelihood inference from several simultaneously monitored response processes of a structure to obtain knowledge about the parameters of other not monitored but important response processes when the structure is subject to some Gaussian load field in space and time. The consi....... The considered example is a ship sailing with a given speed through a Gaussian wave field....

1. Likelihood-based inference for clustered line transect data

DEFF Research Database (Denmark)

Waagepetersen, Rasmus; Schweder, Tore

2006-01-01

The uncertainty in estimation of spatial animal density from line transect surveys depends on the degree of spatial clustering in the animal population. To quantify the clustering we model line transect data as independent thinnings of spatial shot-noise Cox processes. Likelihood-based inference...

2. Likelihood-based Dynamic Factor Analysis for Measurement and Forecasting

NARCIS (Netherlands)

Jungbacker, B.M.J.P.; Koopman, S.J.

2015-01-01

We present new results for the likelihood-based analysis of the dynamic factor model. The latent factors are modelled by linear dynamic stochastic processes. The idiosyncratic disturbance series are specified as autoregressive processes with mutually correlated innovations. The new results lead to

3. Likelihood-based inference for clustered line transect data

DEFF Research Database (Denmark)

Waagepetersen, Rasmus Plenge; Schweder, Tore

The uncertainty in estimation of spatial animal density from line transect surveys depends on the degree of spatial clustering in the animal population. To quantify the clustering we model line transect data as independent thinnings of spatial shot-noise Cox processes. Likelihood-based inference...

4. Composite likelihood and two-stage estimation in family studies

DEFF Research Database (Denmark)

Andersen, Elisabeth Anne Wreford

2004-01-01

In this paper register based family studies provide the motivation for linking a two-stage estimation procedure in copula models for multivariate failure time data with a composite likelihood approach. The asymptotic properties of the estimators in both parametric and semi-parametric models are d...

5. Reconceptualizing Social Influence in Counseling: The Elaboration Likelihood Model.

Science.gov (United States)

McNeill, Brian W.; Stoltenberg, Cal D.

1989-01-01

Presents Elaboration Likelihood Model (ELM) of persuasion (a reconceptualization of the social influence process) as alternative model of attitude change. Contends ELM unifies conflicting social psychology results and can potentially account for inconsistent research findings in counseling psychology. Provides guidelines on integrating…

6. Counseling Pretreatment and the Elaboration Likelihood Model of Attitude Change.

Science.gov (United States)

Heesacker, Martin

1986-01-01

Results of the application of the Elaboration Likelihood Model (ELM) to a counseling context revealed that more favorable attitudes toward counseling occurred as subjects' ego involvement increased and as intervention quality improved. Counselor credibility affected the degree to which subjects' attitudes reflected argument quality differences.…

7. Multilevel maximum likelihood estimation with application to covariance matrices

Czech Academy of Sciences Publication Activity Database

Turčičová, Marie; Mandel, J.; Eben, Kryštof

Published online: 23 January ( 2018 ) ISSN 0361-0926 R&D Projects: GA ČR GA13-34856S Institutional support: RVO:67985807 Keywords : Fisher information * High dimension * Hierarchical maximum likelihood * Nested parameter spaces * Spectral diagonal covariance model * Sparse inverse covariance model Subject RIV: BB - Applied Statistics, Operational Research Impact factor: 0.311, year: 2016

8. Pendeteksian Outlier pada Regresi Nonlinier dengan Metode statistik Likelihood Displacement

Directory of Open Access Journals (Sweden)

Siti Tabi'atul Hasanah

2012-11-01

Full Text Available Outlier is an observation that much different (extreme from the other observational data, or data can be interpreted that do not follow the general pattern of the model. Sometimes outliers provide information that can not be provided by other data. That's why outliers should not just be eliminated. Outliers can also be an influential observation. There are many methods that can be used to detect of outliers. In previous studies done on outlier detection of linear regression. Next will be developed detection of outliers in nonlinear regression. Nonlinear regression here is devoted to multiplicative nonlinear regression. To detect is use of statistical method likelihood displacement. Statistical methods abbreviated likelihood displacement (LD is a method to detect outliers by removing the suspected outlier data. To estimate the parameters are used to the maximum likelihood method, so we get the estimate of the maximum. By using LD method is obtained i.e likelihood displacement is thought to contain outliers. Further accuracy of LD method in detecting the outliers are shown by comparing the MSE of LD with the MSE from the regression in general. Statistic test used is Λ. Initial hypothesis was rejected when proved so is an outlier.

9. Gaussian copula as a likelihood function for environmental models

Science.gov (United States)

Wani, O.; Espadas, G.; Cecinati, F.; Rieckermann, J.

2017-12-01

Parameter estimation of environmental models always comes with uncertainty. To formally quantify this parametric uncertainty, a likelihood function needs to be formulated, which is defined as the probability of observations given fixed values of the parameter set. A likelihood function allows us to infer parameter values from observations using Bayes' theorem. The challenge is to formulate a likelihood function that reliably describes the error generating processes which lead to the observed monitoring data, such as rainfall and runoff. If the likelihood function is not representative of the error statistics, the parameter inference will give biased parameter values. Several uncertainty estimation methods that are currently being used employ Gaussian processes as a likelihood function, because of their favourable analytical properties. Box-Cox transformation is suggested to deal with non-symmetric and heteroscedastic errors e.g. for flow data which are typically more uncertain in high flows than in periods with low flows. Problem with transformations is that the results are conditional on hyper-parameters, for which it is difficult to formulate the analyst's belief a priori. In an attempt to address this problem, in this research work we suggest learning the nature of the error distribution from the errors made by the model in the "past" forecasts. We use a Gaussian copula to generate semiparametric error distributions . 1) We show that this copula can be then used as a likelihood function to infer parameters, breaking away from the practice of using multivariate normal distributions. Based on the results from a didactical example of predicting rainfall runoff, 2) we demonstrate that the copula captures the predictive uncertainty of the model. 3) Finally, we find that the properties of autocorrelation and heteroscedasticity of errors are captured well by the copula, eliminating the need to use transforms. In summary, our findings suggest that copulas are an

10. Modeling gene expression measurement error: a quasi-likelihood approach

Directory of Open Access Journals (Sweden)

Strimmer Korbinian

2003-03-01

Full Text Available Abstract Background Using suitable error models for gene expression measurements is essential in the statistical analysis of microarray data. However, the true probabilistic model underlying gene expression intensity readings is generally not known. Instead, in currently used approaches some simple parametric model is assumed (usually a transformed normal distribution or the empirical distribution is estimated. However, both these strategies may not be optimal for gene expression data, as the non-parametric approach ignores known structural information whereas the fully parametric models run the risk of misspecification. A further related problem is the choice of a suitable scale for the model (e.g. observed vs. log-scale. Results Here a simple semi-parametric model for gene expression measurement error is presented. In this approach inference is based an approximate likelihood function (the extended quasi-likelihood. Only partial knowledge about the unknown true distribution is required to construct this function. In case of gene expression this information is available in the form of the postulated (e.g. quadratic variance structure of the data. As the quasi-likelihood behaves (almost like a proper likelihood, it allows for the estimation of calibration and variance parameters, and it is also straightforward to obtain corresponding approximate confidence intervals. Unlike most other frameworks, it also allows analysis on any preferred scale, i.e. both on the original linear scale as well as on a transformed scale. It can also be employed in regression approaches to model systematic (e.g. array or dye effects. Conclusions The quasi-likelihood framework provides a simple and versatile approach to analyze gene expression data that does not make any strong distributional assumptions about the underlying error model. For several simulated as well as real data sets it provides a better fit to the data than competing models. In an example it also

11. On the performance of social network and likelihood-based expert weighting schemes

International Nuclear Information System (INIS)

Cooke, Roger M.; ElSaadany, Susie; Huang Xinzheng

2008-01-01

Using expert judgment data from the TU Delft's expert judgment database, we compare the performance of different weighting schemes, namely equal weighting, performance-based weighting from the classical model [Cooke RM. Experts in uncertainty. Oxford: Oxford University Press; 1991.], social network (SN) weighting and likelihood weighting. The picture that emerges with regard to SN weights is rather mixed. SN theory does not provide an alternative to performance-based combination of expert judgments, since the statistical accuracy of the SN decision maker is sometimes unacceptably low. On the other hand, it does outperform equal weighting in the majority of cases. The results here, though not overwhelmingly positive, do nonetheless motivate further research into social interaction methods for nominating and weighting experts. Indeed, a full expert judgment study with performance measurement requires an investment in time and effort, with a view to securing external validation. If high confidence in a comparable level of validation can be obtained by less intensive methods, this would be very welcome, and would facilitate the application of structured expert judgment in situations where the resources for a full study are not available. Likelihood weights are just as resource intensive as performance-based weights, and the evidence presented here suggests that they are inferior to performance-based weights with regard to those scoring variables which are optimized in performance weights (calibration and information). Perhaps surprisingly, they are also inferior with regard to likelihood. Their use is further discouraged by the fact that they constitute a strongly improper scoring rule

12. MAXIMUM LIKELIHOOD CLASSIFICATION OF HIGH-RESOLUTION SAR IMAGES IN URBAN AREA

Directory of Open Access Journals (Sweden)

M. Soheili Majd

2012-09-01

Full Text Available In this work, we propose a state-of-the-art on statistical analysis of polarimetric synthetic aperture radar (SAR data, through the modeling of several indices. We concentrate on eight ground classes which have been carried out from amplitudes, co-polarisation ratio, depolarization ratios, and other polarimetric descriptors. To study their different statistical behaviours, we consider Gauss, log- normal, Beta I, Weibull, Gamma, and Fisher statistical models and estimate their parameters using three methods: method of moments (MoM, maximum-likelihood (ML methodology, and log-cumulants method (MoML. Then, we study the opportunity of introducing this information in an adapted supervised classification scheme based on Maximum–Likelihood and Fisher pdf. Our work relies on an image of a suburban area, acquired by the airborne RAMSES SAR sensor of ONERA. The results prove the potential of such data to discriminate urban surfaces and show the usefulness of adapting any classical classification algorithm however classification maps present a persistant class confusion between flat gravelled or concrete roofs and trees.

13. A composite likelihood approach for spatially correlated survival data

Science.gov (United States)

Paik, Jane; Ying, Zhiliang

2013-01-01

The aim of this paper is to provide a composite likelihood approach to handle spatially correlated survival data using pairwise joint distributions. With e-commerce data, a recent question of interest in marketing research has been to describe spatially clustered purchasing behavior and to assess whether geographic distance is the appropriate metric to describe purchasing dependence. We present a model for the dependence structure of time-to-event data subject to spatial dependence to characterize purchasing behavior from the motivating example from e-commerce data. We assume the Farlie-Gumbel-Morgenstern (FGM) distribution and then model the dependence parameter as a function of geographic and demographic pairwise distances. For estimation of the dependence parameters, we present pairwise composite likelihood equations. We prove that the resulting estimators exhibit key properties of consistency and asymptotic normality under certain regularity conditions in the increasing-domain framework of spatial asymptotic theory. PMID:24223450

14. A composite likelihood approach for spatially correlated survival data.

Science.gov (United States)

Paik, Jane; Ying, Zhiliang

2013-01-01

The aim of this paper is to provide a composite likelihood approach to handle spatially correlated survival data using pairwise joint distributions. With e-commerce data, a recent question of interest in marketing research has been to describe spatially clustered purchasing behavior and to assess whether geographic distance is the appropriate metric to describe purchasing dependence. We present a model for the dependence structure of time-to-event data subject to spatial dependence to characterize purchasing behavior from the motivating example from e-commerce data. We assume the Farlie-Gumbel-Morgenstern (FGM) distribution and then model the dependence parameter as a function of geographic and demographic pairwise distances. For estimation of the dependence parameters, we present pairwise composite likelihood equations. We prove that the resulting estimators exhibit key properties of consistency and asymptotic normality under certain regularity conditions in the increasing-domain framework of spatial asymptotic theory.

15. Secondary Analysis under Cohort Sampling Designs Using Conditional Likelihood

Directory of Open Access Journals (Sweden)

Olli Saarela

2012-01-01

Full Text Available Under cohort sampling designs, additional covariate data are collected on cases of a specific type and a randomly selected subset of noncases, primarily for the purpose of studying associations with a time-to-event response of interest. With such data available, an interest may arise to reuse them for studying associations between the additional covariate data and a secondary non-time-to-event response variable, usually collected for the whole study cohort at the outset of the study. Following earlier literature, we refer to such a situation as secondary analysis. We outline a general conditional likelihood approach for secondary analysis under cohort sampling designs and discuss the specific situations of case-cohort and nested case-control designs. We also review alternative methods based on full likelihood and inverse probability weighting. We compare the alternative methods for secondary analysis in two simulated settings and apply them in a real-data example.

16. GENERALIZATION OF RAYLEIGH MAXIMUM LIKELIHOOD DESPECKLING FILTER USING QUADRILATERAL KERNELS

Directory of Open Access Journals (Sweden)

S. Sridevi

2013-02-01

Full Text Available Speckle noise is the most prevalent noise in clinical ultrasound images. It visibly looks like light and dark spots and deduce the pixel intensity as murkiest. Gazing at fetal ultrasound images, the impact of edge and local fine details are more palpable for obstetricians and gynecologists to carry out prenatal diagnosis of congenital heart disease. A robust despeckling filter has to be contrived to proficiently suppress speckle noise and simultaneously preserve the features. The proposed filter is the generalization of Rayleigh maximum likelihood filter by the exploitation of statistical tools as tuning parameters and use different shapes of quadrilateral kernels to estimate the noise free pixel from neighborhood. The performance of various filters namely Median, Kuwahura, Frost, Homogenous mask filter and Rayleigh maximum likelihood filter are compared with the proposed filter in terms PSNR and image profile. Comparatively the proposed filters surpass the conventional filters.

17. Likelihood inference for a nonstationary fractional autoregressive model

DEFF Research Database (Denmark)

Johansen, Søren; Ørregård Nielsen, Morten

2010-01-01

This paper discusses model-based inference in an autoregressive model for fractional processes which allows the process to be fractional of order d or d-b. Fractional differencing involves infinitely many past values and because we are interested in nonstationary processes we model the data X1......,...,X_{T} given the initial values X_{-n}, n=0,1,..., as is usually done. The initial values are not modeled but assumed to be bounded. This represents a considerable generalization relative to all previous work where it is assumed that initial values are zero. For the statistical analysis we assume...... the conditional Gaussian likelihood and for the probability analysis we also condition on initial values but assume that the errors in the autoregressive model are i.i.d. with suitable moment conditions. We analyze the conditional likelihood and its derivatives as stochastic processes in the parameters, including...

18. Physical constraints on the likelihood of life on exoplanets

Science.gov (United States)

Lingam, Manasvi; Loeb, Abraham

2018-04-01

One of the most fundamental questions in exoplanetology is to determine whether a given planet is habitable. We estimate the relative likelihood of a planet's propensity towards habitability by considering key physical characteristics such as the role of temperature on ecological and evolutionary processes, and atmospheric losses via hydrodynamic escape and stellar wind erosion. From our analysis, we demonstrate that Earth-sized exoplanets in the habitable zone around M-dwarfs seemingly display much lower prospects of being habitable relative to Earth, owing to the higher incident ultraviolet fluxes and closer distances to the host star. We illustrate our results by specifically computing the likelihood (of supporting life) for the recently discovered exoplanets, Proxima b and TRAPPIST-1e, which we find to be several orders of magnitude smaller than that of Earth.

19. THESEUS: maximum likelihood superpositioning and analysis of macromolecular structures.

Science.gov (United States)

Theobald, Douglas L; Wuttke, Deborah S

2006-09-01

THESEUS is a command line program for performing maximum likelihood (ML) superpositions and analysis of macromolecular structures. While conventional superpositioning methods use ordinary least-squares (LS) as the optimization criterion, ML superpositions provide substantially improved accuracy by down-weighting variable structural regions and by correcting for correlations among atoms. ML superpositioning is robust and insensitive to the specific atoms included in the analysis, and thus it does not require subjective pruning of selected variable atomic coordinates. Output includes both likelihood-based and frequentist statistics for accurate evaluation of the adequacy of a superposition and for reliable analysis of structural similarities and differences. THESEUS performs principal components analysis for analyzing the complex correlations found among atoms within a structural ensemble. ANSI C source code and selected binaries for various computing platforms are available under the GNU open source license from http://monkshood.colorado.edu/theseus/ or http://www.theseus3d.org.

20. Deformation of log-likelihood loss function for multiclass boosting.

Science.gov (United States)

Kanamori, Takafumi

2010-09-01

The purpose of this paper is to study loss functions in multiclass classification. In classification problems, the decision function is estimated by minimizing an empirical loss function, and then, the output label is predicted by using the estimated decision function. We propose a class of loss functions which is obtained by a deformation of the log-likelihood loss function. There are four main reasons why we focus on the deformed log-likelihood loss function: (1) this is a class of loss functions which has not been deeply investigated so far, (2) in terms of computation, a boosting algorithm with a pseudo-loss is available to minimize the proposed loss function, (3) the proposed loss functions provide a clear correspondence between the decision functions and conditional probabilities of output labels, (4) the proposed loss functions satisfy the statistical consistency of the classification error rate which is a desirable property in classification problems. Based on (3), we show that the deformed log-likelihood loss provides a model of mislabeling which is useful as a statistical model of medical diagnostics. We also propose a robust loss function against outliers in multiclass classification based on our approach. The robust loss function is a natural extension of the existing robust loss function for binary classification. A model of mislabeling and a robust loss function are useful to cope with noisy data. Some numerical studies are presented to show the robustness of the proposed loss function. A mathematical characterization of the deformed log-likelihood loss function is also presented. Copyright 2010 Elsevier Ltd. All rights reserved.

1. Bayesian interpretation of Generalized empirical likelihood by maximum entropy

OpenAIRE

Rochet , Paul

2011-01-01

We study a parametric estimation problem related to moment condition models. As an alternative to the generalized empirical likelihood (GEL) and the generalized method of moments (GMM), a Bayesian approach to the problem can be adopted, extending the MEM procedure to parametric moment conditions. We show in particular that a large number of GEL estimators can be interpreted as a maximum entropy solution. Moreover, we provide a more general field of applications by proving the method to be rob...

2. Menyoal Elaboration Likelihood Model (ELM) dan Teori Retorika

OpenAIRE

Yudi Perbawaningsih

2012-01-01

Abstract: Persuasion is a communication process to establish or change attitudes, which can be understood through theory of Rhetoric and theory of Elaboration Likelihood Model (ELM). This study elaborates these theories in a Public Lecture series which to persuade the students in choosing their concentration of study. The result shows that in term of persuasion effectiveness it is not quite relevant to separate the message and its source. The quality of source is determined by the quality of ...

3. Corporate governance effect on financial distress likelihood: Evidence from Spain

Directory of Open Access Journals (Sweden)

Montserrat Manzaneque

2016-01-01

Full Text Available The paper explores some mechanisms of corporate governance (ownership and board characteristics in Spanish listed companies and their impact on the likelihood of financial distress. An empirical study was conducted between 2007 and 2012 using a matched-pairs research design with 308 observations, with half of them classified as distressed and non-distressed. Based on the previous study by Pindado, Rodrigues, and De la Torre (2008, a broader concept of bankruptcy is used to define business failure. Employing several conditional logistic models, as well as to other previous studies on bankruptcy, the results confirm that in difficult situations prior to bankruptcy, the impact of board ownership and proportion of independent directors on business failure likelihood are similar to those exerted in more extreme situations. These results go one step further, to offer a negative relationship between board size and the likelihood of financial distress. This result is interpreted as a form of creating diversity and to improve the access to the information and resources, especially in contexts where the ownership is highly concentrated and large shareholders have a great power to influence the board structure. However, the results confirm that ownership concentration does not have a significant impact on financial distress likelihood in the Spanish context. It is argued that large shareholders are passive as regards an enhanced monitoring of management and, alternatively, they do not have enough incentives to hold back the financial distress. These findings have important implications in the Spanish context, where several changes in the regulatory listing requirements have been carried out with respect to corporate governance, and where there is no empirical evidence regarding this respect.

4. Maximum Likelihood, Consistency and Data Envelopment Analysis: A Statistical Foundation

OpenAIRE

Rajiv D. Banker

1993-01-01

This paper provides a formal statistical basis for the efficiency evaluation techniques of data envelopment analysis (DEA). DEA estimators of the best practice monotone increasing and concave production function are shown to be also maximum likelihood estimators if the deviation of actual output from the efficient output is regarded as a stochastic variable with a monotone decreasing probability density function. While the best practice frontier estimator is biased below the theoretical front...

5. Maximum likelihood convolutional decoding (MCD) performance due to system losses

Science.gov (United States)

Webster, L.

1976-01-01

A model for predicting the computational performance of a maximum likelihood convolutional decoder (MCD) operating in a noisy carrier reference environment is described. This model is used to develop a subroutine that will be utilized by the Telemetry Analysis Program to compute the MCD bit error rate. When this computational model is averaged over noisy reference phase errors using a high-rate interpolation scheme, the results are found to agree quite favorably with experimental measurements.

6. Menyoal Elaboration Likelihood Model (ELM) Dan Teori Retorika

OpenAIRE

Perbawaningsih, Yudi

2012-01-01

: Persuasion is a communication process to establish or change attitudes, which can be understood through theory of Rhetoric and theory of Elaboration Likelihood Model (ELM). This study elaborates these theories in a Public Lecture series which to persuade the students in choosing their concentration of study. The result shows that in term of persuasion effectiveness it is not quite relevant to separate the message and its source. The quality of source is determined by the quality of the mess...

7. Penggunaan Elaboration Likelihood Model dalam Menganalisis Penerimaan Teknologi Informasi

OpenAIRE

vitrian, vitrian2

2010-01-01

This article discusses some technology acceptance models in an organization. Thorough analysis of how technology is acceptable help managers make any planning to implement new teachnology and make sure that new technology could enhance organization's performance. Elaboration Likelihood Model (ELM) is the one which sheds light on some behavioral factors in acceptance of information technology. The basic tenet of ELM states that human behavior in principle can be influenced through central r...

8. Statistical Bias in Maximum Likelihood Estimators of Item Parameters.

Science.gov (United States)

1982-04-01

34 a> E r’r~e r ,C Ie I# ne,..,.rVi rnd Id.,flfv b1 - bindk numb.r) I; ,t-i i-cd I ’ tiie bias in the maximum likelihood ,st i- i;, ’ t iIeiIrs in...NTC, IL 60088 Psychometric Laboratory University of North Carolina I ERIC Facility-Acquisitions Davie Hall 013A 4833 Rugby Avenue Chapel Hill, NC

9. Empirical Likelihood in Nonignorable Covariate-Missing Data Problems.

Science.gov (United States)

Xie, Yanmei; Zhang, Biao

2017-04-20

Missing covariate data occurs often in regression analysis, which frequently arises in the health and social sciences as well as in survey sampling. We study methods for the analysis of a nonignorable covariate-missing data problem in an assumed conditional mean function when some covariates are completely observed but other covariates are missing for some subjects. We adopt the semiparametric perspective of Bartlett et al. (Improving upon the efficiency of complete case analysis when covariates are MNAR. Biostatistics 2014;15:719-30) on regression analyses with nonignorable missing covariates, in which they have introduced the use of two working models, the working probability model of missingness and the working conditional score model. In this paper, we study an empirical likelihood approach to nonignorable covariate-missing data problems with the objective of effectively utilizing the two working models in the analysis of covariate-missing data. We propose a unified approach to constructing a system of unbiased estimating equations, where there are more equations than unknown parameters of interest. One useful feature of these unbiased estimating equations is that they naturally incorporate the incomplete data into the data analysis, making it possible to seek efficient estimation of the parameter of interest even when the working regression function is not specified to be the optimal regression function. We apply the general methodology of empirical likelihood to optimally combine these unbiased estimating equations. We propose three maximum empirical likelihood estimators of the underlying regression parameters and compare their efficiencies with other existing competitors. We present a simulation study to compare the finite-sample performance of various methods with respect to bias, efficiency, and robustness to model misspecification. The proposed empirical likelihood method is also illustrated by an analysis of a data set from the US National Health and

10. Democracy, Autocracy and the Likelihood of International Conflict

OpenAIRE

Tangerås, Thomas

2008-01-01

This is a game-theoretic analysis of the link between regime type and international conflict. The democratic electorate can credibly punish the leader for bad conflict outcomes, whereas the autocratic selectorate cannot. For the fear of being thrown out of office, democratic leaders are (i) more selective about the wars they initiate and (ii) on average win more of the wars they start. Foreign policy behaviour is found to display strategic complementarities. The likelihood of interstate war, ...

11. Moment Conditions Selection Based on Adaptive Penalized Empirical Likelihood

Directory of Open Access Journals (Sweden)

Yunquan Song

2014-01-01

Full Text Available Empirical likelihood is a very popular method and has been widely used in the fields of artificial intelligence (AI and data mining as tablets and mobile application and social media dominate the technology landscape. This paper proposes an empirical likelihood shrinkage method to efficiently estimate unknown parameters and select correct moment conditions simultaneously, when the model is defined by moment restrictions in which some are possibly misspecified. We show that our method enjoys oracle-like properties; that is, it consistently selects the correct moment conditions and at the same time its estimator is as efficient as the empirical likelihood estimator obtained by all correct moment conditions. Moreover, unlike the GMM, our proposed method allows us to carry out confidence regions for the parameters included in the model without estimating the covariances of the estimators. For empirical implementation, we provide some data-driven procedures for selecting the tuning parameter of the penalty function. The simulation results show that the method works remarkably well in terms of correct moment selection and the finite sample properties of the estimators. Also, a real-life example is carried out to illustrate the new methodology.

12. Caching and interpolated likelihoods: accelerating cosmological Monte Carlo Markov chains

Energy Technology Data Exchange (ETDEWEB)

Bouland, Adam; Easther, Richard; Rosenfeld, Katherine, E-mail: adam.bouland@aya.yale.edu, E-mail: richard.easther@yale.edu, E-mail: krosenfeld@cfa.harvard.edu [Department of Physics, Yale University, New Haven CT 06520 (United States)

2011-05-01

We describe a novel approach to accelerating Monte Carlo Markov Chains. Our focus is cosmological parameter estimation, but the algorithm is applicable to any problem for which the likelihood surface is a smooth function of the free parameters and computationally expensive to evaluate. We generate a high-order interpolating polynomial for the log-likelihood using the first points gathered by the Markov chains as a training set. This polynomial then accurately computes the majority of the likelihoods needed in the latter parts of the chains. We implement a simple version of this algorithm as a patch (InterpMC) to CosmoMC and show that it accelerates parameter estimatation by a factor of between two and four for well-converged chains. The current code is primarily intended as a ''proof of concept'', and we argue that there is considerable room for further performance gains. Unlike other approaches to accelerating parameter fits, we make no use of precomputed training sets or special choices of variables, and InterpMC is almost entirely transparent to the user.

13. Caching and interpolated likelihoods: accelerating cosmological Monte Carlo Markov chains

International Nuclear Information System (INIS)

Bouland, Adam; Easther, Richard; Rosenfeld, Katherine

2011-01-01

We describe a novel approach to accelerating Monte Carlo Markov Chains. Our focus is cosmological parameter estimation, but the algorithm is applicable to any problem for which the likelihood surface is a smooth function of the free parameters and computationally expensive to evaluate. We generate a high-order interpolating polynomial for the log-likelihood using the first points gathered by the Markov chains as a training set. This polynomial then accurately computes the majority of the likelihoods needed in the latter parts of the chains. We implement a simple version of this algorithm as a patch (InterpMC) to CosmoMC and show that it accelerates parameter estimatation by a factor of between two and four for well-converged chains. The current code is primarily intended as a ''proof of concept'', and we argue that there is considerable room for further performance gains. Unlike other approaches to accelerating parameter fits, we make no use of precomputed training sets or special choices of variables, and InterpMC is almost entirely transparent to the user

14. Maximum likelihood as a common computational framework in tomotherapy

International Nuclear Information System (INIS)

Olivera, G.H.; Shepard, D.M.; Reckwerdt, P.J.; Ruchala, K.; Zachman, J.; Fitchard, E.E.; Mackie, T.R.

1998-01-01

Tomotherapy is a dose delivery technique using helical or axial intensity modulated beams. One of the strengths of the tomotherapy concept is that it can incorporate a number of processes into a single piece of equipment. These processes include treatment optimization planning, dose reconstruction and kilovoltage/megavoltage image reconstruction. A common computational technique that could be used for all of these processes would be very appealing. The maximum likelihood estimator, originally developed for emission tomography, can serve as a useful tool in imaging and radiotherapy. We believe that this approach can play an important role in the processes of optimization planning, dose reconstruction and kilovoltage and/or megavoltage image reconstruction. These processes involve computations that require comparable physical methods. They are also based on equivalent assumptions, and they have similar mathematical solutions. As a result, the maximum likelihood approach is able to provide a common framework for all three of these computational problems. We will demonstrate how maximum likelihood methods can be applied to optimization planning, dose reconstruction and megavoltage image reconstruction in tomotherapy. Results for planning optimization, dose reconstruction and megavoltage image reconstruction will be presented. Strengths and weaknesses of the methodology are analysed. Future directions for this work are also suggested. (author)

15. Neural Networks Involved in Adolescent Reward Processing: An Activation Likelihood Estimation Meta-Analysis of Functional Neuroimaging Studies

Science.gov (United States)

Silverman, Merav H.; Jedd, Kelly; Luciana, Monica

2015-01-01

16. Diagnostic accuracy of liver fibrosis based on red cell distribution width (RDW) to platelet ratio with fibroscan in chronic hepatitis B

Science.gov (United States)

Sembiring, J.; Jones, F.

2018-03-01

Red cell Distribution Width (RDW) and platelet ratio (RPR) can predict liver fibrosis and cirrhosis in chronic hepatitis B with relatively high accuracy. RPR was superior to other non-invasive methods to predict liver fibrosis, such as AST and ALT ratio, AST and platelet ratio Index and FIB-4. The aim of this study was to assess diagnostic accuracy liver fibrosis by using RDW and platelets ratio in chronic hepatitis B patients based on compared with Fibroscan. This cross-sectional study was conducted at Adam Malik Hospital from January-June 2015. We examine 34 patients hepatitis B chronic, screen RDW, platelet, and fibroscan. Data were statistically analyzed. The result RPR with ROC procedure has an accuracy of 72.3% (95% CI: 84.1% - 97%). In this study, the RPR had a moderate ability to predict fibrosis degree (p = 0.029 with AUC> 70%). The cutoff value RPR was 0.0591, sensitivity and spesificity were 71.4% and 60%, Positive Prediction Value (PPV) was 55.6% and Negative Predictions Value (NPV) was 75%, positive likelihood ratio was 1.79 and negative likelihood ratio was 0.48. RPR have the ability to predict the degree of liver fibrosis in chronic hepatitis B patients with moderate accuracy.

17. Communicating likelihoods and probabilities in forecasts of volcanic eruptions

Science.gov (United States)

Doyle, Emma E. H.; McClure, John; Johnston, David M.; Paton, Douglas

2014-02-01

The issuing of forecasts and warnings of natural hazard events, such as volcanic eruptions, earthquake aftershock sequences and extreme weather often involves the use of probabilistic terms, particularly when communicated by scientific advisory groups to key decision-makers, who can differ greatly in relative expertise and function in the decision making process. Recipients may also differ in their perception of relative importance of political and economic influences on interpretation. Consequently, the interpretation of these probabilistic terms can vary greatly due to the framing of the statements, and whether verbal or numerical terms are used. We present a review from the psychology literature on how the framing of information influences communication of these probability terms. It is also unclear as to how people rate their perception of an event's likelihood throughout a time frame when a forecast time window is stated. Previous research has identified that, when presented with a 10-year time window forecast, participants viewed the likelihood of an event occurring ‘today’ as being of less than that in year 10. Here we show that this skew in perception also occurs for short-term time windows (under one week) that are of most relevance for emergency warnings. In addition, unlike the long-time window statements, the use of the phrasing “within the next…” instead of “in the next…” does not mitigate this skew, nor do we observe significant differences between the perceived likelihoods of scientists and non-scientists. This finding suggests that effects occurring due to the shorter time window may be ‘masking’ any differences in perception due to wording or career background observed for long-time window forecasts. These results have implications for scientific advice, warning forecasts, emergency management decision-making, and public information as any skew in perceived event likelihood towards the end of a forecast time window may result in

18. Comparisons of likelihood and machine learning methods of individual classification

Science.gov (United States)

Guinand, B.; Topchy, A.; Page, K.S.; Burnham-Curtis, M. K.; Punch, W.F.; Scribner, K.T.

2002-01-01

Classification methods used in machine learning (e.g., artificial neural networks, decision trees, and k-nearest neighbor clustering) are rarely used with population genetic data. We compare different nonparametric machine learning techniques with parametric likelihood estimations commonly employed in population genetics for purposes of assigning individuals to their population of origin (“assignment tests”). Classifier accuracy was compared across simulated data sets representing different levels of population differentiation (low and high FST), number of loci surveyed (5 and 10), and allelic diversity (average of three or eight alleles per locus). Empirical data for the lake trout (Salvelinus namaycush) exhibiting levels of population differentiation comparable to those used in simulations were examined to further evaluate and compare classification methods. Classification error rates associated with artificial neural networks and likelihood estimators were lower for simulated data sets compared to k-nearest neighbor and decision tree classifiers over the entire range of parameters considered. Artificial neural networks only marginally outperformed the likelihood method for simulated data (0–2.8% lower error rates). The relative performance of each machine learning classifier improved relative likelihood estimators for empirical data sets, suggesting an ability to “learn” and utilize properties of empirical genotypic arrays intrinsic to each population. Likelihood-based estimation methods provide a more accessible option for reliable assignment of individuals to the population of origin due to the intricacies in development and evaluation of artificial neural networks. In recent years, characterization of highly polymorphic molecular markers such as mini- and microsatellites and development of novel methods of analysis have enabled researchers to extend investigations of ecological and evolutionary processes below the population level to the level of

19. High ratio recirculating gas compressor

Science.gov (United States)

Weinbrecht, J.F.

1989-08-22

A high ratio positive displacement recirculating rotary compressor is disclosed. The compressor includes an integral heat exchanger and recirculation conduits for returning cooled, high pressure discharge gas to the compressor housing to reducing heating of the compressor and enable higher pressure ratios to be sustained. The compressor features a recirculation system which results in continuous and uninterrupted flow of recirculation gas to the compressor with no direct leakage to either the discharge port or the intake port of the compressor, resulting in a capability of higher sustained pressure ratios without overheating of the compressor. 10 figs.

20. The impact of food intake and social habits on embryo quality and the likelihood of blastocyst formation.

Science.gov (United States)

Braga, Daniela Paes Almeida Ferreira; Halpern, Gabriela; Setti, Amanda S; Figueira, Rita Cássia S; Iaconelli, Assumpto; Borges, Edson

2015-07-01

1. US adolescents and MyPyramid: associations between fast-food consumption and lower likelihood of meeting recommendations.

Science.gov (United States)

Sebastian, Rhonda S; Wilkinson Enns, Cecilia; Goldman, Joseph D

2009-02-01

To determine whether fast-food consumption is associated with adolescents' food group intakes and likelihood of meeting recommendations outlined in the MyPyramid Food Guidance System. Data from two 24-hour recalls collected in What We Eat in America, National Health and Nutrition Examination Survey 2003-2004 were analyzed. Fast-food consumers were divided into tertiles based on the proportion of 2-day energy intake derived from fast food. Adolescent boys and nonpregnant girls aged 12 to 19 years (n=1,956). All statistical analyses included sample weights to account for the survey design. Regression analyses were used to detect associations between fast-food consumption and both food group intakes and percentages of individuals meeting MyPyramid recommendations, and to predict odds of meeting recommendations by fast-food consumption level. Fast-food consumption was associated negatively with MyPyramid fruit and milk group intakes (boys and girls) and positively with discretionary energy and solid fats (girls only). Negative associations were also found between fast-food consumption and percentages of adolescents meeting recommendations for milk (boys), fruits (girls), and vegetables and discretionary energy (boys and girls). Compared with those consuming no fast food, adolescents in the highest tertile of energy from fast food were less likely to meet recommendations for vegetables (odds ratio [OR]=0.16, 95% confidence interval [CI]: 0.05 to 0.52 for boys; OR=0.18, 95% CI: 0.04 to 0.79 for girls) and discretionary energy (OR=0.41, 95% CI: 0.22 to 0.77 for boys; OR=0.04, 95% CI: 0.01 to 0.24 for girls). No relationships were found between fast-food consumption and grains, meat/beans, and oils. Adolescents' intakes, whether containing fast food or not, need improvement. Fast food is one factor that impacts adolescents' intake of MyPyramid groups and their likelihood of meeting recommendations. Awareness of fast-food's role in discrepancies between adolescent intakes

2. Likelihood of being seen within emergency departments’ assigned urgency times for poisoned and injured individuals

Directory of Open Access Journals (Sweden)

Rachel L. Rosenthal

2014-10-01

Full Text Available The objective of the present study is to determine the likelihood of injured or poisoned patients in special populations, such as those patients that are elderly and self-injurious, being seen within an emergency department’s triage nurse assigned urgency. Data from the National Hospital Ambulatory Medical Care Survey (2007 was utilized in this study. Multi-level models and multivariate linear regression models were used; patient age, sex, reported pain levels, wait time, and injury type were examined as potential predictors of being seen within assigned urgency. From a random sample across all US Emergency Departments, 5616 patients nested in 312 hospital emergency departments were included into the study. Typically, approximately 1 in 5 emergency department patients were not seen within their triage nurse assigned urgencies. The typical patient in the average hospital had an 81% likelihood of being seen within their assigned urgency. P atients who were oldest [odds ratio (OR=0.0990] and had self-inflicted injuries (vs assault OR=1.246 and OR=1.596 had the least likelihood to be seen within their assigned urgencies. As actual wait-time increased for patients, they were less likely to be seen within their assigned urgencies. The most powerful predictors of the study’s outcome were injury type and age, indicating that patients from special populations such as the elderly or those with injuries resulting from deliberate self-harm are less likely to be actually priority patients independent of triage nurse assigned urgencies.

3. Applying exclusion likelihoods from LHC searches to extended Higgs sectors

International Nuclear Information System (INIS)

Bechtle, Philip; Heinemeyer, Sven; Staal, Oscar; Stefaniak, Tim; Weiglein, Georg

2015-01-01

LHC searches for non-standard Higgs bosons decaying into tau lepton pairs constitute a sensitive experimental probe for physics beyond the Standard Model (BSM), such as supersymmetry (SUSY). Recently, the limits obtained from these searches have been presented by the CMS collaboration in a nearly model-independent fashion - as a narrow resonance model - based on the full 8 TeV dataset. In addition to publishing a 95 % C.L. exclusion limit, the full likelihood information for the narrowresonance model has been released. This provides valuable information that can be incorporated into global BSM fits. We present a simple algorithm that maps an arbitrary model with multiple neutral Higgs bosons onto the narrow resonance model and derives the corresponding value for the exclusion likelihood from the CMS search. This procedure has been implemented into the public computer code HiggsBounds (version 4.2.0 and higher). We validate our implementation by cross-checking against the official CMS exclusion contours in three Higgs benchmark scenarios in the Minimal Supersymmetric Standard Model (MSSM), and find very good agreement. Going beyond validation, we discuss the combined constraints of the ττ search and the rate measurements of the SM-like Higgs at 125 GeV in a recently proposed MSSM benchmark scenario, where the lightest Higgs boson obtains SM-like couplings independently of the decoupling of the heavier Higgs states. Technical details for how to access the likelihood information within HiggsBounds are given in the appendix. The program is available at http:// higgsbounds.hepforge.org. (orig.)

4. Australian food life style segments and elaboration likelihood differences

DEFF Research Database (Denmark)

Brunsø, Karen; Reid, Mike

As the global food marketing environment becomes more competitive, the international and comparative perspective of consumers' attitudes and behaviours becomes more important for both practitioners and academics. This research employs the Food-Related Life Style (FRL) instrument in Australia...... in order to 1) determine Australian Life Style Segments and compare these with their European counterparts, and to 2) explore differences in elaboration likelihood among the Australian segments, e.g. consumers' interest and motivation to perceive product related communication. The results provide new...

5. Maximum-likelihood method for numerical inversion of Mellin transform

International Nuclear Information System (INIS)

Iqbal, M.

1997-01-01

A method is described for inverting the Mellin transform which uses an expansion in Laguerre polynomials and converts the Mellin transform to Laplace transform, then the maximum-likelihood regularization method is used to recover the original function of the Mellin transform. The performance of the method is illustrated by the inversion of the test functions available in the literature (J. Inst. Math. Appl., 20 (1977) 73; Math. Comput., 53 (1989) 589). Effectiveness of the method is shown by results obtained through demonstration by means of tables and diagrams

6. How to Improve the Likelihood of CDM Approval?

DEFF Research Database (Denmark)

Brandt, Urs Steiner; Svendsen, Gert Tinggaard

2014-01-01

How can the likelihood of Clean Development Mechanism (CDM) approval be improved in the face of institutional shortcomings? To answer this question, we focus on the three institutional shortcomings of income sharing, risk sharing and corruption prevention concerning afforestation/reforestation (A....../R). Furthermore, three main stakeholders are identified, namely investors, governments and agents in a principal-agent model regarding monitoring and enforcement capacity. Developing countries such as West Africa have, despite huge potentials, not been integrated in A/R CDM projects yet. Remote sensing, however...

7. Maximum Likelihood and Bayes Estimation in Randomly Censored Geometric Distribution

Directory of Open Access Journals (Sweden)

Hare Krishna

2017-01-01

Full Text Available In this article, we study the geometric distribution under randomly censored data. Maximum likelihood estimators and confidence intervals based on Fisher information matrix are derived for the unknown parameters with randomly censored data. Bayes estimators are also developed using beta priors under generalized entropy and LINEX loss functions. Also, Bayesian credible and highest posterior density (HPD credible intervals are obtained for the parameters. Expected time on test and reliability characteristics are also analyzed in this article. To compare various estimates developed in the article, a Monte Carlo simulation study is carried out. Finally, for illustration purpose, a randomly censored real data set is discussed.

8. Elemental composition of cosmic rays using a maximum likelihood method

International Nuclear Information System (INIS)

Ruddick, K.

1996-01-01

We present a progress report on our attempts to determine the composition of cosmic rays in the knee region of the energy spectrum. We have used three different devices to measure properties of the extensive air showers produced by primary cosmic rays: the Soudan 2 underground detector measures the muon flux deep underground, a proportional tube array samples shower density at the surface of the earth, and a Cherenkov array observes light produced high in the atmosphere. We have begun maximum likelihood fits to these measurements with the hope of determining the nuclear mass number A on an event by event basis. (orig.)

9. Likelihood-Based Inference in Nonlinear Error-Correction Models

DEFF Research Database (Denmark)

Kristensen, Dennis; Rahbæk, Anders

We consider a class of vector nonlinear error correction models where the transfer function (or loadings) of the stationary relation- ships is nonlinear. This includes in particular the smooth transition models. A general representation theorem is given which establishes the dynamic properties...... and a linear trend in general. Gaussian likelihood-based estimators are considered for the long- run cointegration parameters, and the short-run parameters. Asymp- totic theory is provided for these and it is discussed to what extend asymptotic normality and mixed normaity can be found. A simulation study...

10. Process criticality accident likelihoods, consequences and emergency planning

International Nuclear Information System (INIS)

McLaughlin, T.P.

1992-01-01

Evaluation of criticality accident risks in the processing of significant quantities of fissile materials is both complex and subjective, largely due to the lack of accident statistics. Thus, complying with national and international standards and regulations which require an evaluation of the net benefit of a criticality accident alarm system, is also subjective. A review of guidance found in the literature on potential accident magnitudes is presented for different material forms and arrangements. Reasoned arguments are also presented concerning accident prevention and accident likelihoods for these material forms and arrangements. (Author)

11. Likelihood Estimation of Gamma Ray Bursts Duration Distribution

OpenAIRE

Horvath, Istvan

2005-01-01

Two classes of Gamma Ray Bursts have been identified so far, characterized by T90 durations shorter and longer than approximately 2 seconds. It was shown that the BATSE 3B data allow a good fit with three Gaussian distributions in log T90. In the same Volume in ApJ. another paper suggested that the third class of GRBs is may exist. Using the full BATSE catalog here we present the maximum likelihood estimation, which gives us 0.5% probability to having only two subclasses. The MC simulation co...

12. Process criticality accident likelihoods, consequences, and emergency planning

Energy Technology Data Exchange (ETDEWEB)

McLaughlin, T.P.

1991-01-01

Evaluation of criticality accident risks in the processing of significant quantities of fissile materials is both complex and subjective, largely due to the lack of accident statistics. Thus, complying with standards such as ISO 7753 which mandates that the need for an alarm system be evaluated, is also subjective. A review of guidance found in the literature on potential accident magnitudes is presented for different material forms and arrangements. Reasoned arguments are also presented concerning accident prevention and accident likelihoods for these material forms and arrangements. 13 refs., 1 fig., 1 tab.

13. Estimating likelihood of future crashes for crash-prone drivers

OpenAIRE

Subasish Das; Xiaoduan Sun; Fan Wang; Charles Leboeuf

2015-01-01

At-fault crash-prone drivers are usually considered as the high risk group for possible future incidents or crashes. In Louisiana, 34% of crashes are repeatedly committed by the at-fault crash-prone drivers who represent only 5% of the total licensed drivers in the state. This research has conducted an exploratory data analysis based on the driver faultiness and proneness. The objective of this study is to develop a crash prediction model to estimate the likelihood of future crashes for the a...

14. Failed refutations: further comments on parsimony and likelihood methods and their relationship to Popper's degree of corroboration.

Science.gov (United States)

de Queiroz, Kevin; Poe, Steven

2003-06-01

single body of data; however, both classes of methods can be used to perform severe tests. The assumption of descent with modification is insufficient background knowledge to justify cladistic parsimony as a method for assessing degree of corroboration. Invoking equivalency between parsimony methods and likelihood models that assume no common mechanism emphasizes the necessity of additional assumptions, at least some of which are probabilistic in nature. Incongruent characters do not qualify as falsifiers of phylogenetic hypotheses except under extremely unrealistic evolutionary models; therefore, justifications of parsimony methods as falsificationist based on the idea that they minimize the ad hoc dismissal of falsifiers are questionable. Probabilistic concepts such as degree of corroboration and likelihood provide a more appropriate framework for understanding how phylogenetics conforms with Popper's philosophy of science. Likelihood ratio tests do not assume what is at issue but instead are methods for testing hypotheses according to an accepted standard of statistical significance and for incorporating considerations about test severity. These tests are fundamentally similar to Popper's degree of corroboration in being based on the relationship between the probability of the evidence e in the presence versus absence of the hypothesis h, i.e., between p(e|hb) and p(e|b), where b is the background knowledge. Both parsimony and likelihood methods are inductive in that their inferences (particular trees) contain more information than (and therefore do not follow necessarily from) the observations upon which they are based; however, both are deductive in that their conclusions (tree lengths and likelihoods) follow necessarily from their premises (particular trees, observed character state distributions, and evolutionary models). For these and other reasons, phylogenetic likelihood methods are highly compatible with Karl Popper's philosophy of science and offer several

15. How are important life events disclosed on facebook? Relationships with likelihood of sharing and privacy.

Science.gov (United States)

Bevan, Jennifer L; Cummings, Megan B; Kubiniec, Ashley; Mogannam, Megan; Price, Madison; Todd, Rachel

2015-01-01

This study examined an aspect of Facebook disclosure that has as yet gone unexplored: whether a user prefers to share information directly, for example, through status updates, or indirectly, via photos with no caption or relationship status changes without context or explanation. The focus was on the sharing of important positive and negative life events related to romantic relationships, health, and work/school in relation to likelihood of sharing this type of information on Facebook and general attitudes toward privacy. An online survey of 599 adult Facebook users found that when positive life events were shared, users preferred to do so indirectly, whereas negative life events were more likely to be disclosed directly. Privacy shared little association with how information was shared. Implications for understanding the finer nuances of how news is shared on Facebook are discussed.

16. Likelihood Approximation With Parallel Hierarchical Matrices For Large Spatial Datasets

KAUST Repository

Litvinenko, Alexander

2017-11-01

The main goal of this article is to introduce the parallel hierarchical matrix library HLIBpro to the statistical community. We describe the HLIBCov package, which is an extension of the HLIBpro library for approximating large covariance matrices and maximizing likelihood functions. We show that an approximate Cholesky factorization of a dense matrix of size $2M\\\\times 2M$ can be computed on a modern multi-core desktop in few minutes. Further, HLIBCov is used for estimating the unknown parameters such as the covariance length, variance and smoothness parameter of a Matérn covariance function by maximizing the joint Gaussian log-likelihood function. The computational bottleneck here is expensive linear algebra arithmetics due to large and dense covariance matrices. Therefore covariance matrices are approximated in the hierarchical ($\\\\H$-) matrix format with computational cost $\\\\mathcal{O}(k^2n \\\\log^2 n/p)$ and storage $\\\\mathcal{O}(kn \\\\log n)$, where the rank $k$ is a small integer (typically $k<25$), $p$ the number of cores and $n$ the number of locations on a fairly general mesh. We demonstrate a synthetic example, where the true values of known parameters are known. For reproducibility we provide the C++ code, the documentation, and the synthetic data.

17. Likelihood Approximation With Parallel Hierarchical Matrices For Large Spatial Datasets

KAUST Repository

Litvinenko, Alexander; Sun, Ying; Genton, Marc G.; Keyes, David E.

2017-01-01

The main goal of this article is to introduce the parallel hierarchical matrix library HLIBpro to the statistical community. We describe the HLIBCov package, which is an extension of the HLIBpro library for approximating large covariance matrices and maximizing likelihood functions. We show that an approximate Cholesky factorization of a dense matrix of size $2M\\times 2M$ can be computed on a modern multi-core desktop in few minutes. Further, HLIBCov is used for estimating the unknown parameters such as the covariance length, variance and smoothness parameter of a Matérn covariance function by maximizing the joint Gaussian log-likelihood function. The computational bottleneck here is expensive linear algebra arithmetics due to large and dense covariance matrices. Therefore covariance matrices are approximated in the hierarchical ($\\H$-) matrix format with computational cost $\\mathcal{O}(k^2n \\log^2 n/p)$ and storage $\\mathcal{O}(kn \\log n)$, where the rank $k$ is a small integer (typically $k<25$), $p$ the number of cores and $n$ the number of locations on a fairly general mesh. We demonstrate a synthetic example, where the true values of known parameters are known. For reproducibility we provide the C++ code, the documentation, and the synthetic data.

18. Superfast maximum-likelihood reconstruction for quantum tomography

Science.gov (United States)

Shang, Jiangwei; Zhang, Zhengyun; Ng, Hui Khoon

2017-06-01

Conventional methods for computing maximum-likelihood estimators (MLE) often converge slowly in practical situations, leading to a search for simplifying methods that rely on additional assumptions for their validity. In this work, we provide a fast and reliable algorithm for maximum-likelihood reconstruction that avoids this slow convergence. Our method utilizes the state-of-the-art convex optimization scheme, an accelerated projected-gradient method, that allows one to accommodate the quantum nature of the problem in a different way than in the standard methods. We demonstrate the power of our approach by comparing its performance with other algorithms for n -qubit state tomography. In particular, an eight-qubit situation that purportedly took weeks of computation time in 2005 can now be completed in under a minute for a single set of data, with far higher accuracy than previously possible. This refutes the common claim that MLE reconstruction is slow and reduces the need for alternative methods that often come with difficult-to-verify assumptions. In fact, recent methods assuming Gaussian statistics or relying on compressed sensing ideas are demonstrably inapplicable for the situation under consideration here. Our algorithm can be applied to general optimization problems over the quantum state space; the philosophy of projected gradients can further be utilized for optimization contexts with general constraints.

19. Likelihood inference for a fractionally cointegrated vector autoregressive model

DEFF Research Database (Denmark)

Johansen, Søren; Ørregård Nielsen, Morten

2012-01-01

such that the process X_{t} is fractional of order d and cofractional of order d-b; that is, there exist vectors ß for which ß'X_{t} is fractional of order d-b, and no other fractionality order is possible. We define the statistical model by 0inference when the true values satisfy b0¿1/2 and d0-b0......We consider model based inference in a fractionally cointegrated (or cofractional) vector autoregressive model with a restricted constant term, ¿, based on the Gaussian likelihood conditional on initial values. The model nests the I(d) VAR model. We give conditions on the parameters...... process in the parameters when errors are i.i.d. with suitable moment conditions and initial values are bounded. When the limit is deterministic this implies uniform convergence in probability of the conditional likelihood function. If the true value b0>1/2, we prove that the limit distribution of (ß...

20. Likelihood-Based Inference of B Cell Clonal Families.

Directory of Open Access Journals (Sweden)

Duncan K Ralph

2016-10-01

Full Text Available The human immune system depends on a highly diverse collection of antibody-making B cells. B cell receptor sequence diversity is generated by a random recombination process called "rearrangement" forming progenitor B cells, then a Darwinian process of lineage diversification and selection called "affinity maturation." The resulting receptors can be sequenced in high throughput for research and diagnostics. Such a collection of sequences contains a mixture of various lineages, each of which may be quite numerous, or may consist of only a single member. As a step to understanding the process and result of this diversification, one may wish to reconstruct lineage membership, i.e. to cluster sampled sequences according to which came from the same rearrangement events. We call this clustering problem "clonal family inference." In this paper we describe and validate a likelihood-based framework for clonal family inference based on a multi-hidden Markov Model (multi-HMM framework for B cell receptor sequences. We describe an agglomerative algorithm to find a maximum likelihood clustering, two approximate algorithms with various trade-offs of speed versus accuracy, and a third, fast algorithm for finding specific lineages. We show that under simulation these algorithms greatly improve upon existing clonal family inference methods, and that they also give significantly different clusters than previous methods when applied to two real data sets.

1. Simulation-based marginal likelihood for cluster strong lensing cosmology

Science.gov (United States)

Killedar, M.; Borgani, S.; Fabjan, D.; Dolag, K.; Granato, G.; Meneghetti, M.; Planelles, S.; Ragone-Figueroa, C.

2018-01-01

Comparisons between observed and predicted strong lensing properties of galaxy clusters have been routinely used to claim either tension or consistency with Λ cold dark matter cosmology. However, standard approaches to such cosmological tests are unable to quantify the preference for one cosmology over another. We advocate approximating the relevant Bayes factor using a marginal likelihood that is based on the following summary statistic: the posterior probability distribution function for the parameters of the scaling relation between Einstein radii and cluster mass, α and β. We demonstrate, for the first time, a method of estimating the marginal likelihood using the X-ray selected z > 0.5 Massive Cluster Survey clusters as a case in point and employing both N-body and hydrodynamic simulations of clusters. We investigate the uncertainty in this estimate and consequential ability to compare competing cosmologies, which arises from incomplete descriptions of baryonic processes, discrepancies in cluster selection criteria, redshift distribution and dynamical state. The relation between triaxial cluster masses at various overdensities provides a promising alternative to the strong lensing test.

2. Menyoal Elaboration Likelihood Model (ELM dan Teori Retorika

Directory of Open Access Journals (Sweden)

Yudi Perbawaningsih

2012-06-01

Full Text Available Abstract: Persuasion is a communication process to establish or change attitudes, which can be understood through theory of Rhetoric and theory of Elaboration Likelihood Model (ELM. This study elaborates these theories in a Public Lecture series which to persuade the students in choosing their concentration of study. The result shows that in term of persuasion effectiveness it is not quite relevant to separate the message and its source. The quality of source is determined by the quality of the message, and vice versa. Separating the two routes of the persuasion process as described in the ELM theory would not be relevant. Abstrak: Persuasi adalah proses komunikasi untuk membentuk atau mengubah sikap, yang dapat dipahami dengan teori Retorika dan teori Elaboration Likelihood Model (ELM. Penelitian ini mengelaborasi teori tersebut dalam Kuliah Umum sebagai sarana mempersuasi mahasiswa untuk memilih konsentrasi studi studi yang didasarkan pada proses pengolahan informasi. Menggunakan metode survey, didapatkan hasil yaitu tidaklah cukup relevan memisahkan pesan dan narasumber dalam melihat efektivitas persuasi. Keduanya menyatu yang berarti bahwa kualitas narasumber ditentukan oleh kualitas pesan yang disampaikannya, dan sebaliknya. Memisahkan proses persuasi dalam dua lajur seperti yang dijelaskan dalam ELM teori menjadi tidak relevan.

3. Gauging the likelihood of stable cavitation from ultrasound contrast agents.

Science.gov (United States)

Bader, Kenneth B; Holland, Christy K

2013-01-07

The mechanical index (MI) was formulated to gauge the likelihood of adverse bioeffects from inertial cavitation. However, the MI formulation did not consider bubble activity from stable cavitation. This type of bubble activity can be readily nucleated from ultrasound contrast agents (UCAs) and has the potential to promote beneficial bioeffects. Here, the presence of stable cavitation is determined numerically by tracking the onset of subharmonic oscillations within a population of bubbles for frequencies up to 7 MHz and peak rarefactional pressures up to 3 MPa. In addition, the acoustic pressure rupture threshold of an UCA population was determined using the Marmottant model. The threshold for subharmonic emissions of optimally sized bubbles was found to be lower than the inertial cavitation threshold for all frequencies studied. The rupture thresholds of optimally sized UCAs were found to be lower than the threshold for subharmonic emissions for either single cycle or steady state acoustic excitations. Because the thresholds of both subharmonic emissions and UCA rupture are linearly dependent on frequency, an index of the form I(CAV) = P(r)/f (where P(r) is the peak rarefactional pressure in MPa and f is the frequency in MHz) was derived to gauge the likelihood of subharmonic emissions due to stable cavitation activity nucleated from UCAs.

4. Safe semi-supervised learning based on weighted likelihood.

Science.gov (United States)

Kawakita, Masanori; Takeuchi, Jun'ichi

2014-05-01

We are interested in developing a safe semi-supervised learning that works in any situation. Semi-supervised learning postulates that n(') unlabeled data are available in addition to n labeled data. However, almost all of the previous semi-supervised methods require additional assumptions (not only unlabeled data) to make improvements on supervised learning. If such assumptions are not met, then the methods possibly perform worse than supervised learning. Sokolovska, Cappé, and Yvon (2008) proposed a semi-supervised method based on a weighted likelihood approach. They proved that this method asymptotically never performs worse than supervised learning (i.e., it is safe) without any assumption. Their method is attractive because it is easy to implement and is potentially general. Moreover, it is deeply related to a certain statistical paradox. However, the method of Sokolovska et al. (2008) assumes a very limited situation, i.e., classification, discrete covariates, n(')→∞ and a maximum likelihood estimator. In this paper, we extend their method by modifying the weight. We prove that our proposal is safe in a significantly wide range of situations as long as n≤n('). Further, we give a geometrical interpretation of the proof of safety through the relationship with the above-mentioned statistical paradox. Finally, we show that the above proposal is asymptotically safe even when n(')

5. Opinion Paper: 'Likelihood-ratio' and 'odds' applied to monitoring of patients as a supplement to 'reference change value' (RCV)

DEFF Research Database (Denmark)

Petersen, Per Hyltoft; Sandberg, Sverre; Iglesias, Natàlia

2007-01-01

and odds used for diagnostic test evaluations is applied to monitoring by substituting measured concentrations with measured differences. Thus, two frequency distributions of differences are assumed, one for a stable, steady-state, situation and one for a certain change. Values exceeding a measured...

6. An Accurate Method for Inferring Relatedness in Large Datasets of Unphased Genotypes via an Embedded Likelihood-Ratio Test

KAUST Repository

Rodriguez, Jesse M.; Batzoglou, Serafim; Bercovici, Sivan

2013-01-01

, accurate and efficient detection of hidden relatedness becomes a challenge. To enable disease-mapping studies of increasingly large cohorts, a fast and accurate method to detect IBD segments is required. We present PARENTE, a novel method for detecting

7. A Multiple Model SNR/RCS Likelihood Ratio Score for Radar-Based Feature-Aided Tracking

National Research Council Canada - National Science Library

Slocumb, Benjamin J; Klusman, III, Michael E

2005-01-01

...) and radar cross section (RCS) for use in narrowband radar tracking. The formulation requires an estimate of the target mean RCS, and a key challenge is the tracking of the mean RCS through significant jumps due to aspect dependencies...

8. A dual-mode generalized likelihood ratio approach to self-reorganizing digital flight control system design

Science.gov (United States)

1976-01-01

Analytic techniques have been developed for detecting and identifying abrupt changes in dynamic systems. The GLR technique monitors the output of the Kalman filter and searches for the time that the failure occured, thus allowing it to be sensitive to new data and consequently increasing the chances for fast system recovery following detection of a failure. All failure detections are based on functional redundancy. Performance tests of the F-8 aircraft flight control system and computerized modelling of the technique are presented.

9. Penalized likelihood and multi-objective spatial scans for the detection and inference of irregular clusters

Directory of Open Access Journals (Sweden)

Fonseca Carlos M

2010-10-01

Full Text Available Abstract Background Irregularly shaped spatial clusters are difficult to delineate. A cluster found by an algorithm often spreads through large portions of the map, impacting its geographical meaning. Penalized likelihood methods for Kulldorff's spatial scan statistics have been used to control the excessive freedom of the shape of clusters. Penalty functions based on cluster geometry and non-connectivity have been proposed recently. Another approach involves the use of a multi-objective algorithm to maximize two objectives: the spatial scan statistics and the geometric penalty function. Results & Discussion We present a novel scan statistic algorithm employing a function based on the graph topology to penalize the presence of under-populated disconnection nodes in candidate clusters, the disconnection nodes cohesion function. A disconnection node is defined as a region within a cluster, such that its removal disconnects the cluster. By applying this function, the most geographically meaningful clusters are sifted through the immense set of possible irregularly shaped candidate cluster solutions. To evaluate the statistical significance of solutions for multi-objective scans, a statistical approach based on the concept of attainment function is used. In this paper we compared different penalized likelihoods employing the geometric and non-connectivity regularity functions and the novel disconnection nodes cohesion function. We also build multi-objective scans using those three functions and compare them with the previous penalized likelihood scans. An application is presented using comprehensive state-wide data for Chagas' disease in puerperal women in Minas Gerais state, Brazil. Conclusions We show that, compared to the other single-objective algorithms, multi-objective scans present better performance, regarding power, sensitivity and positive predicted value. The multi-objective non-connectivity scan is faster and better suited for the

10. Parametric Roll Resonance Detection using Phase Correlation and Log-likelihood Testing Techniques

DEFF Research Database (Denmark)

Galeazzi, Roberto; Blanke, Mogens; Poulsen, Niels Kjølstad

2009-01-01

generation warning system the purpose of which is to provide the master with an onboard system able to trigger an alarm when parametric roll is likely to happen within the immediate future. A detection scheme is introduced, which is able to issue a warning within five roll periods after a resonant motion......Real-time detection of parametric roll is still an open issue that is gathering an increasing attention. A first generation warning systems, based on guidelines and polar diagrams, showed their potential to face issues like long-term prediction and risk assessment. This paper presents a second...... started. After having determined statistical properties of the signals at hand, a detector based on the generalised log-likelihood ratio test (GLRT) is designed to look for variation in signal power. The ability of the detector to trigger alarms when parametric roll is going to onset is evaluated on two...

11. A Sum-of-Squares and Semidefinite Programming Approach for Maximum Likelihood DOA Estimation

Directory of Open Access Journals (Sweden)

Shu Cai

2016-12-01

Full Text Available Direction of arrival (DOA estimation using a uniform linear array (ULA is a classical problem in array signal processing. In this paper, we focus on DOA estimation based on the maximum likelihood (ML criterion, transform the estimation problem into a novel formulation, named as sum-of-squares (SOS, and then solve it using semidefinite programming (SDP. We first derive the SOS and SDP method for DOA estimation in the scenario of a single source and then extend it under the framework of alternating projection for multiple DOA estimation. The simulations demonstrate that the SOS- and SDP-based algorithms can provide stable and accurate DOA estimation when the number of snapshots is small and the signal-to-noise ratio (SNR is low. Moveover, it has a higher spatial resolution compared to existing methods based on the ML criterion.

12. Maximum Likelihood based comparison of the specific growth rates for P. aeruginosa and four mutator strains

DEFF Research Database (Denmark)

Philipsen, Kirsten Riber; Christiansen, Lasse Engbo; Mandsberg, Lotte Frigaard

2008-01-01

with an exponentially decaying function of the time between observations is suggested. A model with a full covariance structure containing OD-dependent variance and an autocorrelation structure is compared to a model with variance only and with no variance or correlation implemented. It is shown that the model...... are used for parameter estimation. The data is log-transformed such that a linear model can be applied. The transformation changes the variance structure, and hence an OD-dependent variance is implemented in the model. The autocorrelation in the data is demonstrated, and a correlation model...... that best describes data is a model taking into account the full covariance structure. An inference study is made in order to determine whether the growth rate of the five bacteria strains is the same. After applying a likelihood-ratio test to models with a full covariance structure, it is concluded...

13. Maximum Simulated Likelihood and Expectation-Maximization Methods to Estimate Random Coefficients Logit with Panel Data

DEFF Research Database (Denmark)

Cherchi, Elisabetta; Guevara, Cristian

2012-01-01

with cross-sectional or with panel data, and (d) EM systematically attained more efficient estimators than the MSL method. The results imply that if the purpose of the estimation is only to determine the ratios of the model parameters (e.g., the value of time), the EM method should be preferred. For all......The random coefficients logit model allows a more realistic representation of agents' behavior. However, the estimation of that model may involve simulation, which may become impractical with many random coefficients because of the curse of dimensionality. In this paper, the traditional maximum...... simulated likelihood (MSL) method is compared with the alternative expectation- maximization (EM) method, which does not require simulation. Previous literature had shown that for cross-sectional data, MSL outperforms the EM method in the ability to recover the true parameters and estimation time...

14. The asymptotic behaviour of the maximum likelihood function of Kriging approximations using the Gaussian correlation function

CSIR Research Space (South Africa)

Kok, S

2012-07-01

Full Text Available continuously as the correlation function hyper-parameters approach zero. Since the global minimizer of the maximum likelihood function is an asymptote in this case, it is unclear if maximum likelihood estimation (MLE) remains valid. Numerical ill...

15. Narrow band interference cancelation in OFDM: Astructured maximum likelihood approach

KAUST Repository

2012-06-01

This paper presents a maximum likelihood (ML) approach to mitigate the effect of narrow band interference (NBI) in a zero padded orthogonal frequency division multiplexing (ZP-OFDM) system. The NBI is assumed to be time variant and asynchronous with the frequency grid of the ZP-OFDM system. The proposed structure based technique uses the fact that the NBI signal is sparse as compared to the ZP-OFDM signal in the frequency domain. The structure is also useful in reducing the computational complexity of the proposed method. The paper also presents a data aided approach for improved NBI estimation. The suitability of the proposed method is demonstrated through simulations. © 2012 IEEE.

16. Preliminary attempt on maximum likelihood tomosynthesis reconstruction of DEI data

International Nuclear Information System (INIS)

Wang Zhentian; Huang Zhifeng; Zhang Li; Kang Kejun; Chen Zhiqiang; Zhu Peiping

2009-01-01

Tomosynthesis is a three-dimension reconstruction method that can remove the effect of superimposition with limited angle projections. It is especially promising in mammography where radiation dose is concerned. In this paper, we propose a maximum likelihood tomosynthesis reconstruction algorithm (ML-TS) on the apparent absorption data of diffraction enhanced imaging (DEI). The motivation of this contribution is to develop a tomosynthesis algorithm in low-dose or noisy circumstances and make DEI get closer to clinic application. The theoretical statistical models of DEI data in physics are analyzed and the proposed algorithm is validated with the experimental data at the Beijing Synchrotron Radiation Facility (BSRF). The results of ML-TS have better contrast compared with the well known 'shift-and-add' algorithm and FBP algorithm. (authors)

17. H.264 SVC Complexity Reduction Based on Likelihood Mode Decision

Directory of Open Access Journals (Sweden)

L. Balaji

2015-01-01

Full Text Available H.264 Advanced Video Coding (AVC was prolonged to Scalable Video Coding (SVC. SVC executes in different electronics gadgets such as personal computer, HDTV, SDTV, IPTV, and full-HDTV in which user demands various scaling of the same content. The various scaling is resolution, frame rate, quality, heterogeneous networks, bandwidth, and so forth. Scaling consumes more encoding time and computational complexity during mode selection. In this paper, to reduce encoding time and computational complexity, a fast mode decision algorithm based on likelihood mode decision (LMD is proposed. LMD is evaluated in both temporal and spatial scaling. From the results, we conclude that LMD performs well, when compared to the previous fast mode decision algorithms. The comparison parameters are time, PSNR, and bit rate. LMD achieve time saving of 66.65% with 0.05% detriment in PSNR and 0.17% increment in bit rate compared with the full search method.

18. H.264 SVC Complexity Reduction Based on Likelihood Mode Decision.

Science.gov (United States)

Balaji, L; Thyagharajan, K K

2015-01-01

H.264 Advanced Video Coding (AVC) was prolonged to Scalable Video Coding (SVC). SVC executes in different electronics gadgets such as personal computer, HDTV, SDTV, IPTV, and full-HDTV in which user demands various scaling of the same content. The various scaling is resolution, frame rate, quality, heterogeneous networks, bandwidth, and so forth. Scaling consumes more encoding time and computational complexity during mode selection. In this paper, to reduce encoding time and computational complexity, a fast mode decision algorithm based on likelihood mode decision (LMD) is proposed. LMD is evaluated in both temporal and spatial scaling. From the results, we conclude that LMD performs well, when compared to the previous fast mode decision algorithms. The comparison parameters are time, PSNR, and bit rate. LMD achieve time saving of 66.65% with 0.05% detriment in PSNR and 0.17% increment in bit rate compared with the full search method.

19. Likelihood Approximation With Hierarchical Matrices For Large Spatial Datasets

KAUST Repository

Litvinenko, Alexander

2017-09-03

We use available measurements to estimate the unknown parameters (variance, smoothness parameter, and covariance length) of a covariance function by maximizing the joint Gaussian log-likelihood function. To overcome cubic complexity in the linear algebra, we approximate the discretized covariance function in the hierarchical (H-) matrix format. The H-matrix format has a log-linear computational cost and storage O(kn log n), where the rank k is a small integer and n is the number of locations. The H-matrix technique allows us to work with general covariance matrices in an efficient way, since H-matrices can approximate inhomogeneous covariance functions, with a fairly general mesh that is not necessarily axes-parallel, and neither the covariance matrix itself nor its inverse have to be sparse. We demonstrate our method with Monte Carlo simulations and an application to soil moisture data. The C, C++ codes and data are freely available.

20. Music genre classification via likelihood fusion from multiple feature models

Science.gov (United States)

Shiu, Yu; Kuo, C.-C. J.

2005-01-01

Music genre provides an efficient way to index songs in a music database, and can be used as an effective means to retrieval music of a similar type, i.e. content-based music retrieval. A new two-stage scheme for music genre classification is proposed in this work. At the first stage, we examine a couple of different features, construct their corresponding parametric models (e.g. GMM and HMM) and compute their likelihood functions to yield soft classification results. In particular, the timbre, rhythm and temporal variation features are considered. Then, at the second stage, these soft classification results are integrated to result in a hard decision for final music genre classification. Experimental results are given to demonstrate the performance of the proposed scheme.

1. Marginal Maximum Likelihood Estimation of Item Response Models in R

Directory of Open Access Journals (Sweden)

Matthew S. Johnson

2007-02-01

Full Text Available Item response theory (IRT models are a class of statistical models used by researchers to describe the response behaviors of individuals to a set of categorically scored items. The most common IRT models can be classified as generalized linear fixed- and/or mixed-effect models. Although IRT models appear most often in the psychological testing literature, researchers in other fields have successfully utilized IRT-like models in a wide variety of applications. This paper discusses the three major methods of estimation in IRT and develops R functions utilizing the built-in capabilities of the R environment to find the marginal maximum likelihood estimates of the generalized partial credit model. The currently available R packages ltm is also discussed.

2. Maximum likelihood estimation of phase-type distributions

DEFF Research Database (Denmark)

Esparza, Luz Judith R

for both univariate and multivariate cases. Methods like the EM algorithm and Markov chain Monte Carlo are applied for this purpose. Furthermore, this thesis provides explicit formulae for computing the Fisher information matrix for discrete and continuous phase-type distributions, which is needed to find......This work is concerned with the statistical inference of phase-type distributions and the analysis of distributions with rational Laplace transform, known as matrix-exponential distributions. The thesis is focused on the estimation of the maximum likelihood parameters of phase-type distributions...... confidence regions for their estimated parameters. Finally, a new general class of distributions, called bilateral matrix-exponential distributions, is defined. These distributions have the entire real line as domain and can be used, for instance, for modelling. In addition, this class of distributions...

3. The elaboration likelihood model and communication about food risks.

Science.gov (United States)

Frewer, L J; Howard, C; Hedderley, D; Shepherd, R

1997-12-01

Factors such as hazard type and source credibility have been identified as important in the establishment of effective strategies for risk communication. The elaboration likelihood model was adapted to investigate the potential impact of hazard type, information source, and persuasive content of information on individual engagement in elaborative, or thoughtful, cognitions about risk messages. One hundred sixty respondents were allocated to one of eight experimental groups, and the effects of source credibility, persuasive content of information and hazard type were systematically varied. The impact of the different factors on beliefs about the information and elaborative processing examined. Low credibility was particularly important in reducing risk perceptions, although persuasive content and hazard type were also influential in determining whether elaborative processing occurred.

4. Maximum Likelihood Blood Velocity Estimator Incorporating Properties of Flow Physics

DEFF Research Database (Denmark)

Schlaikjer, Malene; Jensen, Jørgen Arendt

2004-01-01

)-data under investigation. The flow physic properties are exploited in the second term, as the range of velocity values investigated in the cross-correlation analysis are compared to the velocity estimates in the temporal and spatial neighborhood of the signal segment under investigation. The new estimator...... has been compared to the cross-correlation (CC) estimator and the previously developed maximum likelihood estimator (MLE). The results show that the CMLE can handle a larger velocity search range and is capable of estimating even low velocity levels from tissue motion. The CC and the MLE produce...... for the CC and the MLE. When the velocity search range is set to twice the limit of the CC and the MLE, the number of incorrect velocity estimates are 0, 19.1, and 7.2% for the CMLE, CC, and MLE, respectively. The ability to handle a larger search range and estimating low velocity levels was confirmed...

5. Accelerated maximum likelihood parameter estimation for stochastic biochemical systems

Directory of Open Access Journals (Sweden)

Daigle Bernie J

2012-05-01

Full Text Available Abstract Background A prerequisite for the mechanistic simulation of a biochemical system is detailed knowledge of its kinetic parameters. Despite recent experimental advances, the estimation of unknown parameter values from observed data is still a bottleneck for obtaining accurate simulation results. Many methods exist for parameter estimation in deterministic biochemical systems; methods for discrete stochastic systems are less well developed. Given the probabilistic nature of stochastic biochemical models, a natural approach is to choose parameter values that maximize the probability of the observed data with respect to the unknown parameters, a.k.a. the maximum likelihood parameter estimates (MLEs. MLE computation for all but the simplest models requires the simulation of many system trajectories that are consistent with experimental data. For models with unknown parameters, this presents a computational challenge, as the generation of consistent trajectories can be an extremely rare occurrence. Results We have developed Monte Carlo Expectation-Maximization with Modified Cross-Entropy Method (MCEM2: an accelerated method for calculating MLEs that combines advances in rare event simulation with a computationally efficient version of the Monte Carlo expectation-maximization (MCEM algorithm. Our method requires no prior knowledge regarding parameter values, and it automatically provides a multivariate parameter uncertainty estimate. We applied the method to five stochastic systems of increasing complexity, progressing from an analytically tractable pure-birth model to a computationally demanding model of yeast-polarization. Our results demonstrate that MCEM2 substantially accelerates MLE computation on all tested models when compared to a stand-alone version of MCEM. Additionally, we show how our method identifies parameter values for certain classes of models more accurately than two recently proposed computationally efficient methods

6. CONSTRUCTING A FLEXIBLE LIKELIHOOD FUNCTION FOR SPECTROSCOPIC INFERENCE

International Nuclear Information System (INIS)

Czekala, Ian; Andrews, Sean M.; Mandel, Kaisey S.; Green, Gregory M.; Hogg, David W.

2015-01-01

We present a modular, extensible likelihood framework for spectroscopic inference based on synthetic model spectra. The subtraction of an imperfect model from a continuously sampled spectrum introduces covariance between adjacent datapoints (pixels) into the residual spectrum. For the high signal-to-noise data with large spectral range that is commonly employed in stellar astrophysics, that covariant structure can lead to dramatically underestimated parameter uncertainties (and, in some cases, biases). We construct a likelihood function that accounts for the structure of the covariance matrix, utilizing the machinery of Gaussian process kernels. This framework specifically addresses the common problem of mismatches in model spectral line strengths (with respect to data) due to intrinsic model imperfections (e.g., in the atomic/molecular databases or opacity prescriptions) by developing a novel local covariance kernel formalism that identifies and self-consistently downweights pathological spectral line “outliers.” By fitting many spectra in a hierarchical manner, these local kernels provide a mechanism to learn about and build data-driven corrections to synthetic spectral libraries. An open-source software implementation of this approach is available at http://iancze.github.io/Starfish, including a sophisticated probabilistic scheme for spectral interpolation when using model libraries that are sparsely sampled in the stellar parameters. We demonstrate some salient features of the framework by fitting the high-resolution V-band spectrum of WASP-14, an F5 dwarf with a transiting exoplanet, and the moderate-resolution K-band spectrum of Gliese 51, an M5 field dwarf

7. Likelihood of illegal alcohol sales at professional sport stadiums.

Science.gov (United States)

Toomey, Traci L; Erickson, Darin J; Lenk, Kathleen M; Kilian, Gunna R

2008-11-01

Several studies have assessed the propensity for illegal alcohol sales at licensed alcohol establishments and community festivals, but no previous studies examined the propensity for these sales at professional sport stadiums. In this study, we assessed the likelihood of alcohol sales to both underage youth and obviously intoxicated patrons at professional sports stadiums across the United States, and assessed the factors related to likelihood of both types of alcohol sales. We conducted pseudo-underage (i.e., persons age 21 or older who appear under 21) and pseudo-intoxicated (i.e., persons feigning intoxication) alcohol purchase attempts at stadiums that house professional hockey, basketball, baseball, and football teams. We conducted the purchase attempts at 16 sport stadiums located in 5 states. We measured 2 outcome variables: pseudo-underage sale (yes, no) and pseudo-intoxicated sale (yes, no), and 3 types of independent variables: (1) seller characteristics, (2) purchase attempt characteristics, and (3) event characteristics. Following univariate and bivariate analyses, we a separate series of logistic generalized mixed regression models for each outcome variable. The overall sales rates to the pseudo-underage and pseudo-intoxicated buyers were 18% and 74%, respectively. In the multivariate logistic analyses, we found that the odds of a sale to a pseudo-underage buyer in the stands was 2.9 as large as the odds of a sale at the concession booths (30% vs. 13%; p = 0.01). The odds of a sale to an obviously intoxicated buyer in the stands was 2.9 as large as the odds of a sale at the concession booths (89% vs. 73%; p = 0.02). Similar to studies assessing illegal alcohol sales at licensed alcohol establishments and community festivals, findings from this study shows the need for interventions specifically focused on illegal alcohol sales at professional sporting events.

8. Maximum likelihood fitting of FROC curves under an initial-detection-and-candidate-analysis model

International Nuclear Information System (INIS)

Edwards, Darrin C.; Kupinski, Matthew A.; Metz, Charles E.; Nishikawa, Robert M.

2002-01-01

We have developed a model for FROC curve fitting that relates the observer's FROC performance not to the ROC performance that would be obtained if the observer's responses were scored on a per image basis, but rather to a hypothesized ROC performance that the observer would obtain in the task of classifying a set of 'candidate detections' as positive or negative. We adopt the assumptions of the Bunch FROC model, namely that the observer's detections are all mutually independent, as well as assumptions qualitatively similar to, but different in nature from, those made by Chakraborty in his AFROC scoring methodology. Under the assumptions of our model, we show that the observer's FROC performance is a linearly scaled version of the candidate analysis ROC curve, where the scaling factors are just given by the FROC operating point coordinates for detecting initial candidates. Further, we show that the likelihood function of the model parameters given observational data takes on a simple form, and we develop a maximum likelihood method for fitting a FROC curve to this data. FROC and AFROC curves are produced for computer vision observer datasets and compared with the results of the AFROC scoring method. Although developed primarily with computer vision schemes in mind, we hope that the methodology presented here will prove worthy of further study in other applications as well

9. Maximum-likelihood estimation of the hyperbolic parameters from grouped observations

DEFF Research Database (Denmark)

Jensen, Jens Ledet

1988-01-01

a least-squares problem. The second procedure Hypesti first approaches the maximum-likelihood estimate by iterating in the profile-log likelihood function for the scale parameter. Close to the maximum of the likelihood function, the estimation is brought to an end by iteration, using all four parameters...

10. A short proof that phylogenetic tree reconstruction by maximum likelihood is hard.

Science.gov (United States)

Roch, Sebastien

2006-01-01

Maximum likelihood is one of the most widely used techniques to infer evolutionary histories. Although it is thought to be intractable, a proof of its hardness has been lacking. Here, we give a short proof that computing the maximum likelihood tree is NP-hard by exploiting a connection between likelihood and parsimony observed by Tuffley and Steel.

11. A Short Proof that Phylogenetic Tree Reconstruction by Maximum Likelihood is Hard

OpenAIRE

Roch, S.

2005-01-01

Maximum likelihood is one of the most widely used techniques to infer evolutionary histories. Although it is thought to be intractable, a proof of its hardness has been lacking. Here, we give a short proof that computing the maximum likelihood tree is NP-hard by exploiting a connection between likelihood and parsimony observed by Tuffley and Steel.

12. Financial Key Ratios

OpenAIRE

Tănase Alin-Eliodor

2014-01-01

This article focuses on computing techniques starting from trial balance data regarding financial key ratios. There are presented activity, liquidity, solvency and profitability financial key ratios. It is presented a computing methodology in three steps based on a trial balance.

13. Diagnostic Performance of Bronchoalveolar Lavage Fluid CD4/CD8 Ratio for Sarcoidosis: A Meta-analysis.

Science.gov (United States)

Shen, Yongchun; Pang, Caishuang; Wu, Yanqiu; Li, Diandian; Wan, Chun; Liao, Zenglin; Yang, Ting; Chen, Lei; Wen, Fuqiang

2016-06-01

The usefulness of bronchoalveolar lavage fluid (BALF) CD4/CD8 ratio for diagnosing sarcoidosis has been reported in many studies with variable results. Therefore, we performed a meta-analysis to estimate the overall diagnostic accuracy of BALF CD4/CD8 ratio based on the bulk of published evidence. Studies published prior to June 2015 and indexed in PubMed, OVID, Web of Science, Scopus and other databases were evaluated for inclusion. Data on sensitivity, specificity, positive likelihood ratio (PLR), negative likelihood ratio (NLR), and diagnostic odds ratio (DOR) were pooled from included studies. Summary receiver operating characteristic (SROC) curves were used to summarize overall test performance. Deeks's funnel plot was used to detect publication bias. Sixteen publications with 1885 subjects met our inclusion criteria and were included in this meta-analysis. Summary estimates of the diagnostic performance of the BALF CD4/CD8 ratio were as follows: sensitivity, 0.70 (95%CI 0.64-0.75); specificity, 0.83 (95%CI 0.78-0.86); PLR, 4.04 (95%CI 3.13-5.20); NLR, 0.36 (95%CI 0.30-0.44); and DOR, 11.17 (95%CI 7.31-17.07). The area under the SROC curve was 0.84 (95%CI 0.81-0.87). There was no evidence of publication bias. Measuring the BALF CD4/CD8 ratio may assist in the diagnosis of sarcoidosis when interpreted in parallel with other diagnostic factors. Copyright © 2016 The Authors. Published by Elsevier B.V. All rights reserved.

14. Number of Siblings During Childhood and the Likelihood of Divorce in Adulthood.

Science.gov (United States)

Bobbitt-Zeher, Donna; Downey, Douglas B; Merry, Joseph

2016-11-01

Despite fertility decline across economically developed countries, relatively little is known about the social consequences of children being raised with fewer siblings. Much research suggests that growing up with fewer siblings is probably positive, as children tend to do better in school when sibship size is small. Less scholarship, however, has explored how growing up with few siblings influences children's ability to get along with peers and develop long-term meaningful relationships. If siblings serve as important social practice partners during childhood, individuals with few or no siblings may struggle to develop successful social lives later in adulthood. With data from the General Social Surveys 1972-2012 , we explore this possibility by testing whether sibship size during childhood predicts the probability of divorce in adulthood. We find that, among those who ever marry, each additional sibling is associated with a three percent decline in the likelihood of divorce, net of covariates.

15. Planck 2013 results. XV. CMB power spectra and likelihood

Science.gov (United States)

Planck Collaboration; Ade, P. A. R.; Aghanim, N.; Armitage-Caplan, C.; Arnaud, M.; Ashdown, M.; Atrio-Barandela, F.; Aumont, J.; Baccigalupi, C.; Banday, A. J.; Barreiro, R. B.; Bartlett, J. G.; Battaner, E.; Benabed, K.; Benoît, A.; Benoit-Lévy, A.; Bernard, J.-P.; Bersanelli, M.; Bielewicz, P.; Bobin, J.; Bock, J. J.; Bonaldi, A.; Bonavera, L.; Bond, J. R.; Borrill, J.; Bouchet, F. R.; Boulanger, F.; Bridges, M.; Bucher, M.; Burigana, C.; Butler, R. C.; Calabrese, E.; Cardoso, J.-F.; Catalano, A.; Challinor, A.; Chamballu, A.; Chiang, H. C.; Chiang, L.-Y.; Christensen, P. R.; Church, S.; Clements, D. L.; Colombi, S.; Colombo, L. P. L.; Combet, C.; Couchot, F.; Coulais, A.; Crill, B. P.; Curto, A.; Cuttaia, F.; Danese, L.; Davies, R. D.; Davis, R. J.; de Bernardis, P.; de Rosa, A.; de Zotti, G.; Delabrouille, J.; Delouis, J.-M.; Désert, F.-X.; Dickinson, C.; Diego, J. M.; Dole, H.; Donzelli, S.; Doré, O.; Douspis, M.; Dunkley, J.; Dupac, X.; Efstathiou, G.; Elsner, F.; Enßlin, T. A.; Eriksen, H. K.; Finelli, F.; Forni, O.; Frailis, M.; Fraisse, A. A.; Franceschi, E.; Gaier, T. C.; Galeotta, S.; Galli, S.; Ganga, K.; Giard, M.; Giardino, G.; Giraud-Héraud, Y.; Gjerløw, E.; González-Nuevo, J.; Górski, K. M.; Gratton, S.; Gregorio, A.; Gruppuso, A.; Gudmundsson, J. E.; Hansen, F. K.; Hanson, D.; Harrison, D.; Helou, G.; Henrot-Versillé, S.; Hernández-Monteagudo, C.; Herranz, D.; Hildebrandt, S. R.; Hivon, E.; Hobson, M.; Holmes, W. A.; Hornstrup, A.; Hovest, W.; Huffenberger, K. M.; Hurier, G.; Jaffe, A. H.; Jaffe, T. R.; Jewell, J.; Jones, W. C.; Juvela, M.; Keihänen, E.; Keskitalo, R.; Kiiveri, K.; Kisner, T. S.; Kneissl, R.; Knoche, J.; Knox, L.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lähteenmäki, A.; Lamarre, J.-M.; Lasenby, A.; Lattanzi, M.; Laureijs, R. J.; Lawrence, C. R.; Le Jeune, M.; Leach, S.; Leahy, J. P.; Leonardi, R.; León-Tavares, J.; Lesgourgues, J.; Liguori, M.; Lilje, P. B.; Linden-Vørnle, M.; Lindholm, V.; López-Caniego, M.; Lubin, P. M.; Macías-Pérez, J. F.; Maffei, B.; Maino, D.; Mandolesi, N.; Marinucci, D.; Maris, M.; Marshall, D. J.; Martin, P. G.; Martínez-González, E.; Masi, S.; Massardi, M.; Matarrese, S.; Matthai, F.; Mazzotta, P.; Meinhold, P. R.; Melchiorri, A.; Mendes, L.; Menegoni, E.; Mennella, A.; Migliaccio, M.; Millea, M.; Mitra, S.; Miville-Deschênes, M.-A.; Molinari, D.; Moneti, A.; Montier, L.; Morgante, G.; Mortlock, D.; Moss, A.; Munshi, D.; Murphy, J. A.; Naselsky, P.; Nati, F.; Natoli, P.; Netterfield, C. B.; Nørgaard-Nielsen, H. U.; Noviello, F.; Novikov, D.; Novikov, I.; O'Dwyer, I. J.; Orieux, F.; Osborne, S.; Oxborrow, C. A.; Paci, F.; Pagano, L.; Pajot, F.; Paladini, R.; Paoletti, D.; Partridge, B.; Pasian, F.; Patanchon, G.; Paykari, P.; Perdereau, O.; Perotto, L.; Perrotta, F.; Piacentini, F.; Piat, M.; Pierpaoli, E.; Pietrobon, D.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Ponthieu, N.; Popa, L.; Poutanen, T.; Pratt, G. W.; Prézeau, G.; Prunet, S.; Puget, J.-L.; Rachen, J. P.; Rahlin, A.; Rebolo, R.; Reinecke, M.; Remazeilles, M.; Renault, C.; Ricciardi, S.; Riller, T.; Ringeval, C.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Roudier, G.; Rowan-Robinson, M.; Rubiño-Martín, J. A.; Rusholme, B.; Sandri, M.; Sanselme, L.; Santos, D.; Savini, G.; Scott, D.; Seiffert, M. D.; Shellard, E. P. S.; Spencer, L. D.; Starck, J.-L.; Stolyarov, V.; Stompor, R.; Sudiwala, R.; Sureau, F.; Sutton, D.; Suur-Uski, A.-S.; Sygnet, J.-F.; Tauber, J. A.; Tavagnacco, D.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tristram, M.; Tucci, M.; Tuovinen, J.; Türler, M.; Valenziano, L.; Valiviita, J.; Van Tent, B.; Varis, J.; Vielva, P.; Villa, F.; Vittorio, N.; Wade, L. A.; Wandelt, B. D.; Wehus, I. K.; White, M.; White, S. D. M.; Yvon, D.; Zacchei, A.; Zonca, A.

2014-11-01

This paper presents the Planck 2013 likelihood, a complete statistical description of the two-point correlation function of the CMB temperature fluctuations that accounts for all known relevant uncertainties, both instrumental and astrophysical in nature. We use this likelihood to derive our best estimate of the CMB angular power spectrum from Planck over three decades in multipole moment, ℓ, covering 2 ≤ ℓ ≤ 2500. The main source of uncertainty at ℓ ≲ 1500 is cosmic variance. Uncertainties in small-scale foreground modelling and instrumental noise dominate the error budget at higher ℓs. For ℓ impact of residual foreground and instrumental uncertainties on the final cosmological parameters. We find good internal agreement among the high-ℓ cross-spectra with residuals below a few μK2 at ℓ ≲ 1000, in agreement with estimated calibration uncertainties. We compare our results with foreground-cleaned CMB maps derived from all Planck frequencies, as well as with cross-spectra derived from the 70 GHz Planck map, and find broad agreement in terms of spectrum residuals and cosmological parameters. We further show that the best-fit ΛCDM cosmology is in excellent agreement with preliminary PlanckEE and TE polarisation spectra. We find that the standard ΛCDM cosmology is well constrained by Planck from the measurements at ℓ ≲ 1500. One specific example is the spectral index of scalar perturbations, for which we report a 5.4σ deviation from scale invariance, ns = 1. Increasing the multipole range beyond ℓ ≃ 1500 does not increase our accuracy for the ΛCDM parameters, but instead allows us to study extensions beyond the standard model. We find no indication of significant departures from the ΛCDM framework. Finally, we report a tension between the Planck best-fit ΛCDM model and the low-ℓ spectrum in the form of a power deficit of 5-10% at ℓ ≲ 40, with a statistical significance of 2.5-3σ. Without a theoretically motivated model for

16. Efficient algorithms for maximum likelihood decoding in the surface code

Science.gov (United States)

Bravyi, Sergey; Suchara, Martin; Vargo, Alexander

2014-09-01

We describe two implementations of the optimal error correction algorithm known as the maximum likelihood decoder (MLD) for the two-dimensional surface code with a noiseless syndrome extraction. First, we show how to implement MLD exactly in time O (n2), where n is the number of code qubits. Our implementation uses a reduction from MLD to simulation of matchgate quantum circuits. This reduction however requires a special noise model with independent bit-flip and phase-flip errors. Secondly, we show how to implement MLD approximately for more general noise models using matrix product states (MPS). Our implementation has running time O (nχ3), where χ is a parameter that controls the approximation precision. The key step of our algorithm, borrowed from the density matrix renormalization-group method, is a subroutine for contracting a tensor network on the two-dimensional grid. The subroutine uses MPS with a bond dimension χ to approximate the sequence of tensors arising in the course of contraction. We benchmark the MPS-based decoder against the standard minimum weight matching decoder observing a significant reduction of the logical error probability for χ ≥4.

17. Maximum likelihood sequence estimation for optical complex direct modulation.

Science.gov (United States)

Che, Di; Yuan, Feng; Shieh, William

2017-04-17

Semiconductor lasers are versatile optical transmitters in nature. Through the direct modulation (DM), the intensity modulation is realized by the linear mapping between the injection current and the light power, while various angle modulations are enabled by the frequency chirp. Limited by the direct detection, DM lasers used to be exploited only as 1-D (intensity or angle) transmitters by suppressing or simply ignoring the other modulation. Nevertheless, through the digital coherent detection, simultaneous intensity and angle modulations (namely, 2-D complex DM, CDM) can be realized by a single laser diode. The crucial technique of CDM is the joint demodulation of intensity and differential phase with the maximum likelihood sequence estimation (MLSE), supported by a closed-form discrete signal approximation of frequency chirp to characterize the MLSE transition probability. This paper proposes a statistical method for the transition probability to significantly enhance the accuracy of the chirp model. Using the statistical estimation, we demonstrate the first single-channel 100-Gb/s PAM-4 transmission over 1600-km fiber with only 10G-class DM lasers.

18. Maximum likelihood estimation for cytogenetic dose-response curves

International Nuclear Information System (INIS)

Frome, E.L; DuFrain, R.J.

1983-10-01

In vitro dose-response curves are used to describe the relation between the yield of dicentric chromosome aberrations and radiation dose for human lymphocytes. The dicentric yields follow the Poisson distribution, and the expected yield depends on both the magnitude and the temporal distribution of the dose for low LET radiation. A general dose-response model that describes this relation has been obtained by Kellerer and Rossi using the theory of dual radiation action. The yield of elementary lesions is kappa[γd + g(t, tau)d 2 ], where t is the time and d is dose. The coefficient of the d 2 term is determined by the recovery function and the temporal mode of irradiation. Two special cases of practical interest are split-dose and continuous exposure experiments, and the resulting models are intrinsically nonlinear in the parameters. A general purpose maximum likelihood estimation procedure is described and illustrated with numerical examples from both experimental designs. Poisson regression analysis is used for estimation, hypothesis testing, and regression diagnostics. Results are discussed in the context of exposure assessment procedures for both acute and chronic human radiation exposure

19. Ringing Artefact Reduction By An Efficient Likelihood Improvement Method

Science.gov (United States)

Fuderer, Miha

1989-10-01

In MR imaging, the extent of the acquired spatial frequencies of the object is necessarily finite. The resulting image shows artefacts caused by "truncation" of its Fourier components. These are known as Gibbs artefacts or ringing artefacts. These artefacts are particularly. visible when the time-saving reduced acquisition method is used, say, when scanning only the lowest 70% of the 256 data lines. Filtering the data results in loss of resolution. A method is described that estimates the high frequency data from the low-frequency data lines, with the likelihood of the image as criterion. It is a computationally very efficient method, since it requires practically only two extra Fourier transforms, in addition to the normal. reconstruction. The results of this method on MR images of human subjects are promising. Evaluations on a 70% acquisition image show about 20% decrease of the error energy after processing. "Error energy" is defined as the total power of the difference to a 256-data-lines reference image. The elimination of ringing artefacts then appears almost complete..

20. Maximum-likelihood estimation of recent shared ancestry (ERSA).

Science.gov (United States)

Huff, Chad D; Witherspoon, David J; Simonson, Tatum S; Xing, Jinchuan; Watkins, W Scott; Zhang, Yuhua; Tuohy, Therese M; Neklason, Deborah W; Burt, Randall W; Guthery, Stephen L; Woodward, Scott R; Jorde, Lynn B

2011-05-01

Accurate estimation of recent shared ancestry is important for genetics, evolution, medicine, conservation biology, and forensics. Established methods estimate kinship accurately for first-degree through third-degree relatives. We demonstrate that chromosomal segments shared by two individuals due to identity by descent (IBD) provide much additional information about shared ancestry. We developed a maximum-likelihood method for the estimation of recent shared ancestry (ERSA) from the number and lengths of IBD segments derived from high-density SNP or whole-genome sequence data. We used ERSA to estimate relationships from SNP genotypes in 169 individuals from three large, well-defined human pedigrees. ERSA is accurate to within one degree of relationship for 97% of first-degree through fifth-degree relatives and 80% of sixth-degree and seventh-degree relatives. We demonstrate that ERSA's statistical power approaches the maximum theoretical limit imposed by the fact that distant relatives frequently share no DNA through a common ancestor. ERSA greatly expands the range of relationships that can be estimated from genetic data and is implemented in a freely available software package.

1. Quantifying uncertainty, variability and likelihood for ordinary differential equation models

LENUS (Irish Health Repository)

Weisse, Andrea Y

2010-10-28

Abstract Background In many applications, ordinary differential equation (ODE) models are subject to uncertainty or variability in initial conditions and parameters. Both, uncertainty and variability can be quantified in terms of a probability density function on the state and parameter space. Results The partial differential equation that describes the evolution of this probability density function has a form that is particularly amenable to application of the well-known method of characteristics. The value of the density at some point in time is directly accessible by the solution of the original ODE extended by a single extra dimension (for the value of the density). This leads to simple methods for studying uncertainty, variability and likelihood, with significant advantages over more traditional Monte Carlo and related approaches especially when studying regions with low probability. Conclusions While such approaches based on the method of characteristics are common practice in other disciplines, their advantages for the study of biological systems have so far remained unrecognized. Several examples illustrate performance and accuracy of the approach and its limitations.

2. Affective mapping: An activation likelihood estimation (ALE) meta-analysis.

Science.gov (United States)

Kirby, Lauren A J; Robinson, Jennifer L

2017-11-01

Functional neuroimaging has the spatial resolution to explain the neural basis of emotions. Activation likelihood estimation (ALE), as opposed to traditional qualitative meta-analysis, quantifies convergence of activation across studies within affective categories. Others have used ALE to investigate a broad range of emotions, but without the convenience of the BrainMap database. We used the BrainMap database and analysis resources to run separate meta-analyses on coordinates reported for anger, anxiety, disgust, fear, happiness, humor, and sadness. Resultant ALE maps were compared to determine areas of convergence between emotions, as well as to identify affect-specific networks. Five out of the seven emotions demonstrated consistent activation within the amygdala, whereas all emotions consistently activated the right inferior frontal gyrus, which has been implicated as an integration hub for affective and cognitive processes. These data provide the framework for models of affect-specific networks, as well as emotional processing hubs, which can be used for future studies of functional or effective connectivity. Copyright © 2015 Elsevier Inc. All rights reserved.

3. Dark matter CMB constraints and likelihoods for poor particle physicists

Energy Technology Data Exchange (ETDEWEB)

Cline, James M.; Scott, Pat, E-mail: jcline@physics.mcgill.ca, E-mail: patscott@physics.mcgill.ca [Department of Physics, McGill University, 3600 rue University, Montréal, QC, H3A 2T8 (Canada)

2013-03-01

The cosmic microwave background provides constraints on the annihilation and decay of light dark matter at redshifts between 100 and 1000, the strength of which depends upon the fraction of energy ending up in the form of electrons and photons. The resulting constraints are usually presented for a limited selection of annihilation and decay channels. Here we provide constraints on the annihilation cross section and decay rate, at discrete values of the dark matter mass m{sub χ}, for all the annihilation and decay channels whose secondary spectra have been computed using PYTHIA in arXiv:1012.4515 (''PPPC 4 DM ID: a poor particle physicist cookbook for dark matter indirect detection''), namely e, μ, τ, V → e, V → μ, V → τ, u, d s, c, b, t, γ, g, W, Z and h. By interpolating in mass, these can be used to find the CMB constraints and likelihood functions from WMAP7 and Planck for a wide range of dark matter models, including those with annihilation or decay into a linear combination of different channels.

4. Dark matter CMB constraints and likelihoods for poor particle physicists

International Nuclear Information System (INIS)

Cline, James M.; Scott, Pat

2013-01-01

The cosmic microwave background provides constraints on the annihilation and decay of light dark matter at redshifts between 100 and 1000, the strength of which depends upon the fraction of energy ending up in the form of electrons and photons. The resulting constraints are usually presented for a limited selection of annihilation and decay channels. Here we provide constraints on the annihilation cross section and decay rate, at discrete values of the dark matter mass m χ , for all the annihilation and decay channels whose secondary spectra have been computed using PYTHIA in arXiv:1012.4515 (''PPPC 4 DM ID: a poor particle physicist cookbook for dark matter indirect detection''), namely e, μ, τ, V → e, V → μ, V → τ, u, d s, c, b, t, γ, g, W, Z and h. By interpolating in mass, these can be used to find the CMB constraints and likelihood functions from WMAP7 and Planck for a wide range of dark matter models, including those with annihilation or decay into a linear combination of different channels

5. Maximum likelihood estimation for cytogenetic dose-response curves

International Nuclear Information System (INIS)

Frome, E.L.; DuFrain, R.J.

1986-01-01

In vitro dose-response curves are used to describe the relation between chromosome aberrations and radiation dose for human lymphocytes. The lymphocytes are exposed to low-LET radiation, and the resulting dicentric chromosome aberrations follow the Poisson distribution. The expected yield depends on both the magnitude and the temporal distribution of the dose. A general dose-response model that describes this relation has been presented by Kellerer and Rossi (1972, Current Topics on Radiation Research Quarterly 8, 85-158; 1978, Radiation Research 75, 471-488) using the theory of dual radiation action. Two special cases of practical interest are split-dose and continuous exposure experiments, and the resulting dose-time-response models are intrinsically nonlinear in the parameters. A general-purpose maximum likelihood estimation procedure is described, and estimation for the nonlinear models is illustrated with numerical examples from both experimental designs. Poisson regression analysis is used for estimation, hypothesis testing, and regression diagnostics. Results are discussed in the context of exposure assessment procedures for both acute and chronic human radiation exposure

6. Physical activity may decrease the likelihood of children developing constipation.

Science.gov (United States)

Seidenfaden, Sandra; Ormarsson, Orri Thor; Lund, Sigrun H; Bjornsson, Einar S

2018-01-01

Childhood constipation is common. We evaluated children diagnosed with constipation, who were referred to an Icelandic paediatric emergency department, and determined the effect of lifestyle factors on its aetiology. The parents of children who were diagnosed with constipation and participated in a phase IIB clinical trial on laxative suppositories answered an online questionnaire about their children's lifestyle and constipation in March-April 2013. The parents of nonconstipated children that visited the paediatric department of Landspitali University Hospital or an Icelandic outpatient clinic answered the same questionnaire. We analysed responses regarding 190 children aged one year to 18 years: 60 with constipation and 130 without. We found that 40% of the constipated children had recurrent symptoms, 27% had to seek medical attention more than once and 33% received medication per rectum. The 47 of 130 control group subjects aged 10-18 were much more likely to exercise more than three times a week (72%) and for more than a hour (62%) than the 26 of 60 constipated children of the same age (42% and 35%, respectively). Constipation risk factors varied with age and many children diagnosed with constipation had recurrent symptoms. Physical activity may affect the likelihood of developing constipation in older children. ©2017 Foundation Acta Paediatrica. Published by John Wiley & Sons Ltd.

7. Maximum likelihood pedigree reconstruction using integer linear programming.

Science.gov (United States)

Cussens, James; Bartlett, Mark; Jones, Elinor M; Sheehan, Nuala A

2013-01-01

Large population biobanks of unrelated individuals have been highly successful in detecting common genetic variants affecting diseases of public health concern. However, they lack the statistical power to detect more modest gene-gene and gene-environment interaction effects or the effects of rare variants for which related individuals are ideally required. In reality, most large population studies will undoubtedly contain sets of undeclared relatives, or pedigrees. Although a crude measure of relatedness might sometimes suffice, having a good estimate of the true pedigree would be much more informative if this could be obtained efficiently. Relatives are more likely to share longer haplotypes around disease susceptibility loci and are hence biologically more informative for rare variants than unrelated cases and controls. Distant relatives are arguably more useful for detecting variants with small effects because they are less likely to share masking environmental effects. Moreover, the identification of relatives enables appropriate adjustments of statistical analyses that typically assume unrelatedness. We propose to exploit an integer linear programming optimisation approach to pedigree learning, which is adapted to find valid pedigrees by imposing appropriate constraints. Our method is not restricted to small pedigrees and is guaranteed to return a maximum likelihood pedigree. With additional constraints, we can also search for multiple high-probability pedigrees and thus account for the inherent uncertainty in any particular pedigree reconstruction. The true pedigree is found very quickly by comparison with other methods when all individuals are observed. Extensions to more complex problems seem feasible. © 2012 Wiley Periodicals, Inc.

8. Constructing diagnostic likelihood: clinical decisions using subjective versus statistical probability.

Science.gov (United States)

Kinnear, John; Jackson, Ruth

2017-07-01

Although physicians are highly trained in the application of evidence-based medicine, and are assumed to make rational decisions, there is evidence that their decision making is prone to biases. One of the biases that has been shown to affect accuracy of judgements is that of representativeness and base-rate neglect, where the saliency of a person's features leads to overestimation of their likelihood of belonging to a group. This results in the substitution of 'subjective' probability for statistical probability. This study examines clinicians' propensity to make estimations of subjective probability when presented with clinical information that is considered typical of a medical condition. The strength of the representativeness bias is tested by presenting choices in textual and graphic form. Understanding of statistical probability is also tested by omitting all clinical information. For the questions that included clinical information, 46.7% and 45.5% of clinicians made judgements of statistical probability, respectively. Where the question omitted clinical information, 79.9% of clinicians made a judgement consistent with statistical probability. There was a statistically significant difference in responses to the questions with and without representativeness information (χ2 (1, n=254)=54.45, pprobability. One of the causes for this representativeness bias may be the way clinical medicine is taught where stereotypic presentations are emphasised in diagnostic decision making. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.

9. Race of source effects in the elaboration likelihood model.

Science.gov (United States)

White, P H; Harkins, S G

1994-11-01

In a series of experiments, we investigated the effect of race of source on persuasive communications in the Elaboration Likelihood Model (R.E. Petty & J.T. Cacioppo, 1981, 1986). In Experiment 1, we found no evidence that White participants responded to a Black source as a simple negative cue. Experiment 2 suggested the possibility that exposure to a Black source led to low-involvement message processing. In Experiments 3 and 4, a distraction paradigm was used to test this possibility, and it was found that participants under low involvement were highly motivated to process a message presented by a Black source. In Experiment 5, we found that attitudes toward the source's ethnic group, rather than violations of expectancies, accounted for this processing effect. Taken together, the results of these experiments are consistent with S.L. Gaertner and J.F. Dovidio's (1986) theory of aversive racism, which suggests that Whites, because of a combination of egalitarian values and underlying negative racial attitudes, are very concerned about not appearing unfavorable toward Blacks, leading them to be highly motivated to process messages presented by a source from this group.

10. Maximum likelihood estimation for cytogenetic dose-response curves

Energy Technology Data Exchange (ETDEWEB)

Frome, E.L; DuFrain, R.J.

1983-10-01

In vitro dose-response curves are used to describe the relation between the yield of dicentric chromosome aberrations and radiation dose for human lymphocytes. The dicentric yields follow the Poisson distribution, and the expected yield depends on both the magnitude and the temporal distribution of the dose for low LET radiation. A general dose-response model that describes this relation has been obtained by Kellerer and Rossi using the theory of dual radiation action. The yield of elementary lesions is kappa(..gamma..d + g(t, tau)d/sup 2/), where t is the time and d is dose. The coefficient of the d/sup 2/ term is determined by the recovery function and the temporal mode of irradiation. Two special cases of practical interest are split-dose and continuous exposure experiments, and the resulting models are intrinsically nonlinear in the parameters. A general purpose maximum likelihood estimation procedure is described and illustrated with numerical examples from both experimental designs. Poisson regression analysis is used for estimation, hypothesis testing, and regression diagnostics. Results are discussed in the context of exposure assessment procedures for both acute and chronic human radiation exposure.

11. Maximum likelihood approach for several stochastic volatility models

International Nuclear Information System (INIS)

Camprodon, Jordi; Perelló, Josep

2012-01-01

Volatility measures the amplitude of price fluctuations. Despite it being one of the most important quantities in finance, volatility is not directly observable. Here we apply a maximum likelihood method which assumes that price and volatility follow a two-dimensional diffusion process where volatility is the stochastic diffusion coefficient of the log-price dynamics. We apply this method to the simplest versions of the expOU, the OU and the Heston stochastic volatility models and we study their performance in terms of the log-price probability, the volatility probability, and its Mean First-Passage Time. The approach has some predictive power on the future returns amplitude by only knowing the current volatility. The assumed models do not consider long-range volatility autocorrelation and the asymmetric return-volatility cross-correlation but the method still yields very naturally these two important stylized facts. We apply the method to different market indices and with a good performance in all cases. (paper)

12. Some hematological parameters and the prognostic values of CD4, CD8 and total lymphocyte counts and CD4/CD8 cell count ratio in healthy HIV sero-negative, healthy HIV sero-positive and AIDS subjects in Port Harcourt, Nigeria

Directory of Open Access Journals (Sweden)

Blessing Didia

2008-12-01

Full Text Available OBJECTIVE: The present study attempts to determine normal values of CD4, CD8, CD4/CD8 ratio, total WBC and differential counts, hematocrit and total lymphocyte count (TLC in healthy HIV sero-negative and sero-positive subjects, and to assess the prognostic significance of these parameters in these subjects as compared to AIDS subjects.METHODS: A total of 300 subjects (147 M, 153 F aged between 17 and 71 years were recruited into the study. Subjects were separated according to sex and divided into three groups: Group A: healthy HIV sero-negative subjects; Group B: healthy HIV sero-positive newly diagnosed ART-naïve subjects; and Group C: AIDS subjects. CD4 and CD8 counts were determined by flow cytometry; hematocrit was determined using Hawksley micro-capillary tubes; total WBC and differential counts were determined manually with the improved Neubauer counting chamber; and TLC was obtained by multiplying the percentage of lymphocytes by the total WBC count.RESULTS: For male subjects, significant differences were found in CD4 count, CD4/CD8 count ratio, hematocrit, total WBC and TLC, whereas for female subjects, significant differences were found only in CD4 and CD4/CD8 count ratio in the three groups of subjects. In both sexes, however, these parameters were found to be highest in healthy HIV sero-negative subjects and lowest in AIDS subjects, with HIV sero-positive subjects having intermediate values. CONCLUSION: The results confirm previous reports that the CD4 count and CD4/CD8 count ratio are fairly reliable indicators of the progression of HIV infection. In addition, the results also apparently suggest that the prognostic value of CD8 count is limited and that of TLC possibly sex-dependent. The results could be of importance in our environment since previous reports have been relatively scarce.

13. VVER-1000 dominance ratio

International Nuclear Information System (INIS)

Gorodkov, S.

2009-01-01

Dominance ratio, or more precisely, its closeness to unity, is important characteristic of large reactor. It allows evaluate beforehand the number of source iterations required in deterministic calculations of power spatial distribution. Or the minimal number of histories to be modeled for achievement of statistical error level desired in large core Monte Carlo calculations. In this work relatively simple approach for dominance ratio evaluation is proposed. It essentially uses core symmetry. Dependence of dominance ratio on neutron flux spatial distribution is demonstrated. (author)

14. WWER-1000 dominance ratio

International Nuclear Information System (INIS)

Gorodkov, S.S.

2009-01-01

Dominance ratio, or more precisely, its closeness to unity, is important characteristic of large reactor. It allows evaluate beforehand the number of source iterations required in deterministic calculations of power spatial distribution. Or the minimal number of histories to be modeled for achievement of statistical error level desired in large core Monte Carlo calculations. In this work relatively simple approach for dominance ratio evaluation is proposed. It essentially uses core symmetry. Dependence of dominance ratio on neutron flux spatial distribution is demonstrated. (Authors)

15. Sharpening Sharpe Ratios

OpenAIRE

William N. Goetzmann; Jonathan E. Ingersoll Jr.; Matthew I. Spiegel; Ivo Welch

2002-01-01

It is now well known that the Sharpe ratio and other related reward-to-risk measures may be manipulated with option-like strategies. In this paper we derive the general conditions for achieving the maximum expected Sharpe ratio. We derive static rules for achieving the maximum Sharpe ratio with two or more options, as well as a continuum of derivative contracts. The optimal strategy has a truncated right tail and a fat left tail. We also derive dynamic rules for increasing the Sharpe ratio. O...

16. An Evaluation of Two Internal Surrogates for Determining the Three-Dimensional Position of Peripheral Lung Tumors

International Nuclear Information System (INIS)

Spoelstra, Femke; Soernsen de Koste, John R. van; Vincent, Andrew; Cuijpers, Johan P.; Slotman, Ben J.; Senan, Suresh

2009-01-01

Purpose: Both carina and diaphragm positions have been used as surrogates during respiratory-gated radiotherapy. We studied the correlation of both surrogates with three-dimensional (3D) tumor position. Methods and Materials: A total of 59 repeat artifact-free four-dimensional (4D) computed tomography (CT) scans, acquired during uncoached breathing, were identified in 23 patients with Stage I lung cancer. Repeat scans were co-registered to the initial 4D CT scan, and tumor, carina, and ipsilateral diaphragm were manually contoured in all phases of each 4D CT data set. Correlation between positions of carina and diaphragm with 3D tumor position was studied by use of log-likelihood ratio statistics. Models to predict 3D tumor position from internal surrogates at end inspiration (EI) and end expiration (EE) were developed, and model accuracy was tested by calculating SDs of differences between predicted and actual tumor positions. Results: Motion of both the carina and diaphragm significantly correlated with tumor motion, but log-likelihood ratios indicated that the carina was more predictive for tumor position. When craniocaudal tumor position was predicted by use of craniocaudal carina positions, the SDs of the differences between the predicted and observed positions were 2.2 mm and 2.4 mm at EI and EE, respectively. The corresponding SDs derived with the diaphragm positions were 3.7 mm and 3.9 mm at EI and EE, respectively. Prediction errors in the other directions were comparable. Prediction accuracy was similar at EI and EE. Conclusions: The carina is a better surrogate of 3D tumor position than diaphragm position. Because residual prediction errors were observed in this analysis, additional studies will be performed using audio-coached scans.

17. An omnibus likelihood test statistic and its factorization for change detection in time series of polarimetric SAR data

DEFF Research Database (Denmark)

Nielsen, Allan Aasbjerg; Conradsen, Knut; Skriver, Henning

2016-01-01

Based on an omnibus likelihood ratio test statistic for the equality of several variance-covariance matrices following the complex Wishart distribution with an associated p-value and a factorization of this test statistic, change analysis in a short sequence of multilook, polarimetric SAR data...... in the covariance matrix representation is carried out. The omnibus test statistic and its factorization detect if and when change(s) occur. The technique is demonstrated on airborne EMISAR L-band data but may be applied to Sentinel-1, Cosmo-SkyMed, TerraSAR-X, ALOS and RadarSat-2 or other dual- and quad...

18. Detecting isotopic ratio outliers

International Nuclear Information System (INIS)

Bayne, C.K.; Smith, D.H.

1985-01-01

An alternative method is proposed for improving isotopic ratio estimates. This method mathematically models pulse-count data and uses iterative reweighted Poisson regression to estimate model parameters to calculate the isotopic ratios. This computer-oriented approach provides theoretically better methods than conventional techniques to establish error limits and to identify outliers. 6 refs., 3 figs., 3 tabs

19. Detecting isotopic ratio outliers

Science.gov (United States)

Bayne, C. K.; Smith, D. H.

An alternative method is proposed for improving isotopic ratio estimates. This method mathematically models pulse-count data and uses iterative reweighted Poisson regression to estimate model parameters to calculate the isotopic ratios. This computer-oriented approach provides theoretically better methods than conventional techniques to establish error limits and to identify outliers.

20. Detecting isotopic ratio outliers

International Nuclear Information System (INIS)

Bayne, C.K.; Smith, D.H.

1986-01-01

An alternative method is proposed for improving isotopic ratio estimates. This method mathematically models pulse-count data and uses iterative reweighted Poisson regression to estimate model parameters to calculate the isotopic ratios. This computer-oriented approach provides theoretically better methods than conventional techniques to establish error limits and to identify outliers

1. Estimating likelihood of future crashes for crash-prone drivers

Directory of Open Access Journals (Sweden)

Subasish Das

2015-06-01

Full Text Available At-fault crash-prone drivers are usually considered as the high risk group for possible future incidents or crashes. In Louisiana, 34% of crashes are repeatedly committed by the at-fault crash-prone drivers who represent only 5% of the total licensed drivers in the state. This research has conducted an exploratory data analysis based on the driver faultiness and proneness. The objective of this study is to develop a crash prediction model to estimate the likelihood of future crashes for the at-fault drivers. The logistic regression method is used by employing eight years' traffic crash data (2004–2011 in Louisiana. Crash predictors such as the driver's crash involvement, crash and road characteristics, human factors, collision type, and environmental factors are considered in the model. The at-fault and not-at-fault status of the crashes are used as the response variable. The developed model has identified a few important variables, and is used to correctly classify at-fault crashes up to 62.40% with a specificity of 77.25%. This model can identify as many as 62.40% of the crash incidence of at-fault drivers in the upcoming year. Traffic agencies can use the model for monitoring the performance of an at-fault crash-prone drivers and making roadway improvements meant to reduce crash proneness. From the findings, it is recommended that crash-prone drivers should be targeted for special safety programs regularly through education and regulations.

2. New algorithms and methods to estimate maximum-likelihood phylogenies: assessing the performance of PhyML 3.0.

Science.gov (United States)

Guindon, Stéphane; Dufayard, Jean-François; Lefort, Vincent; Anisimova, Maria; Hordijk, Wim; Gascuel, Olivier

2010-05-01

PhyML is a phylogeny software based on the maximum-likelihood principle. Early PhyML versions used a fast algorithm performing nearest neighbor interchanges to improve a reasonable starting tree topology. Since the original publication (Guindon S., Gascuel O. 2003. A simple, fast and accurate algorithm to estimate large phylogenies by maximum likelihood. Syst. Biol. 52:696-704), PhyML has been widely used (>2500 citations in ISI Web of Science) because of its simplicity and a fair compromise between accuracy and speed. In the meantime, research around PhyML has continued, and this article describes the new algorithms and methods implemented in the program. First, we introduce a new algorithm to search the tree space with user-defined intensity using subtree pruning and regrafting topological moves. The parsimony criterion is used here to filter out the least promising topology modifications with respect to the likelihood function. The analysis of a large collection of real nucleotide and amino acid data sets of various sizes demonstrates the good performance of this method. Second, we describe a new test to assess the support of the data for internal branches of a phylogeny. This approach extends the recently proposed approximate likelihood-ratio test and relies on a nonparametric, Shimodaira-Hasegawa-like procedure. A detailed analysis of real alignments sheds light on the links between this new approach and the more classical nonparametric bootstrap method. Overall, our tests show that the last version (3.0) of PhyML is fast, accurate, stable, and ready to use. A Web server and binary files are available from http://www.atgc-montpellier.fr/phyml/.

3. FINANCIAL RATIOS – REVEAL HOW A BUSINESS IS DOING?

OpenAIRE

MIRELA MONEA

2009-01-01

The paper aims to present the main financial ratios which provide a picture about company’s profitability, its financial position, use of its assets efficiency, its long-term debt financing. Discussion is focused on: profitability ratios, short-term financial ratios, activity ratios, long-term debt ratios or dividend policy ratios. Also, will try to answer at the following main questions: What financial ratios analysis tells us? What the users of these needs to know?

4. Phylogeny, ecology, and heart position in snakes.

Science.gov (United States)

Gartner, Gabriel E A; Hicks, James W; Manzani, Paulo R; Andrade, Denis V; Abe, Augusto S; Wang, Tobias; Secor, Stephen M; Garland, Theodore

2010-01-01

The cardiovascular system of all animals is affected by gravitational pressure gradients, the intensity of which varies according to organismic features, behavior, and habitat occupied. A previous nonphylogenetic analysis of heart position in snakes-which often assume vertical postures-found the heart located 15%-25% of total body length from the head in terrestrial and arboreal species but 25%-45% in aquatic species. It was hypothesized that a more anterior heart in arboreal species served to reduce the hydrostatic blood pressure when these animals adopt vertical postures during climbing, whereas an anterior heart position would not be needed in aquatic habitats, where the effects of gravity are less pronounced. We analyzed a new data set of 155 species from five major families of Alethinophidia (one of the two major branches of snakes, the other being blind snakes, Scolecophidia) using both conventional and phylogenetically based statistical methods. General linear models regressing log(10) snout-heart position on log(10) snout-vent length (SVL), as well as dummy variables coding for habitat and/or clade, were compared using likelihood ratio tests and the Akaike Information Criterion. Heart distance to the tip of the snout scaled isometrically with SVL. In all instances, phylogenetic models that incorporated transformation of the branch lengths under an Ornstein-Uhlenbeck model of evolution (to mimic stabilizing selection) better fit the data as compared with their nonphylogenetic counterparts. The best-fit model predicting snake heart position included aspects of both habitat and clade and indicated that arboreal snakes in our study tend to have hearts placed more posteriorly, opposite the trend identified in previous studies. Phylogenetic signal in relative heart position was apparent both within and among clades. Our results suggest that overcoming gravitational pressure gradients in snakes most likely involves the combined action of several cardiovascular and

5. DREAM3: network inference using dynamic context likelihood of relatedness and the inferelator.

Directory of Open Access Journals (Sweden)

2010-03-01

Full Text Available Many current works aiming to learn regulatory networks from systems biology data must balance model complexity with respect to data availability and quality. Methods that learn regulatory associations based on unit-less metrics, such as Mutual Information, are attractive in that they scale well and reduce the number of free parameters (model complexity per interaction to a minimum. In contrast, methods for learning regulatory networks based on explicit dynamical models are more complex and scale less gracefully, but are attractive as they may allow direct prediction of transcriptional dynamics and resolve the directionality of many regulatory interactions.We aim to investigate whether scalable information based methods (like the Context Likelihood of Relatedness method and more explicit dynamical models (like Inferelator 1.0 prove synergistic when combined. We test a pipeline where a novel modification of the Context Likelihood of Relatedness (mixed-CLR, modified to use time series data is first used to define likely regulatory interactions and then Inferelator 1.0 is used for final model selection and to build an explicit dynamical model.Our method ranked 2nd out of 22 in the DREAM3 100-gene in silico networks challenge. Mixed-CLR and Inferelator 1.0 are complementary, demonstrating a large performance gain relative to any single tested method, with precision being especially high at low recall values. Partitioning the provided data set into four groups (knock-down, knock-out, time-series, and combined revealed that using comprehensive knock-out data alone provides optimal performance. Inferelator 1.0 proved particularly powerful at resolving the directionality of regulatory interactions, i.e. "who regulates who" (approximately of identified true positives were correctly resolved. Performance drops for high in-degree genes, i.e. as the number of regulators per target gene increases, but not with out-degree, i.e. performance is not affected by

6. High-order Composite Likelihood Inference for Max-Stable Distributions and Processes

KAUST Repository

Castruccio, Stefano; Huser, Raphaë l; Genton, Marc G.

2015-01-01

In multivariate or spatial extremes, inference for max-stable processes observed at a large collection of locations is a very challenging problem in computational statistics, and current approaches typically rely on less expensive composite likelihoods constructed from small subsets of data. In this work, we explore the limits of modern state-of-the-art computational facilities to perform full likelihood inference and to efficiently evaluate high-order composite likelihoods. With extensive simulations, we assess the loss of information of composite likelihood estimators with respect to a full likelihood approach for some widely-used multivariate or spatial extreme models, we discuss how to choose composite likelihood truncation to improve the efficiency, and we also provide recommendations for practitioners. This article has supplementary material online.

7. High-order Composite Likelihood Inference for Max-Stable Distributions and Processes

KAUST Repository

Castruccio, Stefano

2015-09-29

In multivariate or spatial extremes, inference for max-stable processes observed at a large collection of locations is a very challenging problem in computational statistics, and current approaches typically rely on less expensive composite likelihoods constructed from small subsets of data. In this work, we explore the limits of modern state-of-the-art computational facilities to perform full likelihood inference and to efficiently evaluate high-order composite likelihoods. With extensive simulations, we assess the loss of information of composite likelihood estimators with respect to a full likelihood approach for some widely-used multivariate or spatial extreme models, we discuss how to choose composite likelihood truncation to improve the efficiency, and we also provide recommendations for practitioners. This article has supplementary material online.

8. Supplementary Material for: High-Order Composite Likelihood Inference for Max-Stable Distributions and Processes

KAUST Repository

Castruccio, Stefano; Huser, Raphaë l; Genton, Marc G.

2016-01-01

In multivariate or spatial extremes, inference for max-stable processes observed at a large collection of points is a very challenging problem and current approaches typically rely on less expensive composite likelihoods constructed from small subsets of data. In this work, we explore the limits of modern state-of-the-art computational facilities to perform full likelihood inference and to efficiently evaluate high-order composite likelihoods. With extensive simulations, we assess the loss of information of composite likelihood estimators with respect to a full likelihood approach for some widely used multivariate or spatial extreme models, we discuss how to choose composite likelihood truncation to improve the efficiency, and we also provide recommendations for practitioners. This article has supplementary material online.

9. Charge ratio of muons from atmospheric neutrinos

Energy Technology Data Exchange (ETDEWEB)

Gaisser, T.K.; Stanev, Todor

2003-05-22

We calculate the intensities and angular distributions of positive and negative muons produced by atmospheric neutrinos. We comment on some sources of uncertainty in the charge ratio. We also draw attention to a potentially interesting signature of neutrino oscillations in the muon charge ratio, and we discuss the prospects for its observation (which are not quite within the reach of currently planned magnetized detectors)

10. Maximum Likelihood Estimation and Inference With Examples in R, SAS and ADMB

CERN Document Server

Millar, Russell B

2011-01-01

This book takes a fresh look at the popular and well-established method of maximum likelihood for statistical estimation and inference. It begins with an intuitive introduction to the concepts and background of likelihood, and moves through to the latest developments in maximum likelihood methodology, including general latent variable models and new material for the practical implementation of integrated likelihood using the free ADMB software. Fundamental issues of statistical inference are also examined, with a presentation of some of the philosophical debates underlying the choice of statis

11. Difference and ratio plots

DEFF Research Database (Denmark)

Svendsen, Anders Jørgen; Holmskov, U; Bro, Peter

1995-01-01

and systemic lupus erythematosus from another previously published study (Macanovic, M. and Lachmann, P.J. (1979) Clin. Exp. Immunol. 38, 274) are also represented using ratio plots. Our observations indicate that analysis by regression analysis may often be misleading....... hitherto unnoted differences between controls and patients with either rheumatoid arthritis or systemic lupus erythematosus. For this we use simple, but unconventional, graphic representations of the data, based on difference plots and ratio plots. Differences between patients with Burkitt's lymphoma...

12. Fast Maximum-Likelihood Decoder for Quasi-Orthogonal Space-Time Block Code

Directory of Open Access Journals (Sweden)

2015-01-01

Full Text Available Motivated by the decompositions of sphere and QR-based methods, in this paper we present an extremely fast maximum-likelihood (ML detection approach for quasi-orthogonal space-time block code (QOSTBC. The proposed algorithm with a relatively simple design exploits structure of quadrature amplitude modulation (QAM constellations to achieve its goal and can be extended to any arbitrary constellation. Our decoder utilizes a new decomposition technique for ML metric which divides the metric into independent positive parts and a positive interference part. Search spaces of symbols are substantially reduced by employing the independent parts and statistics of noise. Symbols within the search spaces are successively evaluated until the metric is minimized. Simulation results confirm that the proposed decoder’s performance is superior to many of the recently published state-of-the-art solutions in terms of complexity level. More specifically, it was possible to verify that application of the new algorithms with 1024-QAM would decrease the computational complexity compared to state-of-the-art solution with 16-QAM.

13. pplacer: linear time maximum-likelihood and Bayesian phylogenetic placement of sequences onto a fixed reference tree

Directory of Open Access Journals (Sweden)

Kodner Robin B

2010-10-01

Full Text Available Abstract Background Likelihood-based phylogenetic inference is generally considered to be the most reliable classification method for unknown sequences. However, traditional likelihood-based phylogenetic methods cannot be applied to large volumes of short reads from next-generation sequencing due to computational complexity issues and lack of phylogenetic signal. "Phylogenetic placement," where a reference tree is fixed and the unknown query sequences are placed onto the tree via a reference alignment, is a way to bring the inferential power offered by likelihood-based approaches to large data sets. Results This paper introduces pplacer, a software package for phylogenetic placement and subsequent visualization. The algorithm can place twenty thousand short reads on a reference tree of one thousand taxa per hour per processor, has essentially linear time and memory complexity in the number of reference taxa, and is easy to run in parallel. Pplacer features calculation of the posterior probability of a placement on an edge, which is a statistically rigorous way of quantifying uncertainty on an edge-by-edge basis. It also can inform the user of the positional uncertainty for query sequences by calculating expected distance between placement locations, which is crucial in the estimation of uncertainty with a well-sampled reference tree. The software provides visualizations using branch thickness and color to represent number of placements and their uncertainty. A simulation study using reads generated from 631 COG alignments shows a high level of accuracy for phylogenetic placement over a wide range of alignment diversity, and the power of edge uncertainty estimates to measure placement confidence. Conclusions Pplacer enables efficient phylogenetic placement and subsequent visualization, making likelihood-based phylogenetics methodology practical for large collections of reads; it is freely available as source code, binaries, and a web service.

14. Predicting Teacher Likelihood to Use School Gardens: A Case Study

Science.gov (United States)

Kincy, Natalie; Fuhrman, Nicholas E.; Navarro, Maria; Knauft, David

2016-01-01

A quantitative survey, built around the theory of planned behavior, was used to investigate elementary teachers' attitudes, school norms, perceived behavioral control, and intent in both current and ideal teaching situations toward using gardens in their curriculum. With positive school norms and teachers who garden in their personal time, 77% of…

15. Likelihood-Based Clustering of Meta-Analytic SROC Curves

Science.gov (United States)

Holling, Heinz; Bohning, Walailuck; Bohning, Dankmar

2012-01-01

Meta-analysis of diagnostic studies experience the common problem that different studies might not be comparable since they have been using a different cut-off value for the continuous or ordered categorical diagnostic test value defining different regions for which the diagnostic test is defined to be positive. Hence specificities and…

16. Relationship Formation and Stability in Emerging Adulthood: Do Sex Ratios Matter?

Science.gov (United States)

Warner, Tara D.; Manning, Wendy D.; Giordano, Peggy C.; Longmore, Monica A.

2011-01-01

Research links sex ratios with the likelihood of marriage and divorce. However, whether sex ratios similarly influence precursors to marriage (transitions in and out of dating or cohabiting relationships) is unknown. Utilizing data from the Toledo Adolescent Relationships Study and the 2000 U.S. Census, this study assesses whether sex ratios…

17. Theoretical Study of Penalized-Likelihood Image Reconstruction for Region of Interest Quantification

International Nuclear Information System (INIS)

Qi, Jinyi; Huesman, Ronald H.

2006-01-01

Region of interest (ROI) quantification is an important task in emission tomography (e.g., positron emission tomography and single photon emission computed tomography). It is essential for exploring clinical factors such as tumor activity, growth rate, and the efficacy of therapeutic interventions. Statistical image reconstruction methods based on the penalized maximum-likelihood (PML) or maximum a posteriori principle have been developed for emission tomography to deal with the low signal-to-noise ratio of the emission data. Similar to the filter cut-off frequency in the filtered backprojection method, the regularization parameter in PML reconstruction controls the resolution and noise tradeoff and, hence, affects ROI quantification. In this paper, we theoretically analyze the performance of ROI quantification in PML reconstructions. Building on previous work, we derive simplified theoretical expressions for the bias, variance, and ensemble mean-squared-error (EMSE) of the estimated total activity in an ROI that is surrounded by a uniform background. When the mean and covariance matrix of the activity inside the ROI are known, the theoretical expressions are readily computable and allow for fast evaluation of image quality for ROI quantification with different regularization parameters. The optimum regularization parameter can then be selected to minimize the EMSE. Computer simulations are conducted for small ROIs with variable uniform uptake. The results show that the theoretical predictions match the Monte Carlo results reasonably well

18. Maximum likelihood estimation-based denoising of magnetic resonance images using restricted local neighborhoods

International Nuclear Information System (INIS)

Rajan, Jeny; Jeurissen, Ben; Sijbers, Jan; Verhoye, Marleen; Van Audekerke, Johan

2011-01-01

In this paper, we propose a method to denoise magnitude magnetic resonance (MR) images, which are Rician distributed. Conventionally, maximum likelihood methods incorporate the Rice distribution to estimate the true, underlying signal from a local neighborhood within which the signal is assumed to be constant. However, if this assumption is not met, such filtering will lead to blurred edges and loss of fine structures. As a solution to this problem, we put forward the concept of restricted local neighborhoods where the true intensity for each noisy pixel is estimated from a set of preselected neighboring pixels. To this end, a reference image is created from the noisy image using a recently proposed nonlocal means algorithm. This reference image is used as a prior for further noise reduction. A scheme is developed to locally select an appropriate subset of pixels from which the underlying signal is estimated. Experimental results based on the peak signal to noise ratio, structural similarity index matrix, Bhattacharyya coefficient and mean absolute difference from synthetic and real MR images demonstrate the superior performance of the proposed method over other state-of-the-art methods.

19. Meta-Analysis and Systematic Review to Assess the Role of Soluble FMS-Like Tyrosine Kinase-1 and Placenta Growth Factor Ratio in Prediction of Preeclampsia: The SaPPPhirE Study.

Science.gov (United States)

Agrawal, Swati; Cerdeira, Ana Sofia; Redman, Christopher; Vatish, Manu

2018-02-01

Preeclampsia is a major cause of morbidity and mortality worldwide. Numerous candidate biomarkers have been proposed for diagnosis and prediction of preeclampsia. Measurement of maternal circulating angiogenesis biomarker as the ratio of sFlt-1 (soluble FMS-like tyrosine kinase-1; an antiangiogenic factor)/PlGF (placental growth factor; an angiogenic factor) reflects the antiangiogenic balance that characterizes incipient or overt preeclampsia. The ratio increases before the onset of the disease and thus may help in predicting preeclampsia. We conducted a meta-analysis to explore the predictive accuracy of sFlt-1/PlGF ratio in preeclampsia. We included 15 studies with 534 cases with preeclampsia and 19 587 controls. The ratio has a pooled sensitivity of 80% (95% confidence interval, 0.68-0.88), specificity of 92% (95% confidence interval, 0.87-0.96), positive likelihood ratio of 10.5 (95% confidence interval, 6.2-18.0), and a negative likelihood ratio of 0.22 (95% confidence interval, 0.13-0.35) in predicting preeclampsia in both high- and low-risk patients. Most of the studies have not made a distinction between early- and late-onset disease, and therefore, the analysis for it could not be done. It can prove to be a valuable screening tool for preeclampsia and may also help in decision-making, treatment stratification, and better resource allocation. © 2017 American Heart Association, Inc.

20. Bias Correction for the Maximum Likelihood Estimate of Ability. Research Report. ETS RR-05-15

Science.gov (United States)

Zhang, Jinming

2005-01-01

Lord's bias function and the weighted likelihood estimation method are effective in reducing the bias of the maximum likelihood estimate of an examinee's ability under the assumption that the true item parameters are known. This paper presents simulation studies to determine the effectiveness of these two methods in reducing the bias when the item…

1. Bias correction in the hierarchical likelihood approach to the analysis of multivariate survival data.

Science.gov (United States)

Jeon, Jihyoun; Hsu, Li; Gorfine, Malka

2012-07-01

Frailty models are useful for measuring unobserved heterogeneity in risk of failures across clusters, providing cluster-specific risk prediction. In a frailty model, the latent frailties shared by members within a cluster are assumed to act multiplicatively on the hazard function. In order to obtain parameter and frailty variate estimates, we consider the hierarchical likelihood (H-likelihood) approach (Ha, Lee and Song, 2001. Hierarchical-likelihood approach for frailty models. Biometrika 88, 233-243) in which the latent frailties are treated as "parameters" and estimated jointly with other parameters of interest. We find that the H-likelihood estimators perform well when the censoring rate is low, however, they are substantially biased when the censoring rate is moderate to high. In this paper, we propose a simple and easy-to-implement bias correction method for the H-likelihood estimators under a shared frailty model. We also extend the method to a multivariate frailty model, which incorporates complex dependence structure within clusters. We conduct an extensive simulation study and show that the proposed approach performs very well for censoring rates as high as 80%. We also illustrate the method with a breast cancer data set. Since the H-likelihood is the same as the penalized likelihood function, the proposed bias correction method is also applicable to the penalized likelihood estimators.

2. Analyzing multivariate survival data using composite likelihood and flexible parametric modeling of the hazard functions

DEFF Research Database (Denmark)

Nielsen, Jan; Parner, Erik

2010-01-01

In this paper, we model multivariate time-to-event data by composite likelihood of pairwise frailty likelihoods and marginal hazards using natural cubic splines. Both right- and interval-censored data are considered. The suggested approach is applied on two types of family studies using the gamma...

3. Existence and uniqueness of the maximum likelihood estimator for models with a Kronecker product covariance structure

NARCIS (Netherlands)

Ros, B.P.; Bijma, F.; de Munck, J.C.; de Gunst, M.C.M.

2016-01-01

This paper deals with multivariate Gaussian models for which the covariance matrix is a Kronecker product of two matrices. We consider maximum likelihood estimation of the model parameters, in particular of the covariance matrix. There is no explicit expression for the maximum likelihood estimator

4. Likelihood functions for the analysis of single-molecule binned photon sequences

Energy Technology Data Exchange (ETDEWEB)

Gopich, Irina V., E-mail: irinag@niddk.nih.gov [Laboratory of Chemical Physics, National Institute of Diabetes and Digestive and Kidney Diseases, National Institutes of Health, Bethesda, MD 20892 (United States)

2012-03-02

Graphical abstract: Folding of a protein with attached fluorescent dyes, the underlying conformational trajectory of interest, and the observed binned photon trajectory. Highlights: Black-Right-Pointing-Pointer A sequence of photon counts can be analyzed using a likelihood function. Black-Right-Pointing-Pointer The exact likelihood function for a two-state kinetic model is provided. Black-Right-Pointing-Pointer Several approximations are considered for an arbitrary kinetic model. Black-Right-Pointing-Pointer Improved likelihood functions are obtained to treat sequences of FRET efficiencies. - Abstract: We consider the analysis of a class of experiments in which the number of photons in consecutive time intervals is recorded. Sequence of photon counts or, alternatively, of FRET efficiencies can be studied using likelihood-based methods. For a kinetic model of the conformational dynamics and state-dependent Poisson photon statistics, the formalism to calculate the exact likelihood that this model describes such sequences of photons or FRET efficiencies is developed. Explicit analytic expressions for the likelihood function for a two-state kinetic model are provided. The important special case when conformational dynamics are so slow that at most a single transition occurs in a time bin is considered. By making a series of approximations, we eventually recover the likelihood function used in hidden Markov models. In this way, not only is insight gained into the range of validity of this procedure, but also an improved likelihood function can be obtained.

5. Use of deterministic sampling for exploring likelihoods in linkage analysis for quantitative traits.

NARCIS (Netherlands)

Mackinnon, M.J.; Beek, van der S.; Kinghorn, B.P.

1996-01-01

Deterministic sampling was used to numerically evaluate the expected log-likelihood surfaces of QTL-marker linkage models in large pedigrees with simple structures. By calculating the expected values of likelihoods, questions of power of experimental designs, bias in parameter estimates, approximate

6. Predictors of Self-Reported Likelihood of Working with Older Adults

Science.gov (United States)

Eshbaugh, Elaine M.; Gross, Patricia E.; Satrom, Tatum

2010-01-01

This study examined the self-reported likelihood of working with older adults in a future career among 237 college undergraduates at a midsized Midwestern university. Although aging anxiety was not significantly related to likelihood of working with older adults, those students who had a greater level of death anxiety were less likely than other…

7. Organizational Justice and Men's Likelihood to Sexually Harass: The Moderating Role of Sexism and Personality

Science.gov (United States)

Krings, Franciska; Facchin, Stephanie

2009-01-01

This study demonstrated relations between men's perceptions of organizational justice and increased sexual harassment proclivities. Respondents reported higher likelihood to sexually harass under conditions of low interactional justice, suggesting that sexual harassment likelihood may increase as a response to perceived injustice. Moreover, the…

8. Pulmonary arterial hypertension in children: diagnosis using ratio of main pulmonary artery to ascending aorta diameter as determined by multi-detector computed tomography.

Science.gov (United States)

Caro-Domínguez, Pablo; Compton, Gregory; Humpl, Tilman; Manson, David E

2016-09-01

The ratio of the transverse diameter of the main pulmonary artery (MPA) to ascending aorta as determined at multi-detector CT is a tool that can be used to assess the pulmonary arterial size in cases of pulmonary arterial hypertension in children. To establish a ratio of MPA to ascending aorta diameter using multi-detector CT imaging suggestive of pulmonary arterial hypertension in children. We hypothesize that a defined ratio of MPA to ascending aorta is identifiable on multi-detector CT and that higher ratios can be used to reliably diagnose the presence of pulmonary arterial hypertension in children. We calculated the multi-detector CT ratio of MPA to ascending aorta diameter in 44 children with documented pulmonary arterial hypertension by right heart catheterization and in 44 age- and gender-matched control children with no predisposing factors for pulmonary arterial hypertension. We compared this multi-detector-CT-determined ratio with the MPA pressure in the study group, as well as with the ratio of MPA to ascending aorta in the control group. A threshold ratio value was calculated to accurately identify children with pulmonary arterial hypertension. Children with documented primary pulmonary arterial hypertension have a significantly higher ratio of MPA to ascending aorta (1.46) than children without pulmonary arterial hypertension (1.11). A ratio of 1.3 carries a positive likelihood of 34 and a positive predictive value of 97% for the diagnosis of pulmonary arterial hypertension. The pulmonary arteries were larger in children with pulmonary arterial hypertension than in a control group of normal children. A CT-measured ratio of MPA to ascending aorta of 1.3 should raise the suspicion of pulmonary arterial hypertension in children.

9. Pulmonary arterial hypertension in children: diagnosis using ratio of main pulmonary artery to ascending aorta diameter as determined by multi-detector computed tomography

International Nuclear Information System (INIS)

Caro-Dominguez, Pablo; Manson, David E.; Compton, Gregory; Humpl, Tilman

2016-01-01

The ratio of the transverse diameter of the main pulmonary artery (MPA) to ascending aorta as determined at multi-detector CT is a tool that can be used to assess the pulmonary arterial size in cases of pulmonary arterial hypertension in children. To establish a ratio of MPA to ascending aorta diameter using multi-detector CT imaging suggestive of pulmonary arterial hypertension in children. We hypothesize that a defined ratio of MPA to ascending aorta is identifiable on multi-detector CT and that higher ratios can be used to reliably diagnose the presence of pulmonary arterial hypertension in children. We calculated the multi-detector CT ratio of MPA to ascending aorta diameter in 44 children with documented pulmonary arterial hypertension by right heart catheterization and in 44 age- and gender-matched control children with no predisposing factors for pulmonary arterial hypertension. We compared this multi-detector-CT-determined ratio with the MPA pressure in the study group, as well as with the ratio of MPA to ascending aorta in the control group. A threshold ratio value was calculated to accurately identify children with pulmonary arterial hypertension. Children with documented primary pulmonary arterial hypertension have a significantly higher ratio of MPA to ascending aorta (1.46) than children without pulmonary arterial hypertension (1.11). A ratio of 1.3 carries a positive likelihood of 34 and a positive predictive value of 97% for the diagnosis of pulmonary arterial hypertension. The pulmonary arteries were larger in children with pulmonary arterial hypertension than in a control group of normal children. A CT-measured ratio of MPA to ascending aorta of 1.3 should raise the suspicion of pulmonary arterial hypertension in children. (orig.)

10. Pulmonary arterial hypertension in children: diagnosis using ratio of main pulmonary artery to ascending aorta diameter as determined by multi-detector computed tomography

Energy Technology Data Exchange (ETDEWEB)

Caro-Dominguez, Pablo; Manson, David E. [University of Toronto, Department of Diagnostic Imaging, The Hospital for Sick Children, Department of Medical Imaging, Toronto, ON (Canada); Compton, Gregory [University of Toronto, Department of Diagnostic Imaging, The Hospital for Sick Children, Department of Medical Imaging, Toronto, ON (Canada); Epworth Hospital, Epworth Medical Imaging, Richmond, VIC (Australia); Humpl, Tilman [University of Toronto, Division of Cardiology, Department of Pediatrics, The Hospital for Sick Children, Toronto, ON (Canada)

2016-09-15

The ratio of the transverse diameter of the main pulmonary artery (MPA) to ascending aorta as determined at multi-detector CT is a tool that can be used to assess the pulmonary arterial size in cases of pulmonary arterial hypertension in children. To establish a ratio of MPA to ascending aorta diameter using multi-detector CT imaging suggestive of pulmonary arterial hypertension in children. We hypothesize that a defined ratio of MPA to ascending aorta is identifiable on multi-detector CT and that higher ratios can be used to reliably diagnose the presence of pulmonary arterial hypertension in children. We calculated the multi-detector CT ratio of MPA to ascending aorta diameter in 44 children with documented pulmonary arterial hypertension by right heart catheterization and in 44 age- and gender-matched control children with no predisposing factors for pulmonary arterial hypertension. We compared this multi-detector-CT-determined ratio with the MPA pressure in the study group, as well as with the ratio of MPA to ascending aorta in the control group. A threshold ratio value was calculated to accurately identify children with pulmonary arterial hypertension. Children with documented primary pulmonary arterial hypertension have a significantly higher ratio of MPA to ascending aorta (1.46) than children without pulmonary arterial hypertension (1.11). A ratio of 1.3 carries a positive likelihood of 34 and a positive predictive value of 97% for the diagnosis of pulmonary arterial hypertension. The pulmonary arteries were larger in children with pulmonary arterial hypertension than in a control group of normal children. A CT-measured ratio of MPA to ascending aorta of 1.3 should raise the suspicion of pulmonary arterial hypertension in children. (orig.)

11. Statistical modelling of survival data with random effects h-likelihood approach

CERN Document Server

Ha, Il Do; Lee, Youngjo

2017-01-01

This book provides a groundbreaking introduction to the likelihood inference for correlated survival data via the hierarchical (or h-) likelihood in order to obtain the (marginal) likelihood and to address the computational difficulties in inferences and extensions. The approach presented in the book overcomes shortcomings in the traditional likelihood-based methods for clustered survival data such as intractable integration. The text includes technical materials such as derivations and proofs in each chapter, as well as recently developed software programs in R (“frailtyHL”), while the real-world data examples together with an R package, “frailtyHL” in CRAN, provide readers with useful hands-on tools. Reviewing new developments since the introduction of the h-likelihood to survival analysis (methods for interval estimation of the individual frailty and for variable selection of the fixed effects in the general class of frailty models) and guiding future directions, the book is of interest to research...

12. The likelihood principle and its proof – a never-ending story…

DEFF Research Database (Denmark)

Jørgensen, Thomas Martini

2015-01-01

An ongoing controversy in philosophy of statistics is the so-called “likelihood principle” essentially stating that all evidence which is obtained from an experiment about an unknown quantity θ is contained in the likelihood function of θ. Common classical statistical methodology, such as the use...... of significance tests, and confidence intervals, depends on the experimental procedure and unrealized events and thus violates the likelihood principle. The likelihood principle was identified by that name and proved in a famous paper by Allan Birnbaum in 1962. However, ever since both the principle itself...... as well as the proof has been highly debated. This presentation will illustrate the debate of both the principle and its proof, from 1962 and up to today. An often-used experiment to illustrate the controversy between classical interpretation and evidential confirmation based on the likelihood principle...

13. Low Birthweight Increases the Likelihood of Severe Steatosis in Pediatric Non-Alcoholic Fatty Liver Disease.

Science.gov (United States)

Bugianesi, Elisabetta; Bizzarri, Carla; Rosso, Chiara; Mosca, Antonella; Panera, Nadia; Veraldi, Silvio; Dotta, Andrea; Giannone, Germana; Raponi, Massimiliano; Cappa, Marco; Alisi, Anna; Nobili, Valerio

2017-08-01

Small for gestational age (SGA) is associated with an increased risk of non-alcoholic fatty liver disease (NAFLD). Our aim was to investigate the correlation of birthweight with the severity of liver damage in a large cohort of children with NAFLD. Two hundred and eighty-eight consecutive Caucasian Italian overweight/obese children with biopsy-proven NAFLD were included in the study. We examined the relative association of each histological feature of NAFLD with metabolic alterations, insulin-resistance, I148M polymorphism in the patatin-like phospholipase domain-containing protein 3 (PNPLA3) gene, and birthweight relative to gestational age. In the whole NAFLD cohort, 12.2% of patients were SGA, 62.8% appropriate for gestational age (AGA), and 25% large for gestational age (LGA). SGA children had a higher prevalence of severe steatosis (69%) and severe portal inflammation (14%) compared with the AGA and LGA groups. Notably, severe steatosis (>66%) was decreasing from SGA to AGA and LGA, whereas the prevalence of moderate steatosis (33-66%) was similar in three groups. The prevalence of type 1 NAFLD is higher in the LGA group with respect to the other two groups (25% vs.5.2% vs.9.4%), whereas the SGA group shows a higher prevalence of overlap type (85.8%) with respect to the LGA group (51.4%) but not compared with the AGA group (75%). At multivariable regression analysis, SGA at birth increased fourfold the likelihood of severe steatosis (odds ratio (OR) 4.0, 95% confidence interval (CI) 1.43-10.9, P=0.008) and threefold the likelihood of NAFLD Activity Score (NAS)≥5 (OR 2.98, 95% CI 1.06-8.33, P=0.037) independently of homeostasis model assessment of insulin resistance and PNPLA3 genotype. The PNPLA3-CC wild-type genotype was the strongest independent predictor of the absence of significant fibrosis (OR 0.26, 95% CI 0.13-0.52, P=<0.001). In children with NAFLD, the risk of severe steatosis is increased by SGA at birth, independent of and in addition to other

14. The rectilinear Steiner ratio

Directory of Open Access Journals (Sweden)

PO de Wet

2005-06-01

Full Text Available The rectilinear Steiner ratio was shown to be 3/2 by Hwang [Hwang FK, 1976, On Steiner minimal trees with rectilinear distance, SIAM Journal on Applied Mathematics, 30, pp. 104– 114.]. We use continuity and introduce restricted point sets to obtain an alternative, short and self-contained proof of this result.

15. Sampling of systematic errors to estimate likelihood weights in nuclear data uncertainty propagation

International Nuclear Information System (INIS)

Helgesson, P.; Sjöstrand, H.; Koning, A.J.; Rydén, J.; Rochman, D.; Alhassan, E.; Pomp, S.

2016-01-01

In methodologies for nuclear data (ND) uncertainty assessment and propagation based on random sampling, likelihood weights can be used to infer experimental information into the distributions for the ND. As the included number of correlated experimental points grows large, the computational time for the matrix inversion involved in obtaining the likelihood can become a practical problem. There are also other problems related to the conventional computation of the likelihood, e.g., the assumption that all experimental uncertainties are Gaussian. In this study, a way to estimate the likelihood which avoids matrix inversion is investigated; instead, the experimental correlations are included by sampling of systematic errors. It is shown that the model underlying the sampling methodology (using univariate normal distributions for random and systematic errors) implies a multivariate Gaussian for the experimental points (i.e., the conventional model). It is also shown that the likelihood estimates obtained through sampling of systematic errors approach the likelihood obtained with matrix inversion as the sample size for the systematic errors grows large. In studied practical cases, it is seen that the estimates for the likelihood weights converge impractically slowly with the sample size, compared to matrix inversion. The computational time is estimated to be greater than for matrix inversion in cases with more experimental points, too. Hence, the sampling of systematic errors has little potential to compete with matrix inversion in cases where the latter is applicable. Nevertheless, the underlying model and the likelihood estimates can be easier to intuitively interpret than the conventional model and the likelihood function involving the inverted covariance matrix. Therefore, this work can both have pedagogical value and be used to help motivating the conventional assumption of a multivariate Gaussian for experimental data. The sampling of systematic errors could also

16. Position Information

Data.gov (United States)

Social Security Administration — The Position Information Data Asset provides the ability to search for active SSA position descriptions using various search criteria. An individual may search by PD...

17. Positive Psychology

Science.gov (United States)

Peterson, Christopher

2009-01-01

Positive psychology is a deliberate correction to the focus of psychology on problems. Positive psychology does not deny the difficulties that people may experience but does suggest that sole attention to disorder leads to an incomplete view of the human condition. Positive psychologists concern themselves with four major topics: (1) positive…

18. Chocolate consumption in pregnancy and reduced likelihood of preeclampsia.

Science.gov (United States)

Triche, Elizabeth W; Grosso, Laura M; Belanger, Kathleen; Darefsky, Amy S; Benowitz, Neal L; Bracken, Michael B

2008-05-01

Preeclampsia is a major pregnancy complication with cardiovascular manifestations. Recent studies suggest that chocolate consumption may benefit cardiovascular health. We studied the association of chocolate consumption with risk of preeclampsia in a prospective cohort study of 2291 pregnant women who delivered a singleton livebirth between September 1996 and January 2000. Chocolate consumption was measured by self report in the first and third trimesters, and by umbilical cord serum concentrations of theobromine, the major methylxanthine component of chocolate. Preeclampsia was assessed by detailed medical record review for 1943 of the women. We derived adjusted odds ratios (aOR) and 95% confidence intervals (CIs) from logistic regression models controlling for potential confounders. Preeclampsia developed in 3.7% (n = 63) of 1681 women. Cord serum theobromine concentrations were negatively associated with preeclampsia (aOR = 0.31; CI = 0.11-0.87 for highest compared with lowest quartile). Self-reported chocolate consumption estimates also were inversely associated with preeclampsia. Compared with women consuming under 1 serving of chocolate weekly, women consuming 5+ servings per week had decreased risk: aOR = 0.81 with consumption in the first 3 months of pregnancy (CI = 0.37-1.79) and 0.60 in the last 3 months (0.30-1.24). Our results suggest that chocolate consumption during pregnancy may lower risk of preeclampsia. However, reverse causality may also contribute to these findings.

19. Ubiquitous positioning

CERN Document Server

Mannings, Robin

2008-01-01

This groundbreaking resource offers a practical, in-depth understanding of Ubiquitous Positioning - positioning systems that identify the location and position of people, vehicles and objects in time and space in the digitized networked economy. The future and growth of ubiquitous positioning will be fueled by the convergence of many other areas of technology, from mobile telematics, Internet technology, and location systems, to sensing systems, geographic information systems, and the semantic web. This first-of-its-kind volume explores ubiquitous positioning from a convergence perspective, of

20. Positioning consumption

DEFF Research Database (Denmark)

Halkier, Bente; Keller, Margit

2014-01-01

positionings emerges based on empirical examples of research in parent–children consumption. Positionings are flexible discursive fixations of the relationship between the performances of the practitioner, other practitioners, media discourse and consumption activities. The basic positioning types...... are the practice maintenance and the practice change position, with different sorts of adapting in between. Media discourse can become a resource for a resistant position against social control or for an appropriating position in favour of space for action. Regardless of the current relation to a particular media......This article analyses the ways in which media discourses become a part of contested consumption activities. We apply a positioning perspective with practice theory to focus on how practitioners relate to media discourse as a symbolic resource in their everyday practices. A typology of performance...

1. Maximum-likelihood-based extended-source spatial acquisition and tracking for planetary optical communications

Science.gov (United States)

Tsou, Haiping; Yan, Tsun-Yee

1999-04-01

This paper describes an extended-source spatial acquisition and tracking scheme for planetary optical communications. This scheme uses the Sun-lit Earth image as the beacon signal, which can be computed according to the current Sun-Earth-Probe angle from a pre-stored Earth image or a received snapshot taken by other Earth-orbiting satellite. Onboard the spacecraft, the reference image is correlated in the transform domain with the received image obtained from a detector array, which is assumed to have each of its pixels corrupted by an independent additive white Gaussian noise. The coordinate of the ground station is acquired and tracked, respectively, by an open-loop acquisition algorithm and a closed-loop tracking algorithm derived from the maximum likelihood criterion. As shown in the paper, the optimal spatial acquisition requires solving two nonlinear equations, or iteratively solving their linearized variants, to estimate the coordinate when translation in the relative positions of onboard and ground transceivers is considered. Similar assumption of linearization leads to the closed-loop spatial tracking algorithm in which the loop feedback signals can be derived from the weighted transform-domain correlation. Numerical results using a sample Sun-lit Earth image demonstrate that sub-pixel resolutions can be achieved by this scheme in a high disturbance environment.

2. Penalized maximum-likelihood sinogram restoration for dual focal spot computed tomography

International Nuclear Information System (INIS)

Forthmann, P; Koehler, T; Begemann, P G C; Defrise, M

2007-01-01

Due to various system non-idealities, the raw data generated by a computed tomography (CT) machine are not readily usable for reconstruction. Although the deterministic nature of corruption effects such as crosstalk and afterglow permits correction by deconvolution, there is a drawback because deconvolution usually amplifies noise. Methods that perform raw data correction combined with noise suppression are commonly termed sinogram restoration methods. The need for sinogram restoration arises, for example, when photon counts are low and non-statistical reconstruction algorithms such as filtered backprojection are used. Many modern CT machines offer a dual focal spot (DFS) mode, which serves the goal of increased radial sampling by alternating the focal spot between two positions on the anode plate during the scan. Although the focal spot mode does not play a role with respect to how the data are affected by the above-mentioned corruption effects, it needs to be taken into account if regularized sinogram restoration is to be applied to the data. This work points out the subtle difference in processing that sinogram restoration for DFS requires, how it is correctly employed within the penalized maximum-likelihood sinogram restoration algorithm and what impact it has on image quality

3. Building unbiased estimators from non-Gaussian likelihoods with application to shear estimation

International Nuclear Information System (INIS)

Madhavacheril, Mathew S.; Sehgal, Neelima; McDonald, Patrick; Slosar, Anže

2015-01-01

We develop a general framework for generating estimators of a given quantity which are unbiased to a given order in the difference between the true value of the underlying quantity and the fiducial position in theory space around which we expand the likelihood. We apply this formalism to rederive the optimal quadratic estimator and show how the replacement of the second derivative matrix with the Fisher matrix is a generic way of creating an unbiased estimator (assuming choice of the fiducial model is independent of data). Next we apply the approach to estimation of shear lensing, closely following the work of Bernstein and Armstrong (2014). Our first order estimator reduces to their estimator in the limit of zero shear, but it also naturally allows for the case of non-constant shear and the easy calculation of correlation functions or power spectra using standard methods. Both our first-order estimator and Bernstein and Armstrong's estimator exhibit a bias which is quadratic in true shear. Our third-order estimator is, at least in the realm of the toy problem of Bernstein and Armstrong, unbiased to 0.1% in relative shear errors Δg/g for shears up to |g|=0.2

4. Analysis of Pairwise Interactions in a Maximum Likelihood Sense to Identify Leaders in a Group

Directory of Open Access Journals (Sweden)

Violet Mwaffo

2017-07-01

Full Text Available Collective motion in animal groups manifests itself in the form of highly coordinated maneuvers determined by local interactions among individuals. A particularly critical question in understanding the mechanisms behind such interactions is to detect and classify leader–follower relationships within the group. In the technical literature of coupled dynamical systems, several methods have been proposed to reconstruct interaction networks, including linear correlation analysis, transfer entropy, and event synchronization. While these analyses have been helpful in reconstructing network models from neuroscience to public health, rules on the most appropriate method to use for a specific dataset are lacking. Here, we demonstrate the possibility of detecting leaders in a group from raw positional data in a model-free approach that combines multiple methods in a maximum likelihood sense. We test our framework on synthetic data of groups of self-propelled Vicsek particles, where a single agent acts as a leader and both the size of the interaction region and the level of inherent noise are systematically varied. To assess the feasibility of detecting leaders in real-world applications, we study a synthetic dataset of fish shoaling, generated by using a recent data-driven model for social behavior, and an experimental dataset of pharmacologically treated zebrafish. Not only does our approach offer a robust strategy to detect leaders in synthetic data but it also allows for exploring the role of psychoactive compounds on leader–follower relationships.

5. Transformer ratio enhancement experiment

International Nuclear Information System (INIS)

Gai, W.; Power, J. G.; Kanareykin, A.; Neasheva, E.; Altmark, A.

2004-01-01

Recently, a multibunch scheme for efficient acceleration based on dielectric wakefield accelerator technology was outlined in J.G. Power, W. Gai, A. Kanareykin, X. Sun. PAC 2001 Proceedings, pp. 114-116, 2002. In this paper we present an experimental program for the design, development and demonstration of an Enhanced Transformer Ratio Dielectric Wakefield Accelerator (ETR-DWA). The principal goal is to increase the transformer ratio R, the parameter that characterizes the energy transfer efficiency from the accelerating structure to the accelerated electron beam. We present here an experimental design of a 13.625 GHz dielectric loaded accelerating structure, a laser multisplitter producing a ramped bunch train, and simulations of the bunch train parameters required. Experimental results of the accelerating structure bench testing and ramped pulsed train generation with the laser multisplitter are shown as well. Using beam dynamic simulations, we also obtain the focusing FODO lattice parameters

6. Intake to Production Ratio

DEFF Research Database (Denmark)

Nazaroff, William; Weschler, Charles J.; Little, John C.

2012-01-01

BACKGROUND: Limited data are available to assess human exposure to thousands of chemicals currently in commerce. Information that relates human intake of a chemical to its production and use can help inform understanding of mechanisms and pathways that control exposure and support efforts...... to protect public health.OBJECTIVES: We introduce the intake-to-production ratio (IPR) as an economy-wide quantitative indicator of the extent to which chemical production results in human exposure.METHODS: The IPR was evaluated as the ratio of two terms: aggregate rate of chemical uptake in a human......(n-butyl) phthalate, 1,040 ppm for para-dichlorobenzene, 6,800 ppm for di(isobutyl) phthalate, 7,700 ppm for diethyl phthalate, and 8,000-24,000 ppm (range) for triclosan.CONCLUSION: The IPR is well suited as an aggregate metric of exposure intensity for characterizing population-level exposure to synthesized...

7. Identification of contemporary selection signatures using composite log likelihood and their associations with marbling score in Korean cattle.

Science.gov (United States)

Ryu, Jihye; Lee, Chaeyoung

2014-12-01

Positive selection not only increases beneficial allele frequency but also causes augmentation of allele frequencies of sequence variants in close proximity. Signals for positive selection were detected by the statistical differences in subsequent allele frequencies. To identify selection signatures in Korean cattle, we applied a composite log-likelihood (CLL)-based method, which calculates a composite likelihood of the allelic frequencies observed across sliding windows of five adjacent loci and compares the value with the critical statistic estimated by 50,000 permutations. Data for a total of 11,799 nucleotide polymorphisms were used with 71 Korean cattle and 209 foreign beef cattle. As a result, 147 signals were identified for Korean cattle based on CLL estimates (P selected. Further genetic association analysis with 41 intragenic variants in the selection signatures with the greatest CLL for each chromosome revealed that marbling score was associated with five variants. Intensive association studies with all the selection signatures identified in this study are required to exclude signals associated with other phenotypes or signals falsely detected and thus to identify genetic markers for meat quality. © 2014 Stichting International Foundation for Animal Genetics.

8. Estimation of Model's Marginal likelihood Using Adaptive Sparse Grid Surrogates in Bayesian Model Averaging

Science.gov (United States)

Zeng, X.

2015-12-01

A large number of model executions are required to obtain alternative conceptual models' predictions and their posterior probabilities in Bayesian model averaging (BMA). The posterior model probability is estimated through models' marginal likelihood and prior probability. The heavy computation burden hinders the implementation of BMA prediction, especially for the elaborated marginal likelihood estimator. For overcoming the computation burden of BMA, an adaptive sparse grid (SG) stochastic collocation method is used to build surrogates for alternative conceptual models through the numerical experiment of a synthetical groundwater model. BMA predictions depend on model posterior weights (or marginal likelihoods), and this study also evaluated four marginal likelihood estimators, including arithmetic mean estimator (AME), harmonic mean estimator (HME), stabilized harmonic mean estimator (SHME), and thermodynamic integration estimator (TIE). The results demonstrate that TIE is accurate in estimating conceptual models' marginal likelihoods. The BMA-TIE has better predictive performance than other BMA predictions. TIE has high stability for estimating conceptual model's marginal likelihood. The repeated estimated conceptual model's marginal likelihoods by TIE have significant less variability than that estimated by other estimators. In addition, the SG surrogates are efficient to facilitate BMA predictions, especially for BMA-TIE. The number of model executions needed for building surrogates is 4.13%, 6.89%, 3.44%, and 0.43% of the required model executions of BMA-AME, BMA-HME, BMA-SHME, and BMA-TIE, respectively.

9. Massive optimal data compression and density estimation for scalable, likelihood-free inference in cosmology

Science.gov (United States)

Alsing, Justin; Wandelt, Benjamin; Feeney, Stephen

2018-03-01

Many statistical models in cosmology can be simulated forwards but have intractable likelihood functions. Likelihood-free inference methods allow us to perform Bayesian inference from these models using only forward simulations, free from any likelihood assumptions or approximations. Likelihood-free inference generically involves simulating mock data and comparing to the observed data; this comparison in data-space suffers from the curse of dimensionality and requires compression of the data to a small number of summary statistics to be tractable. In this paper we use massive asymptotically-optimal data compression to reduce the dimensionality of the data-space to just one number per parameter, providing a natural and optimal framework for summary statistic choice for likelihood-free inference. Secondly, we present the first cosmological application of Density Estimation Likelihood-Free Inference (DELFI), which learns a parameterized model for joint distribution of data and parameters, yielding both the parameter posterior and the model evidence. This approach is conceptually simple, requires less tuning than traditional Approximate Bayesian Computation approaches to likelihood-free inference and can give high-fidelity posteriors from orders of magnitude fewer forward simulations. As an additional bonus, it enables parameter inference and Bayesian model comparison simultaneously. We demonstrate Density Estimation Likelihood-Free Inference with massive data compression on an analysis of the joint light-curve analysis supernova data, as a simple validation case study. We show that high-fidelity posterior inference is possible for full-scale cosmological data analyses with as few as ˜104 simulations, with substantial scope for further improvement, demonstrating the scalability of likelihood-free inference to large and complex cosmological datasets.

10. Maximal information analysis: I - various Wayne State plots and the most common likelihood principle

International Nuclear Information System (INIS)

Bonvicini, G.

2005-01-01

Statistical analysis using all moments of the likelihood L(y vertical bar α) (y being the data and α being the fit parameters) is presented. The relevant plots for various data fitting situations are presented. The goodness of fit (GOF) parameter (currently the χ 2 ) is redefined as the isoprobability level in a multidimensional space. Many useful properties of statistical analysis are summarized in a new statistical principle which states that the most common likelihood, and not the tallest, is the best possible likelihood, when comparing experiments or hypotheses

11. Simplified likelihood for the re-interpretation of public CMS results

CERN Document Server

The CMS Collaboration

2017-01-01

In this note, a procedure for the construction of simplified likelihoods for the re-interpretation of the results of CMS searches for new physics is presented. The procedure relies on the use of a reduced set of information on the background models used in these searches which can readily be provided by the CMS collaboration. A toy example is used to demonstrate the procedure and its accuracy in reproducing the full likelihood for setting limits in models for physics beyond the standard model. Finally, two representative searches from the CMS collaboration are used to demonstrate the validity of the simplified likelihood approach under realistic conditions.

12. [Using log-binomial model for estimating the prevalence ratio].

Science.gov (United States)

Ye, Rong; Gao, Yan-hui; Yang, Yi; Chen, Yue

2010-05-01

To estimate the prevalence ratios, using a log-binomial model with or without continuous covariates. Prevalence ratios for individuals' attitude towards smoking-ban legislation associated with smoking status, estimated by using a log-binomial model were compared with odds ratios estimated by logistic regression model. In the log-binomial modeling, maximum likelihood method was used when there were no continuous covariates and COPY approach was used if the model did not converge, for example due to the existence of continuous covariates. We examined the association between individuals' attitude towards smoking-ban legislation and smoking status in men and women. Prevalence ratio and odds ratio estimation provided similar results for the association in women since smoking was not common. In men however, the odds ratio estimates were markedly larger than the prevalence ratios due to a higher prevalence of outcome. The log-binomial model did not converge when age was included as a continuous covariate and COPY method was used to deal with the situation. All analysis was performed by SAS. Prevalence ratio seemed to better measure the association than odds ratio when prevalence is high. SAS programs were provided to calculate the prevalence ratios with or without continuous covariates in the log-binomial regression analysis.

13. [DIN-compatible vision assessment of increased reproducibility using staircase measurement and maximum likelihood analysis].

Science.gov (United States)

Weigmann, U; Petersen, J

1996-08-01

Visual acuity determination according to DIN 58,220 does not make full use of the information received about the patient, in contrast to the staircase method. Thus, testing the same number of optotypes, the staircase method should yield more reproducible acuity results. On the other hand, the staircase method gives systematically higher acuity values because it converges on the 48% point of the psychometric function (for Landolt rings in eight positions) and not on the 65% probability, as DIN 58,220 with criterion 3/5 does. This bias can be avoided by means of a modified evaluation. Using the staircase data we performed a maximum likelihood estimate of the psychometric function as a whole and computed the acuity value for 65% probability of correct answers. We determined monocular visual acuity in 102 persons with widely differing visual performance. Each subject underwent four tests in random order, two according to DIN 58,220 and two using the modified staircase method (Landolt rings in eight positions scaled by a factor 1.26; PC monitor with 1024 x 768 pixels; distance 4.5 m). Each test was performed with 25 optotypes. The two procedures provide the same mean visual acuity values (difference less than 0.02 acuity steps). The test-retest results match in 30.4% of DIN repetitions but in 50% of the staircases. The standard deviation of the test-retest difference is 1.41 (DIN) and 1.06 (modified staircase) acuity steps. Thus the standard deviation of the single test is 1.0 (DIN) and 0.75 (modified staircase) acuity steps. The new method provides visual acuity values identical to DIN 58,220 but is superior with respect to reproducibility.

14. Likelihood inference for COM-Poisson cure rate model with interval-censored data and Weibull lifetimes.

Science.gov (United States)

Pal, Suvra; Balakrishnan, N

2017-10-01

In this paper, we consider a competing cause scenario and assume the number of competing causes to follow a Conway-Maxwell Poisson distribution which can capture both over and under dispersion that is usually encountered in discrete data. Assuming the population of interest having a component cure and the form of the data to be interval censored, as opposed to the usually considered right-censored data, the main contribution is in developing the steps of the expectation maximization algorithm for the determination of the maximum likelihood estimates of the model parameters of the flexible Conway-Maxwell Poisson cure rate model with Weibull lifetimes. An extensive Monte Carlo simulation study is carried out to demonstrate the performance of the proposed estimation method. Model discrimination within the Conway-Maxwell Poisson distribution is addressed using the likelihood ratio test and information-based criteria to select a suitable competing cause distribution that provides the best fit to the data. A simulation study is also carried out to demonstrate the loss in efficiency when selecting an improper competing cause distribution which justifies the use of a flexible family of distributions for the number of competing causes. Finally, the proposed methodology and the flexibility of the Conway-Maxwell Poisson distribution are illustrated with two known data sets from the literature: smoking cessation data and breast cosmesis data.

15. Myocardial perfusion assessment by dual-energy computed tomography in patients with intermediate to high likelihood of coronary artery disease

International Nuclear Information System (INIS)

De Zam, M.C.; Capunay, C.; Rodriguez Granillo, G.A.; Deviggiano, A.; Campisi, R.; Munain, M. López de; Vallejos, J.; Carrascosa, P.M.

2015-01-01

Objectives. We sought to explore the feasibility and diagnostic performance of dual-energy computed tomography (DECT) for the evaluation of myocardial perfusion in patients with intermediate to high likelihood of coronary artery disease (CAD), and to assess the impact of beam hardening artifacts (HAE). Methods. The present prospective study involved patients with known or suspected CAD referred for myocardial perfusion imaging by single-photon emission computed tomography (SPECT). Twenty patients were included in the study protocol, and scanned using DECT imaging (n = 20). The same pharmacological stress was used for DECT and SPECT scans. Results. A total of 680 left ventricular segments were evaluated by DECT and SPECT. The contrast to noise ratio was 8.8±2.9. The diagnostic performance of DECT was very good in identifying perfusion defects [area under ROC curve (AUC) of DECT 0.90 (0.86-0.94)] compared with SPECT, and remained unaffected when including only segments affected by beam hardening artifacts (BHA) [AUC= DECT 0.90 (0.84-0.96)]. Conclusions. In this pilot investigation, myocardial perfusion assessment by DECT imaging in patients with intermediate to high likelihood of CAD was feasible and remained unaffected by the presence of BHA. (authors) [es

16. Gender Differences in the relationship between carbonated sugar-sweetened beverage intake and the likelihood of hypertension according to obesity.

Science.gov (United States)

Song, Hong Ji; Paek, Yu Jin; Choi, Min Kyu; Yoo, Ki-Bong; Kang, Jae-Heon; Lee, Hae-Jeung

2017-06-01

The aim of the present study was to investigate the association between hypertension and carbonated sugar-sweetened beverages (SSB) intake according to gender and obesity. The study used data from 2007, 2008 and 2009 Korea National Health and Nutrition Examination Surveys. A total of 9869 subjects (men = 3845 and women = 6024) were included. SSB intakes were calculated from food frequency questionnaires. Odds ratios (ORs) and 95 % confidence interval (CI) for hypertension were assessed using survey logistic regression and multivariable adjusted models. A total of 14.5 % of individuals were classified as having hypertension. The likelihood of hypertension in the third, fourth and fifth quintiles for SSB intake increased to OR 1.00, 1.20 and 1.42 respectively, after adjusting for confounding factors. Compared to the participants in the lowest tertile for SSB intake, participants in the third tertile showed an increased likelihood of hypertension with ORs (CI) of 2.00 (1.21-3.31) and 1.75 (1.23-2.49) for obese women and non-obese men, respectively. The present study showed gender differences in the relationship between carbonated SSB intake and the hypertension according to obesity.

17. Influence of population size, density, and proximity to talent clubs on the likelihood of becoming elite youth athlete.

Science.gov (United States)

Rossing, N N; Stentoft, D; Flattum, A; Côté, J; Karbing, D S

2018-03-01

Previous studies have found significant differences in the likelihood of becoming an elite athlete depending on community population sizes and densities, an effect known as the place of early development, or birthplace effect. However, the results have not been consistent between sports or European countries. As both professional and voluntary clubs are vital to the talent development systems in Europe, the proximity of an athlete's place of early development to the location of talent clubs may be an important predictor of the likelihood of becoming an elite athlete. Therefore, the primary purpose of this study was to investigate the place of early development effect and the effect of proximity to talent clubs. The samples included elite youth league athletes (579 football and 311 handball) and national youth athletes (85 football and 80 handball) and a comparison group of 147 221 football and 26 290 handball youth athletes. Odds ratios showed variations in the optimal community size and density across sports. Geospatial analyses of proximity to talent clubs highlighted a trend indicating that most national and elite youth league athletes in both sports had their place of early development in their sport near a talent club. The results suggest that proximity is an important predictor in the development of expertise across sports, but future studies need to clarify if proximity is important in other countries and sports. © 2017 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

18. Bleaching increases likelihood of disease on Acropora palmata (Lamarck) in Hawksnest Bay, St John, US Virgin Islands

Science.gov (United States)

Muller, E. M.; Rogers, C. S.; Spitzack, A. S.; van Woesik, R.

2008-03-01

Anomalously high water temperatures may enhance the likelihood of coral disease outbreaks by increasing the abundance or virulence of pathogens, or by increasing host susceptibility. This study tested the compromised-host hypothesis, and documented the relationship between disease and temperature, through monthly monitoring of Acropora palmata colonies from May 2004 to December 2006, in Hawksnest Bay, St John, US Virgin Islands (USVI). Disease prevalence and the rate of change in prevalence showed a positive linear relationship with water temperature and rate of change in water temperature, respectively, but only in 2005 during prolonged periods of elevated temperature. Both bleached and unbleached colonies showed a positive relationship between disease prevalence and temperature in 2005, but the average area of disease-associated mortality increased only for bleached corals, indicating host susceptibility, rather than temperature per se, influenced disease severity on A. palmata.

19. The Reference Return Ratio

DEFF Research Database (Denmark)

Nicolaisen, Jeppe; Faber Frandsen, Tove

2008-01-01

The paper introduces a new journal impact measure called The Reference Return Ratio (3R). Unlike the traditional Journal Impact Factor (JIF), which is based on calculations of publications and citations, the new measure is based on calculations of bibliographic investments (references) and returns...... (citations). A comparative study of the two measures shows a strong relationship between the 3R and the JIF. Yet, the 3R appears to correct for citation habits, citation dynamics, and composition of document types - problems that typically are raised against the JIF. In addition, contrary to traditional...

20. Potential support ratios

DEFF Research Database (Denmark)

Kjærgaard, Søren; Canudas-Romo, Vladimir

2017-01-01

The ‘prospective potential support ratio’ has been proposed by researchers as a measure that accurately quantifies the burden of ageing, by identifying the fraction of a population that has passed a certain measure of longevity, for example, 17 years of life expectancy. Nevertheless......, the prospective potential support ratio usually focuses on the current mortality schedule, or period life expectancy. Instead, in this paper we look at the actual mortality experienced by cohorts in a population, using cohort life tables. We analyse differences between the two perspectives using mortality models...

1. Wald Sequential Probability Ratio Test for Analysis of Orbital Conjunction Data

Science.gov (United States)

Carpenter, J. Russell; Markley, F. Landis; Gold, Dara

2013-01-01

We propose a Wald Sequential Probability Ratio Test for analysis of commonly available predictions associated with spacecraft conjunctions. Such predictions generally consist of a relative state and relative state error covariance at the time of closest approach, under the assumption that prediction errors are Gaussian. We show that under these circumstances, the likelihood ratio of the Wald test reduces to an especially simple form, involving the current best estimate of collision probability, and a similar estimate of collision probability that is based on prior assumptions about the likelihood of collision.

2. Ratios of helicity amplitudes for exclusive ρ{sup 0} electroproduction on transversely polarized protons

Energy Technology Data Exchange (ETDEWEB)

Airapetian, A. [Justus-Liebig Universitaet Giessen, II. Physikalisches Institut, Giessen (Germany); University of Michigan, Randall Laboratory of Physics, Ann Arbor, MI (United States); Akopov, N.; Elbakian, G.; Gharibyan, V.; Marukyan, H.; Petrosyan, A. [Yerevan Physics Institute, Yerevan (Armenia); Akopov, Z.; Borissov, A.; Deconinck, W.; Holler, Y.; Rostomyan, A.; Zihlmann, B. [DESY, Hamburg (Germany); Aschenauer, E.C.; Nowak, W.D. [DESY, Zeuthen (Germany); Augustyniak, W.; Marianski, B.; Trzcinski, A.; Zupranski, P. [National Centre for Nuclear Research, Warsaw (Poland); Belostotski, S.; Kisselev, A.; Manaenkov, S.I.; Veretennikov, D.; Vikhrov, V. [B.P. Konstantinov Petersburg Nuclear Physics Institute, Leningrad Region (Russian Federation); Blok, H.P. [National Institute for Subatomic Physics (Nikhef), Amsterdam (Netherlands); VU University, Department of Physics and Astronomy, Amsterdam (Netherlands); Bryzgalov, V.; Ivanilov, A.; Korotkov, V.; Salomatin, Y. [Institute for High Energy Physics, Moscow Region (Russian Federation); Capitani, G.P.; De Sanctis, E.; Muccifora, V.; Reolon, A.R. [Istituto Nazionale di Fisica Nucleare, Laboratori Nazionali di Frascati, Frascati (Italy); Ciullo, G.; Lenisa, P.; Pappalardo, L.L.; Statera, M. [Istituto Nazionale di Fisica Nucleare, Sezione di Ferrara, Ferrara (Italy); Universita di Ferrara, Dipartimento di Fisica e Scienze della Terra, Ferrara (Italy); Contalbrigo, M. [Istituto Nazionale di Fisica Nucleare, Sezione di Ferrara, Ferrara (Italy); De Leo, R.; Lagamba, L.; Vilardi, I. [Istituto Nazionale di Fisica Nucleare, Sezione di Bari, Bari (Italy); Dueren, M. [Justus-Liebig Universitaet Giessen, II. Physikalisches Institut, Giessen (Germany); Ellinghaus, F. [University of Colorado, Nuclear Physics Laboratory, Boulder, CO (United States); Felawka, L. [TRIUMF, Vancouver, BC (Canada); Frullani, S.; Garibaldi, F. [Istituto Nazionale di Fisica Nucleare, Sezione di Roma, Gruppo Collegato Sanita, Rome (Italy); Istituto Superiore di Sanita, Rome (Italy); Gavrilov, G. [DESY, Hamburg (Germany); B.P. Konstantinov Petersburg Nuclear Physics Institute, Leningrad Region (Russian Federation); TRIUMF, Vancouver, BC (Canada); Goloskokov, S.V.; Shutov, V. [Joint Institute for Nuclear Research, Dubna (Russian Federation); Jackson, H.E.; Reimer, P.E. [Argonne National Laboratory, Physics Division, Argonne, IL (United States); Joosten, S. [Ghent University, Department of Physics and Astronomy, Gent (Belgium); University of Illinois, Department of Physics, Urbana, IL (United States); Kaiser, R.; Lehmann, I.; Rosner, G.; Seitz, B. [University of Glasgow, SUPA, School of Physics and Astronomy, Glasgow (United Kingdom); Karyan, G. [DESY, Hamburg (Germany); Yerevan Physics Institute, Yerevan (Armenia); Kozlov, V.; Terkulov, A. [Lebedev Physical Institute, Moscow (Russian Federation); Kravchenko, P. [Universitaet Erlangen-Nuernberg, Physikalisches Institut, Erlangen (Germany); B.P. Konstantinov Petersburg Nuclear Physics Institute, Leningrad Region (Russian Federation); Kroll, P.; Schaefer, A. [Universitaet Regensburg, Institut fuer Theoretische Physik, Regensburg (Germany); Lapikas, L. [National Institute for Subatomic Physics (Nikhef), Amsterdam (Netherlands); Lorenzon, W. [University of Michigan, Randall Laboratory of Physics, Ann Arbor, MI (United States); Miyachi, Y.; Shibata, T.A. [Tokyo Institute of Technology, Department of Physics, Tokyo (Japan); Movsisyan, A. [Istituto Nazionale di Fisica Nucleare, Sezione di Ferrara, Ferrara (Italy); Yerevan Physics Institute, Yerevan (Armenia); Nass, A.; Rith, K. [Universitaet Erlangen-Nuernberg, Physikalisches Institut, Erlangen (Germany); Riedl, C. [DESY, Zeuthen (Germany); University of Illinois, Department of Physics, Urbana, IL (United States); Ryckbosch, D.; Tytgat, M.; Haarlem, Y. van [Ghent University, Department of Physics and Astronomy, Gent (Belgium); Schnell, G. [University of the Basque Country UPV/EHU, Department of Theoretical Physics, Bilbao (Spain); Basque Foundation for Science, IKERBASQUE, Bilbao (Spain); Ghent University, Department of Physics and Astronomy, Gent (Belgium); Truty, R. [University of Illinois, Department of Physics, Urbana, IL (United States); Hulse, C. van [University of the Basque Country UPV/EHU, Department of Theoretical Physics, Bilbao (Spain); Ghent University, Department of Physics and Astronomy, Gent (Belgium); Yaschenko, S. [DESY, Hamburg (Germany); Universitaet Erlangen-Nuernberg, Physikalisches Institut, Erlangen (Germany); Collaboration: The HERMES Collaboration

2017-06-15

Exclusive ρ{sup 0}-meson electroproduction is studied by the HERMES experiment, using the 27.6 GeV longitudinally polarized electron/positron beam of HERA and a transversely polarized hydrogen target, in the kinematic region 1.0 GeV{sup 2} < Q{sup 2} < 7.0 GeV{sup 2}, 3.0 GeV < W < 6.3 GeV, and -t{sup '} < 0.4 GeV{sup 2}. Using an unbinned maximum-likelihood method, 25 parameters are extracted. These determine the real and imaginary parts of the ratios of several helicity amplitudes describing ρ{sup 0}-meson production by a virtual photon. The denominator of those ratios is the dominant amplitude, the nucleon-helicity-non-flip amplitude F{sub 0(1)/(2)0(1)/(2)}, which describes the production of a longitudinal ρ{sup 0}-meson by a longitudinal virtual photon. The ratios of nucleon-helicity-non-flip amplitudes are found to be in good agreement with those from the previous HERMES analysis. The transverse target polarization allows for the first time the extraction of ratios of a number of nucleon-helicity-flip amplitudes to F{sub 0(1)/(2)0(1)/(2)}. Results obtained in a handbag approach based on generalized parton distributions taking into account the contribution from pion exchange are found to be in good agreement with these ratios. Within the model, the data favor a positive sign for the π - ρ transition form factor. By also exploiting the longitudinal beam polarization, a total of 71 ρ{sup 0} spin-density matrix elements is determined from the extracted 25 parameters, in contrast to only 53 elements as directly determined in earlier analyses. (orig.)

3. Ratios of helicity amplitudes for exclusive ρ{sup 0} electroproduction on transversely polarized protons

Energy Technology Data Exchange (ETDEWEB)

Airapetian, A. [Giessen Univ. (Germany). 2. Physikalisches Inst.; Michigan Univ., Ann Arbor, MI (United States). Randall Lab. of Physics; Akopov, N. [Yerevan Physics Institute (Armenia); Akopov, Z. [DESY, Hamburg (Germany); Collaboration: HERMES Collaboration; and others

2017-06-13

Exclusive ρ{sup 0}-meson electroproduction is studied by the HERMES experiment, using the 27.6 GeV longitudinally polarized electron/positron beam of HERA and a transversely polarized hydrogen target, in the kinematic region 1.0 GeV{sup 2}likelihood method, 25 parameters are extracted. These determine the real and imaginary parts of the ratios of several helicity amplitudes describing ρ{sup 0}-meson production by a virtual photon. The denominator of those ratios is the dominant amplitude, the nucleon-helicity-non-flip amplitude F{sub 0(1)/(2)0(1)/(2)}, which describes the production of a longitudinal ρ{sup 0}-meson by a longitudinal virtual photon. The ratios of nucleon-helicity-non-flip amplitudes are found to be in good agreement with those from the previous HERMES analysis. The transverse target polarization allows for the first time the extraction of ratios of a number of nucleon-helicity-flip amplitudes to F{sub 0(1)/(2)0(1)/(2)}. Results obtained in a handbag approach based on generalized parton distributions taking into account the contribution from pion exchange are found to be in good agreement with these ratios. Within the model, the data favor a positive sign for the π-ρ transition form factor. By also exploiting the longitudinal beam polarization, a total of 71 ρ{sup 0} spin-density matrix elements is determined from the extracted 25 parameters, in contrast to only 53 elements as directly determined in earlier analyses.

4. Uncertainty in a monthly water balance model using the generalized likelihood uncertainty estimation methodology

Science.gov (United States)

Rivera, Diego; Rivas, Yessica; Godoy, Alex

2015-02-01

Hydrological models are simplified representations of natural processes and subject to errors. Uncertainty bounds are a commonly used way to assess the impact of an input or model architecture uncertainty in model outputs. Different sets of parameters could have equally robust goodness-of-fit indicators, which is known as Equifinality. We assessed the outputs from a lumped conceptual hydrological model to an agricultural watershed in central Chile under strong interannual variability (coefficient of variability of 25%) by using the Equifinality concept and uncertainty bounds. The simulation period ran from January 1999 to December 2006. Equifinality and uncertainty bounds from GLUE methodology (Generalized Likelihood Uncertainty Estimation) were used to identify parameter sets as potential representations of the system. The aim of this paper is to exploit the use of uncertainty bounds to differentiate behavioural parameter sets in a simple hydrological model. Then, we analyze the presence of equifinality in order to improve the identification of relevant hydrological processes. The water balance model for Chillan River exhibits, at a first stage, equifinality. However, it was possible to narrow the range for the parameters and eventually identify a set of parameters representing the behaviour of the watershed (a behavioural model) in agreement with observational and soft data (calculation of areal precipitation over the watershed using an isohyetal map). The mean width of the uncertainty bound around the predicted runoff for the simulation period decreased from 50 to 20 m3s-1 after fixing the parameter controlling the areal precipitation over the watershed. This decrement is equivalent to decreasing the ratio between simulated and observed discharge from 5.2 to 2.5. Despite the criticisms against the GLUE methodology, such as the lack of statistical formality, it is identified as a useful tool assisting the modeller with the identification of critical parameters.

5. Tau hadronic branching ratios

CERN Document Server

Buskulic, Damir; De Bonis, I; Décamp, D; Ghez, P; Goy, C; Lees, J P; Lucotte, A; Minard, M N; Odier, P; Pietrzyk, B; Ariztizabal, F; Chmeissani, M; Crespo, J M; Efthymiopoulos, I; Fernández, E; Fernández-Bosman, M; Gaitan, V; Martínez, M; Orteu, S; Pacheco, A; Padilla, C; Palla, Fabrizio; Pascual, A; Perlas, J A; Sánchez, F; Teubert, F; Colaleo, A; Creanza, D; De Palma, M; Farilla, A; Gelao, G; Girone, M; Iaselli, Giuseppe; Maggi, G; Maggi, M; Marinelli, N; Natali, S; Nuzzo, S; Ranieri, A; Raso, G; Romano, F; Ruggieri, F; Selvaggi, G; Silvestris, L; Tempesta, P; Zito, G; Huang, X; Lin, J; Ouyang, Q; Wang, T; Xie, Y; Xu, R; Xue, S; Zhang, J; Zhang, L; Zhao, W; Bonvicini, G; Cattaneo, M; Comas, P; Coyle, P; Drevermann, H; Engelhardt, A; Forty, Roger W; Frank, M; Hagelberg, R; Harvey, J; Jacobsen, R; Janot, P; Jost, B; Kneringer, E; Knobloch, J; Lehraus, Ivan; Markou, C; Martin, E B; Mato, P; Minten, Adolf G; Miquel, R; Oest, T; Palazzi, P; Pater, J R; Pusztaszeri, J F; Ranjard, F; Rensing, P E; Rolandi, Luigi; Schlatter, W D; Schmelling, M; Schneider, O; Tejessy, W; Tomalin, I R; Venturi, A; Wachsmuth, H W; Wiedenmann, W; Wildish, T; Witzeling, W; Wotschack, J; Ajaltouni, Ziad J; Bardadin-Otwinowska, Maria; Barrès, A; Boyer, C; Falvard, A; Gay, P; Guicheney, C; Henrard, P; Jousset, J; Michel, B; Monteil, S; Pallin, D; Perret, P; Podlyski, F; Proriol, J; Rossignol, J M; Saadi, F; Fearnley, Tom; Hansen, J B; Hansen, J D; Hansen, J R; Hansen, P H; Nilsson, B S; Kyriakis, A; Simopoulou, Errietta; Siotis, I; Vayaki, Anna; Zachariadou, K; Blondel, A; Bonneaud, G R; Brient, J C; Bourdon, P; Passalacqua, L; Rougé, A; Rumpf, M; Tanaka, R; Valassi, Andrea; Verderi, M; Videau, H L; Candlin, D J; Parsons, M I; Focardi, E; Parrini, G; Corden, M; Delfino, M C; Georgiopoulos, C H; Jaffe, D E; Antonelli, A; Bencivenni, G; Bologna, G; Bossi, F; Campana, P; Capon, G; Chiarella, V; Felici, G; Laurelli, P; Mannocchi, G; Murtas, F; Murtas, G P; Pepé-Altarelli, M; Dorris, S J; Halley, A W; ten Have, I; Knowles, I G; Lynch, J G; Morton, W T; O'Shea, V; Raine, C; Reeves, P; Scarr, J M; Smith, K; Smith, M G; Thompson, A S; Thomson, F; Thorn, S; Turnbull, R M; Becker, U; Braun, O; Geweniger, C; Graefe, G; Hanke, P; Hepp, V; Kluge, E E; Putzer, A; Rensch, B; Schmidt, M; Sommer, J; Stenzel, H; Tittel, K; Werner, S; Wunsch, M; Beuselinck, R; Binnie, David M; Cameron, W; Colling, D J; Dornan, Peter J; Konstantinidis, N P; Moneta, L; Moutoussi, A; Nash, J; San Martin, G; Sedgbeer, J K; Stacey, A M; Dissertori, G; Girtler, P; Kuhn, D; Rudolph, G; Bowdery, C K; Brodbeck, T J; Colrain, P; Crawford, G; Finch, A J; Foster, F; Hughes, G; Sloan, Terence; Whelan, E P; Williams, M I; Galla, A; Greene, A M; Kleinknecht, K; Quast, G; Raab, J; Renk, B; Sander, H G; Wanke, R; Van Gemmeren, P; Zeitnitz, C; Aubert, Jean-Jacques; Bencheikh, A M; Benchouk, C; Bonissent, A; Bujosa, G; Calvet, D; Carr, J; Diaconu, C A; Etienne, F; Thulasidas, M; Nicod, D; Payre, P; Rousseau, D; Talby, M; Abt, I; Assmann, R W; Bauer, C; Blum, Walter; Brown, D; Dietl, H; Dydak, Friedrich; Ganis, G; Gotzhein, C; Jakobs, K; Kroha, H; Lütjens, G; Lutz, Gerhard; Männer, W; Moser, H G; Richter, R H; Rosado-Schlosser, A; Schael, S; Settles, Ronald; Seywerd, H C J; Saint-Denis, R; Wolf, G; Alemany, R; Boucrot, J; Callot, O; Cordier, A; Courault, F; Davier, M; Duflot, L; Grivaz, J F; Heusse, P; Jacquet, M; Kim, D W; Le Diberder, F R; Lefrançois, J; Lutz, A M; Musolino, G; Nikolic, I A; Park, H J; Park, I C; Schune, M H; Simion, S; Veillet, J J; Videau, I; Abbaneo, D; Azzurri, P; Bagliesi, G; Batignani, G; Bettarini, S; Bozzi, C; Calderini, G; Carpinelli, M; Ciocci, M A; Ciulli, V; Dell'Orso, R; Fantechi, R; Ferrante, I; Foà, L; Forti, F; Giassi, A; Giorgi, M A; Gregorio, A; Ligabue, F; Lusiani, A; Marrocchesi, P S; Messineo, A; Rizzo, G; Sanguinetti, G; Sciabà, A; Spagnolo, P; Steinberger, Jack; Tenchini, Roberto; Tonelli, G; Triggiani, G; Vannini, C; Verdini, P G; Walsh, J; Betteridge, A P; Blair, G A; Bryant, L M; Cerutti, F; Gao, Y; Green, M G; Johnson, D L; Medcalf, T; Mir, L M; Perrodo, P; Strong, J A; Bertin, V; Botterill, David R; Clifft, R W; Edgecock, T R; Haywood, S; Edwards, M; Maley, P; Norton, P R; Thompson, J C; Bloch-Devaux, B; Colas, P; Emery, S; Kozanecki, Witold; Lançon, E; Lemaire, M C; Locci, E; Marx, B; Pérez, P; Rander, J; Renardy, J F; Roussarie, A; Schuller, J P; Schwindling, J; Trabelsi, A; Vallage, B; Johnson, R P; Kim, H Y; Litke, A M; McNeil, M A; Taylor, G; Beddall, A; Booth, C N; Boswell, R; Cartwright, S L; Combley, F; Dawson, I; Köksal, A; Letho, M; Newton, W M; Rankin, C; Thompson, L F; Böhrer, A; Brandt, S; Cowan, G D; Feigl, E; Grupen, Claus; Lutters, G; Minguet-Rodríguez, J A; Rivera, F; Saraiva, P; Smolik, L; Stephan, F; Apollonio, M; Bosisio, L; Della Marina, R; Giannini, G; Gobbo, B; Ragusa, F; Rothberg, J E; Wasserbaech, S R; Armstrong, S R; Bellantoni, L; Elmer, P; Feng, Z; Ferguson, D P S; Gao, Y S; González, S; Grahl, J; Harton, J L; Hayes, O J; Hu, H; McNamara, P A; Nachtman, J M; Orejudos, W; Pan, Y B; Saadi, Y; Schmitt, M; Scott, I J; Sharma, V; Turk, J; Walsh, A M; Wu Sau Lan; Wu, X; Yamartino, J M; Zheng, M; Zobernig, G

1996-01-01

From 64492 selected \\tau-pair events, produced at the Z^0 resonance, the measurement of the tau decays into hadrons from a global analysis using 1991, 1992 and 1993 ALEPH data is presented. Special emphasis is given to the reconstruction of photons and \\pi^0's, and the removal of fake photons. A detailed study of the systematics entering the \\pi^0 reconstruction is also given. A complete and consistent set of tau hadronic branching ratios is presented for 18 exclusive modes. Most measurements are more precise than the present world average. The new level of precision reached allows a stringent test of \\tau-\\mu universality in hadronic decays, g_\\tau/g_\\mu \\ = \\ 1.0013 \\ \\pm \\ 0.0095, and the first measurement of the vector and axial-vector contributions to the non-strange hadronic \\tau decay width: R_{\\tau ,V} \\ = \\ 1.788 \\ \\pm \\ 0.025 and R_{\\tau ,A} \\ = \\ 1.694 \\ \\pm \\ 0.027. The ratio (R_{\\tau ,V} - R_{\\tau ,A}) / (R_{\\tau ,V} + R_{\\tau ,A}), equal to (2.7 \\pm 1.3) \\ \\%, is a measure of the importance of Q...

6. Debris Likelihood, based on GhostNet, NASA Aqua MODIS, and GOES Imager, EXPERIMENTAL

Data.gov (United States)

National Oceanic and Atmospheric Administration, Department of Commerce — Debris Likelihood Index (Estimated) is calculated from GhostNet, NASA Aqua MODIS Chl a and NOAA GOES Imager SST data. THIS IS AN EXPERIMENTAL PRODUCT: intended...

7. A biclustering algorithm for binary matrices based on penalized Bernoulli likelihood

KAUST Repository

Lee, Seokho; Huang, Jianhua Z.

2013-01-01

We propose a new biclustering method for binary data matrices using the maximum penalized Bernoulli likelihood estimation. Our method applies a multi-layer model defined on the logits of the success probabilities, where each layer represents a

8. Finite mixture model: A maximum likelihood estimation approach on time series data

Science.gov (United States)

Yen, Phoong Seuk; Ismail, Mohd Tahir; Hamzah, Firdaus Mohamad

2014-09-01

Recently, statistician emphasized on the fitting of finite mixture model by using maximum likelihood estimation as it provides asymptotic properties. In addition, it shows consistency properties as the sample sizes increases to infinity. This illustrated that maximum likelihood estimation is an unbiased estimator. Moreover, the estimate parameters obtained from the application of maximum likelihood estimation have smallest variance as compared to others statistical method as the sample sizes increases. Thus, maximum likelihood estimation is adopted in this paper to fit the two-component mixture model in order to explore the relationship between rubber price and exchange rate for Malaysia, Thailand, Philippines and Indonesia. Results described that there is a negative effect among rubber price and exchange rate for all selected countries.

9. Moral Identity Predicts Doping Likelihood via Moral Disengagement and Anticipated Guilt.

Science.gov (United States)

Kavussanu, Maria; Ring, Christopher

2017-08-01

In this study, we integrated elements of social cognitive theory of moral thought and action and the social cognitive model of moral identity to better understand doping likelihood in athletes. Participants (N = 398) recruited from a variety of team sports completed measures of moral identity, moral disengagement, anticipated guilt, and doping likelihood. Moral identity predicted doping likelihood indirectly via moral disengagement and anticipated guilt. Anticipated guilt about potential doping mediated the relationship between moral disengagement and doping likelihood. Our findings provide novel evidence to suggest that athletes, who feel that being a moral person is central to their self-concept, are less likely to use banned substances due to their lower tendency to morally disengage and the more intense feelings of guilt they expect to experience for using banned substances.

10. Statistical basis for positive identification in forensic anthropology.

Science.gov (United States)

2006-09-01

Forensic scientists are often expected to present the likelihood of DNA identifications in US courts based on comparative population data, yet forensic anthropologists tend not to quantify the strength of an osteological identification. Because forensic anthropologists are trained first and foremost as physical anthropologists, they emphasize estimation problems at the expense of evidentiary problems, but this approach must be reexamined. In this paper, the statistical bases for presenting osteological and dental evidence are outlined, using a forensic case as a motivating example. A brief overview of Bayesian statistics is provided, and methods to calculate likelihood ratios for five aspects of the biological profile are demonstrated. This paper emphasizes the definition of appropriate reference samples and of the "population at large," and points out the conceptual differences between them. Several databases are introduced for both reference information and to characterize the "population at large," and new data are compiled to calculate the frequency of specific characters, such as age or fractures, within the "population at large." Despite small individual likelihood ratios for age, sex, and stature in the case example, the power of this approach is that, assuming each likelihood ratio is independent, the product rule can be applied. In this particular example, it is over three million times more likely to obtain the observed osteological and dental data if the identification is correct than if the identification is incorrect. This likelihood ratio is a convincing statistic that can support the forensic anthropologist's opinion on personal identity in court. 2006 Wiley-Liss, Inc.

11. Quasi-Maximum Likelihood Estimation and Bootstrap Inference in Fractional Time Series Models with Heteroskedasticity of Unknown Form

DEFF Research Database (Denmark)

Cavaliere, Giuseppe; Nielsen, Morten Ørregaard; Taylor, Robert

We consider the problem of conducting estimation and inference on the parameters of univariate heteroskedastic fractionally integrated time series models. We first extend existing results in the literature, developed for conditional sum-of squares estimators in the context of parametric fractional...... time series models driven by conditionally homoskedastic shocks, to allow for conditional and unconditional heteroskedasticity both of a quite general and unknown form. Global consistency and asymptotic normality are shown to still obtain; however, the covariance matrix of the limiting distribution...... of the estimator now depends on nuisance parameters derived both from the weak dependence and heteroskedasticity present in the shocks. We then investigate classical methods of inference based on the Wald, likelihood ratio and Lagrange multiplier tests for linear hypotheses on either or both of the long and short...

12. Maximum Likelihood Approach for RFID Tag Set Cardinality Estimation with Detection Errors

DEFF Research Database (Denmark)

Nguyen, Chuyen T.; Hayashi, Kazunori; Kaneko, Megumi

2013-01-01

Abstract Estimation schemes of Radio Frequency IDentification (RFID) tag set cardinality are studied in this paper using Maximum Likelihood (ML) approach. We consider the estimation problem under the model of multiple independent reader sessions with detection errors due to unreliable radio...... is evaluated under dierent system parameters and compared with that of the conventional method via computer simulations assuming flat Rayleigh fading environments and framed-slotted ALOHA based protocol. Keywords RFID tag cardinality estimation maximum likelihood detection error...

13. Modified Moment, Maximum Likelihood and Percentile Estimators for the Parameters of the Power Function Distribution

Directory of Open Access Journals (Sweden)

Azam Zaka

2014-10-01

Full Text Available This paper is concerned with the modifications of maximum likelihood, moments and percentile estimators of the two parameter Power function distribution. Sampling behavior of the estimators is indicated by Monte Carlo simulation. For some combinations of parameter values, some of the modified estimators appear better than the traditional maximum likelihood, moments and percentile estimators with respect to bias, mean square error and total deviation.

14. Practical Statistics for LHC Physicists: Descriptive Statistics, Probability and Likelihood (1/3)

CERN Multimedia

CERN. Geneva

2015-01-01

These lectures cover those principles and practices of statistics that are most relevant for work at the LHC. The first lecture discusses the basic ideas of descriptive statistics, probability and likelihood. The second lecture covers the key ideas in the frequentist approach, including confidence limits, profile likelihoods, p-values, and hypothesis testing. The third lecture covers inference in the Bayesian approach. Throughout, real-world examples will be used to illustrate the practical application of the ideas. No previous knowledge is assumed.

15. Anticipating cognitive effort: roles of perceived error-likelihood and time demands.

Science.gov (United States)

Dunn, Timothy L; Inzlicht, Michael; Risko, Evan F

2017-11-13

Why are some actions evaluated as effortful? In the present set of experiments we address this question by examining individuals' perception of effort when faced with a trade-off between two putative cognitive costs: how much time a task takes vs. how error-prone it is. Specifically, we were interested in whether individuals anticipate engaging in a small amount of hard work (i.e., low time requirement, but high error-likelihood) vs. a large amount of easy work (i.e., high time requirement, but low error-likelihood) as being more effortful. In between-subject designs, Experiments 1 through 3 demonstrated that individuals anticipate options that are high in perceived error-likelihood (yet less time consuming) as more effortful than options that are perceived to be more time consuming (yet low in error-likelihood). Further, when asked to evaluate which of the two tasks was (a) more effortful, (b) more error-prone, and (c) more time consuming, effort-based and error-based choices closely tracked one another, but this was not the case for time-based choices. Utilizing a within-subject design, Experiment 4 demonstrated overall similar pattern of judgments as Experiments 1 through 3. However, both judgments of error-likelihood and time demand similarly predicted effort judgments. Results are discussed within the context of extant accounts of cognitive control, with considerations of how error-likelihood and time demands may independently and conjunctively factor into judgments of cognitive effort.

16. Peak power ratio generator

Science.gov (United States)

Moyer, R.D.

A peak power ratio generator is described for measuring, in combination with a conventional power meter, the peak power level of extremely narrow pulses in the gigahertz radio frequency bands. The present invention in a preferred embodiment utilizes a tunnel diode and a back diode combination in a detector circuit as the only high speed elements. The high speed tunnel diode provides a bistable signal and serves as a memory device of the input pulses for the remaining, slower components. A hybrid digital and analog loop maintains the peak power level of a reference channel at a known amount. Thus, by measuring the average power levels of the reference signal and the source signal, the peak power level of the source signal can be determined.

17. Researcher positioning

DEFF Research Database (Denmark)

Mørck, Line Lerche; Khawaja, Iram

2009-01-01

abstract  This article focuses on the complex and multi-layered process of researcher positioning, specifically in relation to the politically sensitive study of marginalised and ‘othered' groups such as Muslims living in Denmark. We discuss the impact of different ethnic, religious and racial...... political and personal involvement by the researcher, which challenges traditional perspectives on research and researcher positioning. A key point in this regard is the importance of constant awareness of and reflection on the multiple ways in which one's positioning as a researcher influences the research...

International Nuclear Information System (INIS)

Eisenberg, R.L.; Dennis, C.A.; May, C.

1989-01-01

This book concentrates on the routine radiographic examinations commonly performed. It details the wide variety of examinations possible and their place in initial learning and in the radiology department as references for those occasions when an unusual examination is requested. This book provides information ranging from basic terminology to skeletal positioning to special procedures. Positions are discussed and supplemented with a picture of a patient, the resulting radiograph, and a labeled diagram. Immobilization and proper shielding of the patient are also shown

19. Position encoder

International Nuclear Information System (INIS)

Goursky, Vsevolod

1975-01-01

A circuitry for deriving the quotient of signal delivered by position-sensitive detectors is described. Digital output is obtained in the form of 10- to 12-bit words. Impact position may be determined with 0.25% accuracy when the dynamic range of the energy signal is less 1:10, and 0.5% accuracy when the dynamic range is 1:20. The division requires an average time of 5μs for 10-bit words

20. Position encoder

International Nuclear Information System (INIS)

Goursky, V.

1975-05-01

This paper describes circuitry for deriving the quotient of signals delivered by position-sensitive detectors. Digital output is obtained in the form of 10 to 12 bit words. Impact position may be determined with 0.25% accuracy when the dynamic range of the energy signal is less than 1:10, and 0.5% accuracy when the dynamic range is 1:20. The division requires an average time of 5μs for 10-bit words [fr