WorldWideScience

Sample records for high counting statistics

  1. Radiation counting statistics

    Energy Technology Data Exchange (ETDEWEB)

    Suh, M. Y.; Jee, K. Y.; Park, K. K.; Park, Y. J.; Kim, W. H

    1999-08-01

    This report is intended to describe the statistical methods necessary to design and conduct radiation counting experiments and evaluate the data from the experiment. The methods are described for the evaluation of the stability of a counting system and the estimation of the precision of counting data by application of probability distribution models. The methods for the determination of the uncertainty of the results calculated from the number of counts, as well as various statistical methods for the reduction of counting error are also described. (Author). 11 refs., 8 tabs., 8 figs.

  2. Radiation counting statistics

    Energy Technology Data Exchange (ETDEWEB)

    Suh, M. Y.; Jee, K. Y.; Park, K. K. [Korea Atomic Energy Research Institute, Taejon (Korea)

    1999-08-01

    This report is intended to describe the statistical methods necessary to design and conduct radiation counting experiments and evaluate the data from the experiments. The methods are described for the evaluation of the stability of a counting system and the estimation of the precision of counting data by application of probability distribution models. The methods for the determination of the uncertainty of the results calculated from the number of counts, as well as various statistical methods for the reduction of counting error are also described. 11 refs., 6 figs., 8 tabs. (Author)

  3. Radiation counting statistics

    International Nuclear Information System (INIS)

    Suh, M. Y.; Jee, K. Y.; Park, K. K.; Park, Y. J.; Kim, W. H.

    1999-08-01

    This report is intended to describe the statistical methods necessary to design and conduct radiation counting experiments and evaluate the data from the experiment. The methods are described for the evaluation of the stability of a counting system and the estimation of the precision of counting data by application of probability distribution models. The methods for the determination of the uncertainty of the results calculated from the number of counts, as well as various statistical methods for the reduction of counting error are also described. (Author). 11 refs., 8 tabs., 8 figs

  4. Counting statistics in radioactivity measurements

    International Nuclear Information System (INIS)

    Martin, J.

    1975-01-01

    The application of statistical methods to radioactivity measurement problems is analyzed in several chapters devoted successively to: the statistical nature of radioactivity counts; the application to radioactive counting of two theoretical probability distributions, Poisson's distribution law and the Laplace-Gauss law; true counting laws; corrections related to the nature of the apparatus; statistical techniques in gamma spectrometry [fr

  5. Full counting statistics of multiple Andreev reflections in incoherent diffusive superconducting junctions

    International Nuclear Information System (INIS)

    Samuelsson, P.

    2007-01-01

    We present a theory for the full distribution of current fluctuations in incoherent diffusive superconducting junctions, subjected to a voltage bias. This theory of full counting statistics of incoherent multiple Andreev reflections is valid for an arbitrary applied voltage. We present a detailed discussion of the properties of the first four cumulants as well as the low and high voltage regimes of the full counting statistics. (orig.)

  6. Radiation measurement practice for understanding statistical fluctuation of radiation count using natural radiation sources

    International Nuclear Information System (INIS)

    Kawano, Takao

    2014-01-01

    It is known that radiation is detected at random and the radiation counts fluctuate statistically. In the present study, a radiation measurement experiment was performed to understand the randomness and statistical fluctuation of radiation counts. In the measurement, three natural radiation sources were used. The sources were fabricated from potassium chloride chemicals, chemical fertilizers and kelps. These materials contain naturally occurring potassium-40 that is a radionuclide. From high schools, junior high schools and elementary schools, nine teachers participated to the radiation measurement experiment. Each participant measured the 1-min integration counts of radiation five times using GM survey meters, and 45 sets of data were obtained for the respective natural radiation sources. It was found that the frequency of occurrence of radiation counts was distributed according to a Gaussian distribution curve, although the obtained 45 data sets of radiation counts superficially looked to be fluctuating meaninglessly. (author)

  7. Statistical tests to compare motif count exceptionalities

    Directory of Open Access Journals (Sweden)

    Vandewalle Vincent

    2007-03-01

    Full Text Available Abstract Background Finding over- or under-represented motifs in biological sequences is now a common task in genomics. Thanks to p-value calculation for motif counts, exceptional motifs are identified and represent candidate functional motifs. The present work addresses the related question of comparing the exceptionality of one motif in two different sequences. Just comparing the motif count p-values in each sequence is indeed not sufficient to decide if this motif is significantly more exceptional in one sequence compared to the other one. A statistical test is required. Results We develop and analyze two statistical tests, an exact binomial one and an asymptotic likelihood ratio test, to decide whether the exceptionality of a given motif is equivalent or significantly different in two sequences of interest. For that purpose, motif occurrences are modeled by Poisson processes, with a special care for overlapping motifs. Both tests can take the sequence compositions into account. As an illustration, we compare the octamer exceptionalities in the Escherichia coli K-12 backbone versus variable strain-specific loops. Conclusion The exact binomial test is particularly adapted for small counts. For large counts, we advise to use the likelihood ratio test which is asymptotic but strongly correlated with the exact binomial test and very simple to use.

  8. Statistical Methods for Unusual Count Data

    DEFF Research Database (Denmark)

    Guthrie, Katherine A.; Gammill, Hilary S.; Kamper-Jørgensen, Mads

    2016-01-01

    microchimerism data present challenges for statistical analysis, including a skewed distribution, excess zero values, and occasional large values. Methods for comparing microchimerism levels across groups while controlling for covariates are not well established. We compared statistical models for quantitative...... microchimerism values, applied to simulated data sets and 2 observed data sets, to make recommendations for analytic practice. Modeling the level of quantitative microchimerism as a rate via Poisson or negative binomial model with the rate of detection defined as a count of microchimerism genome equivalents per...

  9. Statistical data filtration in neutron coincidence counting

    International Nuclear Information System (INIS)

    Beddingfield, D.H.; Menlove, H.O.

    1992-11-01

    We assessed the effectiveness of statistical data filtration to minimize the contribution of matrix materials in 200-ell drums to the nondestructive assay of plutonium. Those matrices were examined: polyethylene, concrete, aluminum, iron, cadmium, and lead. Statistical filtration of neutron coincidence data improved the low-end sensitivity of coincidence counters. Spurious data arising from electrical noise, matrix spallation, and geometric effects were smoothed in a predictable fashion by the statistical filter. The filter effectively lowers the minimum detectable mass limit that can be achieved for plutonium assay using passive neutron coincidence counting

  10. Binomial distribution of Poisson statistics and tracks overlapping probability to estimate total tracks count with low uncertainty

    International Nuclear Information System (INIS)

    Khayat, Omid; Afarideh, Hossein; Mohammadnia, Meisam

    2015-01-01

    In the solid state nuclear track detectors of chemically etched type, nuclear tracks with center-to-center neighborhood of distance shorter than two times the radius of tracks will emerge as overlapping tracks. Track overlapping in this type of detectors causes tracks count losses and it becomes rather severe in high track densities. Therefore, tracks counting in this condition should include a correction factor for count losses of different tracks overlapping orders since a number of overlapping tracks may be counted as one track. Another aspect of the problem is the cases where imaging the whole area of the detector and counting all tracks are not possible. In these conditions a statistical generalization method is desired to be applicable in counting a segmented area of the detector and the results can be generalized to the whole surface of the detector. Also there is a challenge in counting the tracks in densely overlapped tracks because not sufficient geometrical or contextual information are available. It this paper we present a statistical counting method which gives the user a relation between the tracks overlapping probabilities on a segmented area of the detector surface and the total number of tracks. To apply the proposed method one can estimate the total number of tracks on a solid state detector of arbitrary shape and dimensions by approximating the tracks averaged area, whole detector surface area and some orders of tracks overlapping probabilities. It will be shown that this method is applicable in high and ultra high density tracks images and the count loss error can be enervated using a statistical generalization approach. - Highlights: • A correction factor for count losses of different tracks overlapping orders. • For the cases imaging the whole area of the detector is not possible. • Presenting a statistical generalization method for segmented areas. • Giving a relation between the tracks overlapping probabilities and the total tracks

  11. Theory of overdispersion in counting statistics caused by fluctuating probabilities

    International Nuclear Information System (INIS)

    Semkow, Thomas M.

    1999-01-01

    It is shown that the random Lexis fluctuations of probabilities such as probability of decay or detection cause the counting statistics to be overdispersed with respect to the classical binomial, Poisson, or Gaussian distributions. The generating and the distribution functions for the overdispersed counting statistics are derived. Applications to radioactive decay with detection and more complex experiments are given, as well as distinguishing between the source and background, in the presence of overdispersion. Monte-Carlo verifications are provided

  12. Hybrid statistics-simulations based method for atom-counting from ADF STEM images

    Energy Technology Data Exchange (ETDEWEB)

    De wael, Annelies, E-mail: annelies.dewael@uantwerpen.be [Electron Microscopy for Materials Science (EMAT), University of Antwerp, Groenenborgerlaan 171, 2020 Antwerp (Belgium); De Backer, Annick [Electron Microscopy for Materials Science (EMAT), University of Antwerp, Groenenborgerlaan 171, 2020 Antwerp (Belgium); Jones, Lewys; Nellist, Peter D. [Department of Materials, University of Oxford, Parks Road, OX1 3PH Oxford (United Kingdom); Van Aert, Sandra, E-mail: sandra.vanaert@uantwerpen.be [Electron Microscopy for Materials Science (EMAT), University of Antwerp, Groenenborgerlaan 171, 2020 Antwerp (Belgium)

    2017-06-15

    A hybrid statistics-simulations based method for atom-counting from annular dark field scanning transmission electron microscopy (ADF STEM) images of monotype crystalline nanostructures is presented. Different atom-counting methods already exist for model-like systems. However, the increasing relevance of radiation damage in the study of nanostructures demands a method that allows atom-counting from low dose images with a low signal-to-noise ratio. Therefore, the hybrid method directly includes prior knowledge from image simulations into the existing statistics-based method for atom-counting, and accounts in this manner for possible discrepancies between actual and simulated experimental conditions. It is shown by means of simulations and experiments that this hybrid method outperforms the statistics-based method, especially for low electron doses and small nanoparticles. The analysis of a simulated low dose image of a small nanoparticle suggests that this method allows for far more reliable quantitative analysis of beam-sensitive materials. - Highlights: • A hybrid method for atom-counting from ADF STEM images is introduced. • Image simulations are incorporated into a statistical framework in a reliable manner. • Limits of the existing methods for atom-counting are far exceeded. • Reliable counting results from an experimental low dose image are obtained. • Progress towards reliable quantitative analysis of beam-sensitive materials is made.

  13. Counting statistics of transport through Coulomb blockade nanostructures: High-order cumulants and non-Markovian effects

    DEFF Research Database (Denmark)

    Flindt, Christian; Novotny, Tomás; Braggio, Alessandro

    2010-01-01

    Recent experimental progress has made it possible to detect in real-time single electrons tunneling through Coulomb blockade nanostructures, thereby allowing for precise measurements of the statistical distribution of the number of transferred charges, the so-called full counting statistics...... interactions. Our recursive method can treat systems with many states as well as non-Markovian dynamics. We illustrate our approach with three examples of current experimental relevance: bunching transport through a two-level quantum dot, transport through a nanoelectromechanical system with dynamical Franck...

  14. Counting statistics and loss corrections for the APS

    International Nuclear Information System (INIS)

    Lee, W.K.; Mills, D.M.

    1992-01-01

    It has been suggested that for timing experiments, it might be advantageous to arrange the bunches in the storage ring in an asymmetrical mode. In this paper, we determine the counting losses from pulsed x-ray sources from basic probabilistic arguments and from Poisson statistics. In particular the impact on single photon counting losses of a variety of possible filling modes for the Advanced Photon Source (APS) is examined. For bunches of equal current, a loss of 10% occurs whenever the count rate exceeds 21% of the bunch repetition rate. This changes slightly when bunches containing unequal numbers of particles are considered. The results are applied to several common detector/electronics systems

  15. Counting statistics and loss corrections for the APS

    International Nuclear Information System (INIS)

    Lee, W.K.; Mills, D.M.

    1992-01-01

    It has been suggested that for timing experiments, it might be advantageous to arrange the bunches in the storage ring in an asymmetrical mode. In this paper, we determine the counting losses from pulsed x-ray sources from basic probabilistic arguments and from Poisson statistics. In particular the impact on single-photon counting losses of a variety of possible filling modes for the Advanced Photon Source (APS) is examined. For bunches of equal current, a loss of 10% occurs whenever the count rate exceeds 21% of the bunch repetition rate. This changes slightly when bunches containing unequal numbers of particles are considered. The results are applied to several common detector/electronics systems

  16. Counting in Lattices: Combinatorial Problems from Statistical Mechanics.

    Science.gov (United States)

    Randall, Dana Jill

    In this thesis we consider two classical combinatorial problems arising in statistical mechanics: counting matchings and self-avoiding walks in lattice graphs. The first problem arises in the study of the thermodynamical properties of monomers and dimers (diatomic molecules) in crystals. Fisher, Kasteleyn and Temperley discovered an elegant technique to exactly count the number of perfect matchings in two dimensional lattices, but it is not applicable for matchings of arbitrary size, or in higher dimensional lattices. We present the first efficient approximation algorithm for computing the number of matchings of any size in any periodic lattice in arbitrary dimension. The algorithm is based on Monte Carlo simulation of a suitable Markov chain and has rigorously derived performance guarantees that do not rely on any assumptions. In addition, we show that these results generalize to counting matchings in any graph which is the Cayley graph of a finite group. The second problem is counting self-avoiding walks in lattices. This problem arises in the study of the thermodynamics of long polymer chains in dilute solution. While there are a number of Monte Carlo algorithms used to count self -avoiding walks in practice, these are heuristic and their correctness relies on unproven conjectures. In contrast, we present an efficient algorithm which relies on a single, widely-believed conjecture that is simpler than preceding assumptions and, more importantly, is one which the algorithm itself can test. Thus our algorithm is reliable, in the sense that it either outputs answers that are guaranteed, with high probability, to be correct, or finds a counterexample to the conjecture. In either case we know we can trust our results and the algorithm is guaranteed to run in polynomial time. This is the first algorithm for counting self-avoiding walks in which the error bounds are rigorously controlled. This work was supported in part by an AT&T graduate fellowship, a University of

  17. Counting statistics of many-particle quantum walks

    Science.gov (United States)

    Mayer, Klaus; Tichy, Malte C.; Mintert, Florian; Konrad, Thomas; Buchleitner, Andreas

    2011-06-01

    We study quantum walks of many noninteracting particles on a beam splitter array as a paradigmatic testing ground for the competition of single- and many-particle interference in a multimode system. We derive a general expression for multimode particle-number correlation functions, valid for bosons and fermions, and infer pronounced signatures of many-particle interferences in the counting statistics.

  18. Counting statistics of many-particle quantum walks

    International Nuclear Information System (INIS)

    Mayer, Klaus; Tichy, Malte C.; Buchleitner, Andreas; Mintert, Florian; Konrad, Thomas

    2011-01-01

    We study quantum walks of many noninteracting particles on a beam splitter array as a paradigmatic testing ground for the competition of single- and many-particle interference in a multimode system. We derive a general expression for multimode particle-number correlation functions, valid for bosons and fermions, and infer pronounced signatures of many-particle interferences in the counting statistics.

  19. The intensity detection of single-photon detectors based on photon counting probability density statistics

    International Nuclear Information System (INIS)

    Zhang Zijing; Song Jie; Zhao Yuan; Wu Long

    2017-01-01

    Single-photon detectors possess the ultra-high sensitivity, but they cannot directly respond to signal intensity. Conventional methods adopt sampling gates with fixed width and count the triggered number of sampling gates, which is capable of obtaining photon counting probability to estimate the echo signal intensity. In this paper, we not only count the number of triggered sampling gates, but also record the triggered time position of photon counting pulses. The photon counting probability density distribution is obtained through the statistics of a series of the triggered time positions. Then Minimum Variance Unbiased Estimation (MVUE) method is used to estimate the echo signal intensity. Compared with conventional methods, this method can improve the estimation accuracy of echo signal intensity due to the acquisition of more detected information. Finally, a proof-of-principle laboratory system is established. The estimation accuracy of echo signal intensity is discussed and a high accuracy intensity image is acquired under low-light level environments. (paper)

  20. Hybrid statistics-simulations based method for atom-counting from ADF STEM images.

    Science.gov (United States)

    De Wael, Annelies; De Backer, Annick; Jones, Lewys; Nellist, Peter D; Van Aert, Sandra

    2017-06-01

    A hybrid statistics-simulations based method for atom-counting from annular dark field scanning transmission electron microscopy (ADF STEM) images of monotype crystalline nanostructures is presented. Different atom-counting methods already exist for model-like systems. However, the increasing relevance of radiation damage in the study of nanostructures demands a method that allows atom-counting from low dose images with a low signal-to-noise ratio. Therefore, the hybrid method directly includes prior knowledge from image simulations into the existing statistics-based method for atom-counting, and accounts in this manner for possible discrepancies between actual and simulated experimental conditions. It is shown by means of simulations and experiments that this hybrid method outperforms the statistics-based method, especially for low electron doses and small nanoparticles. The analysis of a simulated low dose image of a small nanoparticle suggests that this method allows for far more reliable quantitative analysis of beam-sensitive materials. Copyright © 2017 Elsevier B.V. All rights reserved.

  1. Full Counting Statistics for Interacting Fermions with Determinantal Quantum Monte Carlo Simulations.

    Science.gov (United States)

    Humeniuk, Stephan; Büchler, Hans Peter

    2017-12-08

    We present a method for computing the full probability distribution function of quadratic observables such as particle number or magnetization for the Fermi-Hubbard model within the framework of determinantal quantum Monte Carlo calculations. Especially in cold atom experiments with single-site resolution, such a full counting statistics can be obtained from repeated projective measurements. We demonstrate that the full counting statistics can provide important information on the size of preformed pairs. Furthermore, we compute the full counting statistics of the staggered magnetization in the repulsive Hubbard model at half filling and find excellent agreement with recent experimental results. We show that current experiments are capable of probing the difference between the Hubbard model and the limiting Heisenberg model.

  2. Statistical method for resolving the photon-photoelectron-counting inversion problem

    International Nuclear Information System (INIS)

    Wu Jinlong; Li Tiejun; Peng, Xiang; Guo Hong

    2011-01-01

    A statistical inversion method is proposed for the photon-photoelectron-counting statistics in quantum key distribution experiment. With the statistical viewpoint, this problem is equivalent to the parameter estimation for an infinite binomial mixture model. The coarse-graining idea and Bayesian methods are applied to deal with this ill-posed problem, which is a good simple example to show the successful application of the statistical methods to the inverse problem. Numerical results show the applicability of the proposed strategy. The coarse-graining idea for the infinite mixture models should be general to be used in the future.

  3. Non-Poisson counting statistics of a hybrid G-M counter dead time model

    International Nuclear Information System (INIS)

    Lee, Sang Hoon; Jae, Moosung; Gardner, Robin P.

    2007-01-01

    The counting statistics of a G-M counter with a considerable dead time event rate deviates from Poisson statistics. Important characteristics such as observed counting rates as a function true counting rates, variances and interval distributions were analyzed for three dead time models, non-paralyzable, paralyzable and hybrid, with the help of GMSIM, a Monte Carlo dead time effect simulator. The simulation results showed good agreements with the models in observed counting rates and variances. It was found through GMSIM simulations that the interval distribution for the hybrid model showed three distinctive regions, a complete cutoff region for the duration of the total dead time, a degraded exponential and an enhanced exponential regions. By measuring the cutoff and the duration of degraded exponential from the pulse interval distribution, it is possible to evaluate the two dead times in the hybrid model

  4. Counting statistics in low level radioactivity measurements fluctuating counting efficiency

    International Nuclear Information System (INIS)

    Pazdur, M.F.

    1976-01-01

    A divergence between the probability distribution of the number of nuclear disintegrations and the number of observed counts, caused by counting efficiency fluctuation, is discussed. The negative binominal distribution is proposed to describe the probability distribution of the number of counts, instead of Poisson distribution, which is assumed to hold for the number of nuclear disintegrations only. From actual measurements the r.m.s. amplitude of counting efficiency fluctuation is estimated. Some consequences of counting efficiency fluctuation are investigated and the corresponding formulae are derived: (1) for detection limit as a function of the number of partial measurements and the relative amplitude of counting efficiency fluctuation, and (2) for optimum allocation of the number of partial measurements between sample and background. (author)

  5. Experimental investigation of statistical models describing distribution of counts

    International Nuclear Information System (INIS)

    Salma, I.; Zemplen-Papp, E.

    1992-01-01

    The binomial, Poisson and modified Poisson models which are used for describing the statistical nature of the distribution of counts are compared theoretically, and conclusions for application are considered. The validity of the Poisson and the modified Poisson statistical distribution for observing k events in a short time interval is investigated experimentally for various measuring times. The experiments to measure the influence of the significant radioactive decay were performed with 89 Y m (T 1/2 =16.06 s), using a multichannel analyser (4096 channels) in the multiscaling mode. According to the results, Poisson statistics describe the counting experiment for short measuring times (up to T=0.5T 1/2 ) and its application is recommended. However, analysis of the data demonstrated, with confidence, that for long measurements (T≥T 1/2 ) Poisson distribution is not valid and the modified Poisson function is preferable. The practical implications in calculating uncertainties and in optimizing the measuring time are discussed. Differences between the standard deviations evaluated on the basis of the Poisson and binomial models are especially significant for experiments with long measuring time (T/T 1/2 ≥2) and/or large detection efficiency (ε>0.30). Optimization of the measuring time for paired observations yields the same solution for either the binomial or the Poisson distribution. (orig.)

  6. Spatial statistics of pitting corrosion patterning: Quadrat counts and the non-homogeneous Poisson process

    International Nuclear Information System (INIS)

    Lopez de la Cruz, J.; Gutierrez, M.A.

    2008-01-01

    This paper presents a stochastic analysis of spatial point patterns as effect of localized pitting corrosion. The Quadrat Counts method is studied with two empirical pit patterns. The results are dependent on the quadrat size and bias is introduced when empty quadrats are accounted for the analysis. The spatially inhomogeneous Poisson process is used to improve the performance of the Quadrat Counts method. The latter combines Quadrat Counts with distance-based statistics in the analysis of pit patterns. The Inter-Event and the Nearest-Neighbour statistics are here implemented in order to compare their results. Further, the treatment of patterns in irregular domains is discussed

  7. Reducing bias in the analysis of counting statistics data

    International Nuclear Information System (INIS)

    Hammersley, A.P.; Antoniadis, A.

    1997-01-01

    In the analysis of counting statistics data it is common practice to estimate the variance of the measured data points as the data points themselves. This practice introduces a bias into the results of further analysis which may be significant, and under certain circumstances lead to false conclusions. In the case of normal weighted least squares fitting this bias is quantified and methods to avoid it are proposed. (orig.)

  8. Gene coexpression measures in large heterogeneous samples using count statistics.

    Science.gov (United States)

    Wang, Y X Rachel; Waterman, Michael S; Huang, Haiyan

    2014-11-18

    With the advent of high-throughput technologies making large-scale gene expression data readily available, developing appropriate computational tools to process these data and distill insights into systems biology has been an important part of the "big data" challenge. Gene coexpression is one of the earliest techniques developed that is still widely in use for functional annotation, pathway analysis, and, most importantly, the reconstruction of gene regulatory networks, based on gene expression data. However, most coexpression measures do not specifically account for local features in expression profiles. For example, it is very likely that the patterns of gene association may change or only exist in a subset of the samples, especially when the samples are pooled from a range of experiments. We propose two new gene coexpression statistics based on counting local patterns of gene expression ranks to take into account the potentially diverse nature of gene interactions. In particular, one of our statistics is designed for time-course data with local dependence structures, such as time series coupled over a subregion of the time domain. We provide asymptotic analysis of their distributions and power, and evaluate their performance against a wide range of existing coexpression measures on simulated and real data. Our new statistics are fast to compute, robust against outliers, and show comparable and often better general performance.

  9. High Reproducibility of ELISPOT Counts from Nine Different Laboratories

    Directory of Open Access Journals (Sweden)

    Srividya Sundararaman

    2015-01-01

    Full Text Available The primary goal of immune monitoring with ELISPOT is to measure the number of T cells, specific for any antigen, accurately and reproducibly between different laboratories. In ELISPOT assays, antigen-specific T cells secrete cytokines, forming spots of different sizes on a membrane with variable background intensities. Due to the subjective nature of judging maximal and minimal spot sizes, different investigators come up with different numbers. This study aims to determine whether statistics-based, automated size-gating can harmonize the number of spot counts calculated between different laboratories. We plated PBMC at four different concentrations, 24 replicates each, in an IFN-γ ELISPOT assay with HCMV pp65 antigen. The ELISPOT plate, and an image file of the plate was counted in nine different laboratories using ImmunoSpot® Analyzers by (A Basic Count™ relying on subjective counting parameters set by the respective investigators and (B SmartCount™, an automated counting protocol by the ImmunoSpot® Software that uses statistics-based spot size auto-gating with spot intensity auto-thresholding. The average coefficient of variation (CV for the mean values between independent laboratories was 26.7% when counting with Basic Count™, and 6.7% when counting with SmartCount™. Our data indicates that SmartCount™ allows harmonization of counting ELISPOT results between different laboratories and investigators.

  10. Radon counting statistics - a Monte Carlo investigation

    International Nuclear Information System (INIS)

    Scott, A.G.

    1996-01-01

    Radioactive decay is a Poisson process, and so the Coefficient of Variation (COV) of open-quotes nclose quotes counts of a single nuclide is usually estimated as 1/√n. This is only true if the count duration is much shorter than the half-life of the nuclide. At longer count durations, the COV is smaller than the Poisson estimate. Most radon measurement methods count the alpha decays of 222 Rn, plus the progeny 218 Po and 214 Po, and estimate the 222 Rn activity from the sum of the counts. At long count durations, the chain decay of these nuclides means that every 222 Rn decay must be followed by two other alpha decays. The total number of decays is open-quotes 3Nclose quotes, where N is the number of radon decays, and the true COV of the radon concentration estimate is 1/√(N), √3 larger than the Poisson total count estimate of 1/√3N. Most count periods are comparable to the half lives of the progeny, so the relationship between COV and count time is complex. A Monte-Carlo estimate of the ratio of true COV to Poisson estimate was carried out for a range of count periods from 1 min to 16 h and three common radon measurement methods: liquid scintillation, scintillation cell, and electrostatic precipitation of progeny. The Poisson approximation underestimates COV by less than 20% for count durations of less than 60 min

  11. Selecting the right statistical model for analysis of insect count data by using information theoretic measures.

    Science.gov (United States)

    Sileshi, G

    2006-10-01

    Researchers and regulatory agencies often make statistical inferences from insect count data using modelling approaches that assume homogeneous variance. Such models do not allow for formal appraisal of variability which in its different forms is the subject of interest in ecology. Therefore, the objectives of this paper were to (i) compare models suitable for handling variance heterogeneity and (ii) select optimal models to ensure valid statistical inferences from insect count data. The log-normal, standard Poisson, Poisson corrected for overdispersion, zero-inflated Poisson, the negative binomial distribution and zero-inflated negative binomial models were compared using six count datasets on foliage-dwelling insects and five families of soil-dwelling insects. Akaike's and Schwarz Bayesian information criteria were used for comparing the various models. Over 50% of the counts were zeros even in locally abundant species such as Ootheca bennigseni Weise, Mesoplatys ochroptera Stål and Diaecoderus spp. The Poisson model after correction for overdispersion and the standard negative binomial distribution model provided better description of the probability distribution of seven out of the 11 insects than the log-normal, standard Poisson, zero-inflated Poisson or zero-inflated negative binomial models. It is concluded that excess zeros and variance heterogeneity are common data phenomena in insect counts. If not properly modelled, these properties can invalidate the normal distribution assumptions resulting in biased estimation of ecological effects and jeopardizing the integrity of the scientific inferences. Therefore, it is recommended that statistical models appropriate for handling these data properties be selected using objective criteria to ensure efficient statistical inference.

  12. Application of statistical methods to the testing of nuclear counting assemblies

    International Nuclear Information System (INIS)

    Gilbert, J.P.; Friedling, G.

    1965-01-01

    This report describes the application of the hypothesis test theory to the control of the 'statistical purity' and of the stability of the counting batteries used for measurements on activation detectors in research reactors. The principles involved and the experimental results obtained at Cadarache on batteries operating with the reactors PEGGY and AZUR are given. (authors) [fr

  13. PREFACE: Counting Complexity: An international workshop on statistical mechanics and combinatorics

    Science.gov (United States)

    de Gier, Jan; Warnaar, Ole

    2006-07-01

    On 10-15 July 2005 the conference `Counting Complexity: An international workshop on statistical mechanics and combinatorics' was held on Dunk Island, Queensland, Australia in celebration of Tony Guttmann's 60th birthday. Dunk Island provided the perfect setting for engaging in almost all of Tony's life-long passions: swimming, running, food, wine and, of course, plenty of mathematics and physics. The conference was attended by many of Tony's close scientific friends from all over the world, and most talks were presented by his past and present collaborators. This volume contains the proceedings of the meeting and consists of 24 refereed research papers in the fields of statistical mechanics, condensed matter physics and combinatorics. These papers provide an excellent illustration of the breadth and scope of Tony's work. The very first contribution, written by Stu Whittington, contains an overview of the many scientific achievements of Tony over the past 40 years in mathematics and physics. The organizing committee, consisting of Richard Brak, Aleks Owczarek, Jan de Gier, Emma Lockwood, Andrew Rechnitzer and Ole Warnaar, gratefully acknowledges the Australian Mathematical Society (AustMS), the Australian Mathematical Sciences Institute (AMSI), the ARC Centre of Excellence for Mathematics and Statistics of Complex Systems (MASCOS), the ARC Complex Open Systems Research Network (COSNet), the Institute of Physics (IOP) and the Department of Mathematics and Statistics of The University of Melbourne for financial support in organizing the conference. Tony, we hope that your future years in mathematics will be numerous. Count yourself lucky! Tony Guttman

  14. Local box-counting dimensions of discrete quantum eigenvalue spectra: Analytical connection to quantum spectral statistics

    Science.gov (United States)

    Sakhr, Jamal; Nieminen, John M.

    2018-03-01

    Two decades ago, Wang and Ong, [Phys. Rev. A 55, 1522 (1997)], 10.1103/PhysRevA.55.1522 hypothesized that the local box-counting dimension of a discrete quantum spectrum should depend exclusively on the nearest-neighbor spacing distribution (NNSD) of the spectrum. In this Rapid Communication, we validate their hypothesis by deriving an explicit formula for the local box-counting dimension of a countably-infinite discrete quantum spectrum. This formula expresses the local box-counting dimension of a spectrum in terms of single and double integrals of the NNSD of the spectrum. As applications, we derive an analytical formula for Poisson spectra and closed-form approximations to the local box-counting dimension for spectra having Gaussian orthogonal ensemble (GOE), Gaussian unitary ensemble (GUE), and Gaussian symplectic ensemble (GSE) spacing statistics. In the Poisson and GOE cases, we compare our theoretical formulas with the published numerical data of Wang and Ong and observe excellent agreement between their data and our theory. We also study numerically the local box-counting dimensions of the Riemann zeta function zeros and the alternate levels of GOE spectra, which are often used as numerical models of spectra possessing GUE and GSE spacing statistics, respectively. In each case, the corresponding theoretical formula is found to accurately describe the numerically computed local box-counting dimension.

  15. Gaussian point count statistics for families of curves over a fixed finite field

    OpenAIRE

    Kurlberg, Par; Wigman, Igor

    2010-01-01

    We produce a collection of families of curves, whose point count statistics over F_p becomes Gaussian for p fixed. In particular, the average number of F_p points on curves in these families tends to infinity.

  16. Polychromatic Iterative Statistical Material Image Reconstruction for Photon-Counting Computed Tomography

    Directory of Open Access Journals (Sweden)

    Thomas Weidinger

    2016-01-01

    Full Text Available This work proposes a dedicated statistical algorithm to perform a direct reconstruction of material-decomposed images from data acquired with photon-counting detectors (PCDs in computed tomography. It is based on local approximations (surrogates of the negative logarithmic Poisson probability function. Exploiting the convexity of this function allows for parallel updates of all image pixels. Parallel updates can compensate for the rather slow convergence that is intrinsic to statistical algorithms. We investigate the accuracy of the algorithm for ideal photon-counting detectors. Complementarily, we apply the algorithm to simulation data of a realistic PCD with its spectral resolution limited by K-escape, charge sharing, and pulse-pileup. For data from both an ideal and realistic PCD, the proposed algorithm is able to correct beam-hardening artifacts and quantitatively determine the material fractions of the chosen basis materials. Via regularization we were able to achieve a reduction of image noise for the realistic PCD that is up to 90% lower compared to material images form a linear, image-based material decomposition using FBP images. Additionally, we find a dependence of the algorithms convergence speed on the threshold selection within the PCD.

  17. Homicides by Police: Comparing Counts From the National Violent Death Reporting System, Vital Statistics, and Supplementary Homicide Reports.

    Science.gov (United States)

    Barber, Catherine; Azrael, Deborah; Cohen, Amy; Miller, Matthew; Thymes, Deonza; Wang, David Enze; Hemenway, David

    2016-05-01

    To evaluate the National Violent Death Reporting System (NVDRS) as a surveillance system for homicides by law enforcement officers. We assessed sensitivity and positive predictive value of the NVDRS "type of death" variable against our study count of homicides by police, which we derived from NVDRS coded and narrative data for states participating in NVDRS 2005 to 2012. We compared state counts of police homicides from NVDRS, Vital Statistics, and Federal Bureau of Investigation Supplementary Homicide Reports. We identified 1552 police homicides in the 16 states. Positive predictive value and sensitivity of the NVDRS "type of death" variable for police homicides were high (98% and 90%, respectively). Counts from Vital Statistics and Supplementary Homicide Reports were 58% and 48%, respectively, of our study total; gaps varied widely by state. The annual rate of police homicide (0.24/100,000) varied 5-fold by state and 8-fold by race/ethnicity. NVDRS provides more complete data on police homicides than do existing systems. Expanding NVDRS to all 50 states and making 2 improvements we identify will be an efficient way to provide the nation with more accurate, detailed data on homicides by law enforcement.

  18. Atom counting in HAADF STEM using a statistical model-based approach: methodology, possibilities, and inherent limitations.

    Science.gov (United States)

    De Backer, A; Martinez, G T; Rosenauer, A; Van Aert, S

    2013-11-01

    In the present paper, a statistical model-based method to count the number of atoms of monotype crystalline nanostructures from high resolution high-angle annular dark-field (HAADF) scanning transmission electron microscopy (STEM) images is discussed in detail together with a thorough study on the possibilities and inherent limitations. In order to count the number of atoms, it is assumed that the total scattered intensity scales with the number of atoms per atom column. These intensities are quantitatively determined using model-based statistical parameter estimation theory. The distribution describing the probability that intensity values are generated by atomic columns containing a specific number of atoms is inferred on the basis of the experimental scattered intensities. Finally, the number of atoms per atom column is quantified using this estimated probability distribution. The number of atom columns available in the observed STEM image, the number of components in the estimated probability distribution, the width of the components of the probability distribution, and the typical shape of a criterion to assess the number of components in the probability distribution directly affect the accuracy and precision with which the number of atoms in a particular atom column can be estimated. It is shown that single atom sensitivity is feasible taking the latter aspects into consideration. © 2013 Elsevier B.V. All rights reserved.

  19. Theory of photoelectron counting statistics

    International Nuclear Information System (INIS)

    Blake, J.

    1980-01-01

    The purpose of the present essay is to provide a detailed analysis of those theoretical aspects of photoelectron counting which are capable of experimental verification. Most of our interest is in the physical phenomena themselves, while part is in the mathematical techniques. Many of the mathematical methods used in the analysis of the photoelectron counting problem are generally unfamiliar to physicists interested in the subject. For this reason we have developed the essay in such a fashion that, although primary interest is focused on the physical phenomena, we have also taken pains to carry out enough of the analysis so that the reader can follow the main details. We have chosen to present a consistently quantum mechanical version of the subject, in that we follow the Glauber theory throughout. (orig./WL)

  20. Statistical measurement of the gamma-ray source-count distribution as a function of energy

    Science.gov (United States)

    Zechlin, H.-S.; Cuoco, A.; Donato, F.; Fornengo, N.; Regis, M.

    2017-01-01

    Photon counts statistics have recently been proven to provide a sensitive observable for characterizing gamma-ray source populations and for measuring the composition of the gamma-ray sky. In this work, we generalize the use of the standard 1-point probability distribution function (1pPDF) to decompose the high-latitude gamma-ray emission observed with Fermi-LAT into: (i) point-source contributions, (ii) the Galactic foreground contribution, and (iii) a diffuse isotropic background contribution. We analyze gamma-ray data in five adjacent energy bands between 1 and 171 GeV. We measure the source-count distribution dN/dS as a function of energy, and demonstrate that our results extend current measurements from source catalogs to the regime of so far undetected sources. Our method improves the sensitivity for resolving point-source populations by about one order of magnitude in flux. The dN/dS distribution as a function of flux is found to be compatible with a broken power law. We derive upper limits on further possible breaks as well as the angular power of unresolved sources. We discuss the composition of the gamma-ray sky and capabilities of the 1pPDF method.

  1. The estimation of differential counting measurements of possitive quantities with relatively large statistical errors

    International Nuclear Information System (INIS)

    Vincent, C.H.

    1982-01-01

    Bayes' principle is applied to the differential counting measurement of a positive quantity in which the statistical errors are not necessarily small in relation to the true value of the quantity. The methods of estimation derived are found to give consistent results and to avoid the anomalous negative estimates sometimes obtained by conventional methods. One of the methods given provides a simple means of deriving the required estimates from conventionally presented results and appears to have wide potential applications. Both methods provide the actual posterior probability distribution of the quantity to be measured. A particularly important potential application is the correction of counts on low radioacitvity samples for background. (orig.)

  2. Farey Statistics in Time n^{2/3} and Counting Primitive Lattice Points in Polygons

    OpenAIRE

    Patrascu, Mihai

    2007-01-01

    We present algorithms for computing ranks and order statistics in the Farey sequence, taking time O (n^{2/3}). This improves on the recent algorithms of Pawlewicz [European Symp. Alg. 2007], running in time O (n^{3/4}). We also initiate the study of a more general algorithmic problem: counting primitive lattice points in planar shapes.

  3. Metrology and statistical analysis for the precise standardisation of cobalt-60 by 4πβ-γ coincidence counting

    International Nuclear Information System (INIS)

    Buckman, S.M.

    1995-03-01

    The major part of the thesis is devoted to the theoretical development of a comprehensive PC-based statistical package for the analysis of data from coincidence-counting experiments. This analysis is applied to primary standardizations of Co performed in Australia and Japan. The Australian standardisation, the accuracy of which is confirmed through international comparison, is used to re-calibrate the ionisation chamber. Both Australian and Japanese coincidence-counting systems are interfaced to personal computers to enable replicated sets of measurements to be made under computer control. Further research to confirm the validity of the statistical model includes an experimental investigation into the non-Poisson behaviour of radiation detectors due to the effect of deadtime. Experimental investigation is conducted to determine which areas are most likely to limit the ultimate accuracy achievable with coincidence counting. The thesis concludes by discussing the possibilities of digital coincidence counting and outlines the design of a prototype system presently under development. The accuracy of the Australian standardisation is confirmed by international comparison. From this result a more accurate Co calibration is obtained for the Australian working standard. Based on the work of this thesis, uncertainties in coincidence counting experiments can be better handled with resulting improvements in measurement reliability. The concept and benefits of digital coincidence counting are discussed and a proposed design is given for such a system. All of the data and software associated with this thesis is provided on computer discs. 237 refs., figs., tabs

  4. The statistical interpretations of counting data from measurements of low-level radioactivity

    International Nuclear Information System (INIS)

    Donn, J.J.; Wolke, R.L.

    1977-01-01

    The statistical model appropriate to measurements of low-level or background-dominant radioactivity is examined and the derived relationships are applied to two practical problems involving hypothesis testing: 'Does the sample exhibit a net activity above background' and 'Is the activity of the sample below some preselected limit'. In each of these cases, the appropriate decision rule is formulated, procedures are developed for estimating the preset count which is necessary to achieve a desired probability of detection, and a specific sequence of operations is provided for the worker in the field. (author)

  5. Full counting statistics of level renormalization in electron transport through double quantum dots

    International Nuclear Information System (INIS)

    Luo Junyan; Shen Yu; Cen Gang; He Xiaoling; Wang Changrong; Jiao Hujun

    2011-01-01

    We examine the full counting statistics of electron transport through double quantum dots coupled in series, with particular attention being paid to the unique features originating from level renormalization. It is clearly illustrated that the energy renormalization gives rise to a dynamic charge blockade mechanism, which eventually results in super-Poissonian noise. Coupling of the double dots to an external heat bath leads to dephasing and relaxation mechanisms, which are demonstrated to suppress the noise in a unique way.

  6. Bayesian Penalized Likelihood Image Reconstruction (Q.Clear) in 82Rb Cardiac PET: Impact of Count Statistics

    DEFF Research Database (Denmark)

    Christensen, Nana Louise; Tolbod, Lars Poulsen

    PET scans. 3) Static and dynamic images from a set of 7 patients (BSA: 1.6-2.2 m2) referred for 82Rb cardiac PET was analyzed using a range of beta factors. Results were compared to the institution’s standard clinical practice reconstruction protocol. All scans were performed on GE DMI Digital......Aim: Q.Clear reconstruction is expected to improve detection of perfusion defects in cardiac PET due to the high degree of image convergence and effective noise suppression. However, 82Rb (T½=76s) possess a special problem, since count statistics vary significantly not only between patients...... statistics using a cardiac PET phantom as well as a selection of clinical patients referred for 82Rb cardiac PET. Methods: The study consistent of 3 parts: 1) A thorax-cardiac phantom was scanned for 10 minutes after injection of 1110 MBq 82Rb. Frames at 3 different times after infusion were reconstructed...

  7. RCT: Module 2.03, Counting Errors and Statistics, Course 8768

    Energy Technology Data Exchange (ETDEWEB)

    Hillmer, Kurt T. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2017-04-01

    Radiological sample analysis involves the observation of a random process that may or may not occur and an estimation of the amount of radioactive material present based on that observation. Across the country, radiological control personnel are using the activity measurements to make decisions that may affect the health and safety of workers at those facilities and their surrounding environments. This course will present an overview of measurement processes, a statistical evaluation of both measurements and equipment performance, and some actions to take to minimize the sources of error in count room operations. This course will prepare the student with the skills necessary for radiological control technician (RCT) qualification by passing quizzes, tests, and the RCT Comprehensive Phase 1, Unit 2 Examination (TEST 27566) and by providing in the field skills.

  8. Assessment of the statistical uncertainty affecting a counting; Evaluation de l'incertitude statistique affectant un comptage

    Energy Technology Data Exchange (ETDEWEB)

    Cluchet, J.

    1960-07-01

    After a recall of some aspects regarding the Gauss law and the Gauss curve, this note addresses the case of performance of a large number of measurements of a source activity by means of a sensor (counter, scintillator, nuclear emulsion, etc.) at equal intervals, and with a number of events which is not rigorously constant. Thus, it addresses measurements, and more particularly counting operations in a random or statistical environment. It more particularly addresses the case of a counting rate due to a source greater (and then lower) than twenty times the Eigen movement. The validity of curves is discussed.

  9. High order statistical signatures from source-driven measurements of subcritical fissile systems

    International Nuclear Information System (INIS)

    Mattingly, J.K.

    1998-01-01

    This research focuses on the development and application of high order statistical analyses applied to measurements performed with subcritical fissile systems driven by an introduced neutron source. The signatures presented are derived from counting statistics of the introduced source and radiation detectors that observe the response of the fissile system. It is demonstrated that successively higher order counting statistics possess progressively higher sensitivity to reactivity. Consequently, these signatures are more sensitive to changes in the composition, fissile mass, and configuration of the fissile assembly. Furthermore, it is shown that these techniques are capable of distinguishing the response of the fissile system to the introduced source from its response to any internal or inherent sources. This ability combined with the enhanced sensitivity of higher order signatures indicates that these techniques will be of significant utility in a variety of applications. Potential applications include enhanced radiation signature identification of weapons components for nuclear disarmament and safeguards applications and augmented nondestructive analysis of spent nuclear fuel. In general, these techniques expand present capabilities in the analysis of subcritical measurements

  10. Full counting statistics of a charge pump in the Coulomb blockade regime

    Science.gov (United States)

    Andreev, A. V.; Mishchenko, E. G.

    2001-12-01

    We study full charge counting statistics (FCCS) of a charge pump based on a nearly open single electron transistor. The problem is mapped onto an exactly soluble problem of a nonequilibrium g=1/2 Luttinger liquid with an impurity. We obtain an analytic expression for the generating function of the transmitted charge for an arbitrary pumping strength. Although this model contains fractionally charged excitations only integer transmitted charges can be observed. In the weak pumping limit FCCS correspond to a Poissonian transmission of particles with charge e*=e/2 from which all events with odd numbers of transferred particles are excluded.

  11. Statistical approaches to the analysis of point count data: A little extra information can go a long way

    Science.gov (United States)

    Farnsworth, G.L.; Nichols, J.D.; Sauer, J.R.; Fancy, S.G.; Pollock, K.H.; Shriner, S.A.; Simons, T.R.; Ralph, C. John; Rich, Terrell D.

    2005-01-01

    Point counts are a standard sampling procedure for many bird species, but lingering concerns still exist about the quality of information produced from the method. It is well known that variation in observer ability and environmental conditions can influence the detection probability of birds in point counts, but many biologists have been reluctant to abandon point counts in favor of more intensive approaches to counting. However, over the past few years a variety of statistical and methodological developments have begun to provide practical ways of overcoming some of the problems with point counts. We describe some of these approaches, and show how they can be integrated into standard point count protocols to greatly enhance the quality of the information. Several tools now exist for estimation of detection probability of birds during counts, including distance sampling, double observer methods, time-depletion (removal) methods, and hybrid methods that combine these approaches. Many counts are conducted in habitats that make auditory detection of birds much more likely than visual detection. As a framework for understanding detection probability during such counts, we propose separating two components of the probability a bird is detected during a count into (1) the probability a bird vocalizes during the count and (2) the probability this vocalization is detected by an observer. In addition, we propose that some measure of the area sampled during a count is necessary for valid inferences about bird populations. This can be done by employing fixed-radius counts or more sophisticated distance-sampling models. We recommend any studies employing point counts be designed to estimate detection probability and to include a measure of the area sampled.

  12. Optimization of statistical methods for HpGe gamma-ray spectrometer used in wide count rate ranges

    Energy Technology Data Exchange (ETDEWEB)

    Gervino, G., E-mail: gervino@to.infn.it [UNITO - Università di Torino, Dipartimento di Fisica, Turin (Italy); INFN - Istituto Nazionale di Fisica Nucleare, Sez. Torino, Turin (Italy); Mana, G. [INRIM - Istituto Nazionale di Ricerca Metrologica, Turin (Italy); Palmisano, C. [UNITO - Università di Torino, Dipartimento di Fisica, Turin (Italy); INRIM - Istituto Nazionale di Ricerca Metrologica, Turin (Italy)

    2016-07-11

    The need to perform γ-ray measurements with HpGe detectors is a common technique in many fields such as nuclear physics, radiochemistry, nuclear medicine and neutron activation analysis. The use of HpGe detectors is chosen in situations where isotope identification is needed because of their excellent resolution. Our challenge is to obtain the “best” spectroscopy data possible in every measurement situation. “Best” is a combination of statistical (number of counts) and spectral quality (peak, width and position) over a wide range of counting rates. In this framework, we applied Bayesian methods and the Ellipsoidal Nested Sampling (a multidimensional integration technique) to study the most likely distribution for the shape of HpGe spectra. In treating these experiments, the prior information suggests to model the likelihood function with a product of Poisson distributions. We present the efforts that have been done in order to optimize the statistical methods to HpGe detector outputs with the aim to evaluate to a better order of precision the detector efficiency, the absolute measured activity and the spectra background. Reaching a more precise knowledge of statistical and systematic uncertainties for the measured physical observables is the final goal of this research project.

  13. On-line statistical processing of radiation detector pulse trains with time-varying count rates

    International Nuclear Information System (INIS)

    Apostolopoulos, G.

    2008-01-01

    Statistical analysis is of primary importance for the correct interpretation of nuclear measurements, due to the inherent random nature of radioactive decay processes. This paper discusses the application of statistical signal processing techniques to the random pulse trains generated by radiation detectors. The aims of the presented algorithms are: (i) continuous, on-line estimation of the underlying time-varying count rate θ(t) and its first-order derivative dθ/dt; (ii) detection of abrupt changes in both of these quantities and estimation of their new value after the change point. Maximum-likelihood techniques, based on the Poisson probability distribution, are employed for the on-line estimation of θ and dθ/dt. Detection of abrupt changes is achieved on the basis of the generalized likelihood ratio statistical test. The properties of the proposed algorithms are evaluated by extensive simulations and possible applications for on-line radiation monitoring are discussed

  14. Improving EWMA Plans for Detecting Unusual Increases in Poisson Counts

    Directory of Open Access Journals (Sweden)

    R. S. Sparks

    2009-01-01

    adaptive exponentially weighted moving average (EWMA plan is developed for signalling unusually high incidence when monitoring a time series of nonhomogeneous daily disease counts. A Poisson transitional regression model is used to fit background/expected trend in counts and provides “one-day-ahead” forecasts of the next day's count. Departures of counts from their forecasts are monitored. The paper outlines an approach for improving early outbreak data signals by dynamically adjusting the exponential weights to be efficient at signalling local persistent high side changes. We emphasise outbreak signals in steady-state situations; that is, changes that occur after the EWMA statistic had run through several in-control counts.

  15. High resolution gamma-ray spectroscopy at high count rates with a prototype High Purity Germanium detector

    Science.gov (United States)

    Cooper, R. J.; Amman, M.; Vetter, K.

    2018-04-01

    High-resolution gamma-ray spectrometers are required for applications in nuclear safeguards, emergency response, and fundamental nuclear physics. To overcome one of the shortcomings of conventional High Purity Germanium (HPGe) detectors, we have developed a prototype device capable of achieving high event throughput and high energy resolution at very high count rates. This device, the design of which we have previously reported on, features a planar HPGe crystal with a reduced-capacitance strip electrode geometry. This design is intended to provide good energy resolution at the short shaping or digital filter times that are required for high rate operation and which are enabled by the fast charge collection afforded by the planar geometry crystal. In this work, we report on the initial performance of the system at count rates up to and including two million counts per second.

  16. A high dynamic range pulse counting detection system for mass spectrometry.

    Science.gov (United States)

    Collings, Bruce A; Dima, Martian D; Ivosev, Gordana; Zhong, Feng

    2014-01-30

    A high dynamic range pulse counting system has been developed that demonstrates an ability to operate at up to 2e8 counts per second (cps) on a triple quadrupole mass spectrometer. Previous pulse counting detection systems have typically been limited to about 1e7 cps at the upper end of the systems dynamic range. Modifications to the detection electronics and dead time correction algorithm are described in this paper. A high gain transimpedance amplifier is employed that allows a multi-channel electron multiplier to be operated at a significantly lower bias potential than in previous pulse counting systems. The system utilises a high-energy conversion dynode, a multi-channel electron multiplier, a high gain transimpedance amplifier, non-paralysing detection electronics and a modified dead time correction algorithm. Modification of the dead time correction algorithm is necessary due to a characteristic of the pulse counting electronics. A pulse counting detection system with the capability to count at ion arrival rates of up to 2e8 cps is described. This is shown to provide a linear dynamic range of nearly five orders of magnitude for a sample of aprazolam with concentrations ranging from 0.0006970 ng/mL to 3333 ng/mL while monitoring the m/z 309.1 → m/z 205.2 transition. This represents an upward extension of the detector's linear dynamic range of about two orders of magnitude. A new high dynamic range pulse counting system has been developed demonstrating the ability to operate at up to 2e8 cps on a triple quadrupole mass spectrometer. This provides an upward extension of the detector's linear dynamic range by about two orders of magnitude over previous pulse counting systems. Copyright © 2013 John Wiley & Sons, Ltd.

  17. Cerenkov counting and Cerenkov-scintillation counting with high refractive index organic liquids using a liquid scintillation counter

    International Nuclear Information System (INIS)

    Wiebe, L.I.; Helus, F.; Maier-Borst, W.

    1978-01-01

    18 F and 14 C radioactivity was measured in methyl salicylate (MS), a high refractive index hybrid Cherenkov-scintillation generating medium, using a liquid scintillation counter. At concentrations of up to 21.4%, in MS, dimethyl sulfoxide (DMSO) quenched 14 C fluorescence, and with a 10-fold excess of DMSO over MS, 18 F count rates were reduced below that for DMSO alone, probably as a result of concentration-independent self-quenching due to 'dark-complex' formation. DMSO in lower concentrations did not reduce the counting efficiency of 18 F in MS. Nitrobenzene was a concentration-dependent quencher for both 14 C and 18 F in MS. Chlorobenzene (CB) and DMSO were both found to be weak Cherenkov generators with 18 F. Counting efficiencies for 18 F in MS, CB, and DMSO were 50.3, 7.8 and 4.3% respectively in the coincidence counting mode, and 58.1, 13.0 and 6.8% in the singles mode. 14 C efficiencies were 14.4 and 22.3% for coincidence and singles respectively, and 15.3 and 42.0% using a modern counter designed for coincidence and single photon counting. The high 14 C and 18 F counting efficiency in MS are discussed with respect to excitation mechanism, on the basis of quench and channels ratios changes observed. It is proposed that MS functions as an efficient Cherenkov-scintillation generator for high-energy beta emitters such as 18 F, and as a low-efficiency scintillator for weak beta emitting radionuclides such as 14 C. (author)

  18. Accuracy in activation analysis: count rate effects

    International Nuclear Information System (INIS)

    Lindstrom, R.M.; Fleming, R.F.

    1980-01-01

    The accuracy inherent in activation analysis is ultimately limited by the uncertainty of counting statistics. When careful attention is paid to detail, several workers have shown that all systematic errors can be reduced to an insignificant fraction of the total uncertainty, even when the statistical limit is well below one percent. A matter of particular importance is the reduction of errors due to high counting rate. The loss of counts due to random coincidence (pulse pileup) in the amplifier and to digitization time in the ADC may be treated as a series combination of extending and non-extending dead times, respectively. The two effects are experimentally distinct. Live timer circuits in commercial multi-channel analyzers compensate properly for ADC dead time for long-lived sources, but not for pileup. Several satisfactory solutions are available, including pileup rejection and dead time correction circuits, loss-free ADCs, and computed corrections in a calibrated system. These methods are sufficiently reliable and well understood that a decaying source can be measured routinely with acceptably small errors at a dead time as high as 20 percent

  19. Unifying quantum heat transfer in a nonequilibrium spin-boson model with full counting statistics

    Science.gov (United States)

    Wang, Chen; Ren, Jie; Cao, Jianshu

    2017-02-01

    To study the full counting statistics of quantum heat transfer in a driven nonequilibrium spin-boson model, we develop a generalized nonequilibrium polaron-transformed Redfield equation with an auxiliary counting field. This enables us to study the impact of qubit-bath coupling ranging from weak to strong regimes. Without external modulations, we observe maximal values of both steady-state heat flux and noise power in moderate coupling regimes, below which we find that these two transport quantities are enhanced by the finite-qubit-energy bias. With external modulations, the geometric-phase-induced heat flux shows a monotonic decrease upon increasing the qubit-bath coupling at zero qubit energy bias (without bias). While under the finite-qubit-energy bias (with bias), the geometric-phase-induced heat flux exhibits an interesting reversal behavior in the strong coupling regime. Our results unify the seemingly contradictory results in weak and strong qubit-bath coupling regimes and provide detailed dissections for the quantum fluctuation of nonequilibrium heat transfer.

  20. Application of high intensity ultrasound treatment on Enterobacteriae count in milk

    Directory of Open Access Journals (Sweden)

    Anet Režek Jambrak

    2011-06-01

    Full Text Available Ultrasonication is a non-thermal method of food preservation that has the advantage of inactivating microbes in food without causing the common side-effects associated with conventional heat treatments, such as nutrient and flavour loss. In this work high intensity ultrasound was used to investigate inactivation Enterobacteriae count in raw milk. Raw milk with 4% of milk fat was treated with ultrasonic probe that was 12 mm in diameter and with 20 kHz frequency immerged in milk directly. For ultrasounds treatment, three parameters varied according to the statistical experimental design. Centre composite design was used to optimize and design experimental parameters: temperature (20, 40 and 60 °C, amplitude (120, 90 and 60 μm and time (6, 9 and 12 minutes. All analyses were performed immediately after sonication and after 3 and 5 days of storage in refrigeration at 4 °C. The facts that substantially affect the inactivation of microorganisms using ultrasound are the amplitude of the ultrasonic waves, the exposure/contact time with the microorganisms, and the temperatureof treatment. The achieved results indicate significant inactivation of microorganisms under longer period of treatments with ultrasonic probe particularly in combination with higher temperature andamplitude. Output optimal value of Enterobacteriae count has been defined by Statgraphics where lowest Enterobacteriae count (1.06151 log CFU mL-1 was as follows for specific ultrasound parameters: amplitude of 120 μm, treatment time for 12 min and temperature of 60 °C.

  1. Cerenkov counting and Cerenkov-scintillation counting with high refractive index organic liquids using a liquid scintillation counter

    Energy Technology Data Exchange (ETDEWEB)

    Wiebe, L I; Helus, F; Maier-Borst, W [Deutsches Krebsforschungszentrum, Heidelberg (Germany, F.R.). Inst. fuer Nuklearmedizin

    1978-06-01

    /sup 18/F and /sup 14/C radioactivity was measured in methyl salicylate (MS), a high refractive index hybrid Cherenkov-scintillation generating medium, using a liquid scintillation counter. At concentrations of up to 21.4%, in MS, dimethyl sulfoxide (DMSO) quenched /sup 14/C fluorescence, and with a 10-fold excess of DMSO over MS, /sup 18/F count rates were reduced below that for DMSO alone, probably as a result of concentration-independent self-quenching due to 'dark-complex' formation. DMSO in lower concentrations did not reduce the counting efficiency of /sup 18/F in MS. Nitrobenzene was a concentration-dependent quencher for both /sup 14/C and /sup 18/F in MS. Chlorobenzene (CB) and DMSO were both found to be weak Cherenkov generators with /sup 18/F. Counting efficiencies for /sup 18/F in MS, CB, and DMSO were 50.3, 7.8 and 4.3% respectively in the coincidence counting mode, and 58.1, 13.0 and 6.8% in the singles mode. /sup 14/C efficiencies were 14.4 and 22.3% for coincidence and singles respectively, and 15.3 and 42.0% using a modern counter designed for coincidence and single photon counting. The high /sup 14/C and /sup 18/F counting efficiency in MS are discussed with respect to excitation mechanism, on the basis of quench and channels ratios changes observed. It is proposed that MS functions as an efficient Cherenkov-scintillation generator for high-energy beta emitters such as /sup 18/F, and as a low-efficiency scintillator for weak beta emitting radionuclides such as /sup 14/C.

  2. Statistical Methods for Unusual Count Data: Examples From Studies of Microchimerism

    Science.gov (United States)

    Guthrie, Katherine A.; Gammill, Hilary S.; Kamper-Jørgensen, Mads; Tjønneland, Anne; Gadi, Vijayakrishna K.; Nelson, J. Lee; Leisenring, Wendy

    2016-01-01

    Natural acquisition of small amounts of foreign cells or DNA, referred to as microchimerism, occurs primarily through maternal-fetal exchange during pregnancy. Microchimerism can persist long-term and has been associated with both beneficial and adverse human health outcomes. Quantitative microchimerism data present challenges for statistical analysis, including a skewed distribution, excess zero values, and occasional large values. Methods for comparing microchimerism levels across groups while controlling for covariates are not well established. We compared statistical models for quantitative microchimerism values, applied to simulated data sets and 2 observed data sets, to make recommendations for analytic practice. Modeling the level of quantitative microchimerism as a rate via Poisson or negative binomial model with the rate of detection defined as a count of microchimerism genome equivalents per total cell equivalents tested utilizes all available data and facilitates a comparison of rates between groups. We found that both the marginalized zero-inflated Poisson model and the negative binomial model can provide unbiased and consistent estimates of the overall association of exposure or study group with microchimerism detection rates. The negative binomial model remains the more accessible of these 2 approaches; thus, we conclude that the negative binomial model may be most appropriate for analyzing quantitative microchimerism data. PMID:27769989

  3. Variability in faecal egg counts – a statistical model to achieve reliable determination of anthelmintic resistance in livestock

    DEFF Research Database (Denmark)

    Nielsen, Martin Krarup; Vidyashankar, Anand N.; Hanlon, Bret

    statistical model was therefore developed for analysis of FECRT data from multiple farms. Horse age, gender, zip code and pre-treatment egg count were incorporated into the model. Horses and farms were kept as random effects. Resistance classifications were based on model-based 95% lower confidence limit (LCL...

  4. On temporal correlations in high-resolution frequency counting

    OpenAIRE

    Dunker, Tim; Hauglin, Harald; Rønningen, Ole Petter

    2016-01-01

    We analyze noise properties of time series of frequency data from different counting modes of a Keysight 53230A frequency counter. We use a 10 MHz reference signal from a passive hydrogen maser connected via phase-stable Huber+Suhner Sucoflex 104 cables to the reference and input connectors of the counter. We find that the high resolution gap-free (CONT) frequency counting process imposes long-term correlations in the output data, resulting in a modified Allan deviation that is characteristic...

  5. Evaluation of high-energy electron detectors for probing the inner magnetosphere under high-counting condition

    International Nuclear Information System (INIS)

    Tamada, Yukihiro; Takashima, Takeshi; Mitani, Takefumi; Miyake, Wataru

    2013-01-01

    An ERG (Energization and Radiation in Geospace) satellite will be launched to study the acceleration processes of energetic particles in the radiation belt surrounding the earth. It is very important to reveal the acceleration process of high-energy particles for both science and the application to space weather forecast. Drastic increases of high-energy electrons in the radiation belt is sometimes observed during a geomagnetic storm. When a large magnetic storm occurs, energetic electron count rates may exceed flux limits expected in the nominal design and large number of incident electrons leading to detection loss. The purpose of this study is to demonstrate that the count rate range of a single detection on board ERG satellite can be expanded by means of reading circuit operations to decrease an area of detection. In our ground experiment, we also found an unexpected result that count peaks shift to the higher energy side under high counting conditions. (author)

  6. High Channel Count, High Density Microphone Arrays for Wind Tunnel Environments, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — The Interdisciplinary Consulting Corporation (IC2) proposes the development of high channel count, high density, reduced cost per channel, directional microphone...

  7. Assessment of noise in a digital image using the join-count statistic and the Moran test

    International Nuclear Information System (INIS)

    Kehshih Chuang; Huang, H.K.

    1992-01-01

    It is assumed that data bits of a pixel in digital images can be divided into signal and noise bits. The signal bits occupy the most significant part of the pixel. The signal parts of each pixel are correlated while the noise parts are uncorrelated. Two statistical methods, the Moran test and the join-count statistic, are used to examine the noise parts. Images from computerized tomography, magnetic resonance and computed radiography are used for the evaluation of the noise bits. A residual image is formed by subtracting the original image from its smoothed version. The noise level in the residual image is then identical to that in the original image. Both statistical tests are then performed on the bit planes of the residual image. Results show that most digital images contain only 8-9 bits of correlated information. Both methods are easy to implement and fast to perform. (author)

  8. A pulse shape discriminator with high precision of neutron and gamma ray selection at high counting rate

    International Nuclear Information System (INIS)

    Bialkowski, J.; Moszynski, M.; Wolski, D.

    1989-01-01

    A pulse shape discriminator based on the zero-crossing principle is described. Due to dc negative feedback loops stabilizing the shaping amplifier and the zero-crossing discriminator, the working of the circuit is not affected by the high counting rate and the temperature variations. The pileup rejection circuit built into the discriminator improves the quality of the n-γ separation at high counting rates. A full γ-ray rejection is obtained for a recoil energy of electrons down to 25 keV. At high counting rates the remaining γ-ray contribution is evidently due to the pileup effect which is equal to about 2% at 4x10 5 counts/s. (orig.)

  9. Determining Gate Count Reliability in a Library Setting

    Directory of Open Access Journals (Sweden)

    Jeffrey Phillips

    2016-09-01

    Full Text Available Objective – Patron counts are a common form of measurement for library assessment. To develop accurate library statistics, it is necessary to determine any differences between various counting devices. A yearlong comparison between card reader turnstiles and laser gate counters in a university library sought to offer a standard percentage of variance and provide suggestions to increase the precision of counts. Methods – The collection of library exit counts identified the differences between turnstile and laser gate counter data. Statistical software helped to eliminate any inaccuracies in the collection of turnstile data, allowing this data set to be the base for comparison. Collection intervals were randomly determined and demonstrated periods of slow, average, and heavy traffic. Results – After analyzing 1,039,766 patron visits throughout a year, the final totals only showed a difference of .43% (.0043 between the two devices. The majority of collection periods did not exceed a difference of 3% between the counting instruments. Conclusion – Turnstiles card readers and laser gate counters provide similar levels of reliability when measuring patron activity. Each system has potential counting inaccuracies, but several methods exist to create more precise totals. Turnstile card readers are capable of offering greater detail involving patron identity, but their high cost makes them inaccessible for libraries with lower budgets. This makes laser gate counters an affordable alternative for reliable patron counting in an academic library.

  10. An automated approach for annual layer counting in ice cores

    Directory of Open Access Journals (Sweden)

    M. Winstrup

    2012-11-01

    Full Text Available A novel method for automated annual layer counting in seasonally-resolved paleoclimate records has been developed. It relies on algorithms from the statistical framework of hidden Markov models (HMMs, which originally was developed for use in machine speech recognition. The strength of the layer detection algorithm lies in the way it is able to imitate the manual procedures for annual layer counting, while being based on statistical criteria for annual layer identification. The most likely positions of multiple layer boundaries in a section of ice core data are determined simultaneously, and a probabilistic uncertainty estimate of the resulting layer count is provided, ensuring an objective treatment of ambiguous layers in the data. Furthermore, multiple data series can be incorporated and used simultaneously. In this study, the automated layer counting algorithm has been applied to two ice core records from Greenland: one displaying a distinct annual signal and one which is more challenging. The algorithm shows high skill in reproducing the results from manual layer counts, and the resulting timescale compares well to absolute-dated volcanic marker horizons where these exist.

  11. Fast pulse discriminator for photon counting at high photon densities

    International Nuclear Information System (INIS)

    Benoit, R.; Pedrini, A.

    1977-03-01

    A fast tunnel diode discriminator for photon counting up to 200MHz count frequency is described. The tunnel diode is operated on its apparent I.V. characteristics displayed when the diode is driven into its oscillating region. The pulse shaper-discriminator is completely D.C. coupled in order to avoid base-line shift at high pulse rates

  12. Full counting statistics in a serially coupled double quantum dot system with spin-orbit coupling

    Science.gov (United States)

    Wang, Qiang; Xue, Hai-Bin; Xie, Hai-Qing

    2018-04-01

    We study the full counting statistics of electron transport through a serially coupled double quantum dot (QD) system with spin-orbit coupling (SOC) weakly coupled to two electrodes. We demonstrate that the spin polarizations of the source and drain electrodes determine whether the shot noise maintains super-Poissonian distribution, and whether the sign transitions of the skewness from positive to negative values and of the kurtosis from negative to positive values take place. In particular, the interplay between the spin polarizations of the source and drain electrodes and the magnitude of the external magnetic field, can give rise to a gate-voltage-tunable strong negative differential conductance (NDC) and the shot noise in this NDC region is significantly enhanced. Importantly, for a given SOC parameter, the obvious variation of the high-order current cumulants as a function of the energy-level detuning in a certain range, especially the dip position of the Fano factor of the skewness can be used to qualitatively extract the information about the magnitude of the SOC.

  13. Physics colloquium: Single-electron counting in quantum metrology and in statistical mechanics

    CERN Multimedia

    Geneva University

    2011-01-01

    GENEVA UNIVERSITY Ecole de physique Département de physique nucléaire et corspusculaire 24, quai Ernest-Ansermet 1211 Genève 4 Tél.: (022) 379 62 73 Fax: (022) 379 69 92olé   Lundi 17 octobre 2011 17h00 - Ecole de Physique, Auditoire Stueckelberg PHYSICS COLLOQUIUM « Single-electron counting in quantum metrology and in statistical mechanics » Prof. Jukka Pekola Low Temperature Laboratory, Aalto University Helsinki, Finland   First I discuss the basics of single-electron tunneling and its potential applications in metrology. My main focus is in developing an accurate source of single-electron current for the realization of the unit ampere. I discuss the principle and the present status of the so-called single- electron turnstile. Investigation of errors in transporting electrons one by one has revealed a wealth of observations on fundamental phenomena in mesoscopic superconductivity, including individual Andreev...

  14. Smart pile-up consideration for evaluation of high count rate EDS spectra

    International Nuclear Information System (INIS)

    Eggert, F; Anderhalt, R; Nicolosi, J; Elam, T

    2012-01-01

    This work describes a new pile-up consideration for the very high count rate spectra which are possible to acquire with silicon drift detector (SDD) technology. Pile-up effects are the major and still remaining challenge with the use of SDD for EDS in scanning electron microscopes (SEM) with ultra thin windows for soft X-ray detection. The ability to increase the count rates up to a factor of 100 compared with conventional Si(Li) detectors, comes with the problem that the pile-up recognition (pile-up rejection) in pulse processors is not able to improve by the same order of magnitude, just only with a factor of about 3. Therefore, it is common that spectra will show significant pile-up effects if count rates of more than 10000 counts per second (10 kcps) are used. These false counts affect both automatic qualitative analysis and quantitative evaluation of the spectra. The new idea is to use additional inputs for pile-up calculation to shift the applicability towards very high count rates of up to 200 kcps and more, which can be easily acquired with the SDD. The additional input is the 'known' (estimated) background distribution, calculated iteratively during all automated qualitative or quantitative evaluations. This additional knowledge gives the opportunity for self adjustment of the pile-up calculation parameters and avoids over-corrections which challenge the evaluation as well as the pile-up artefacts themselves. With the proposed method the pile-up correction is no longer a 'correction' but an integral part of all spectra evaluation steps. Examples for the application are given with evaluation of very high count rate spectra.

  15. The power of statistical tests using field trial count data of non-target organisms in enviromental risk assessment of genetically modified plants

    NARCIS (Netherlands)

    Voet, van der H.; Goedhart, P.W.

    2015-01-01

    Publications on power analyses for field trial count data comparing transgenic and conventional crops have reported widely varying requirements for the replication needed to obtain statistical tests with adequate power. These studies are critically reviewed and complemented with a new simulation

  16. Studies on the Pulse Rate, Pedometer Count and Satisfactoin Degree at Various Exercise

    OpenAIRE

    小原, 史朗

    2004-01-01

    This investigation examined whether free exercise of students became good stimulation of breathing circulation function from relation of pulse rate and pedometer count. And, I examined it on satisfaction degree after exercise. Object person was 432 man students (total of 1391) and 94 woman students (total of 472). As for relation of pulse rate and pedometer count, statistical meaning was recognized by man and women. The exercise that a pulse rate and pedometer count were high together seemed ...

  17. A unified statistical framework for material decomposition using multienergy photon counting x-ray detectors

    International Nuclear Information System (INIS)

    Choi, Jiyoung; Kang, Dong-Goo; Kang, Sunghoon; Sung, Younghun; Ye, Jong Chul

    2013-01-01

    Purpose: Material decomposition using multienergy photon counting x-ray detectors (PCXD) has been an active research area over the past few years. Even with some success, the problem of optimal energy selection and three material decomposition including malignant tissue is still on going research topic, and more systematic studies are required. This paper aims to address this in a unified statistical framework in a mammographic environment.Methods: A unified statistical framework for energy level optimization and decomposition of three materials is proposed. In particular, an energy level optimization algorithm is derived using the theory of the minimum variance unbiased estimator, and an iterative algorithm is proposed for material composition as well as system parameter estimation under the unified statistical estimation framework. To verify the performance of the proposed algorithm, the authors performed simulation studies as well as real experiments using physical breast phantom and ex vivo breast specimen. Quantitative comparisons using various performance measures were conducted, and qualitative performance evaluations for ex vivo breast specimen were also performed by comparing the ground-truth malignant tissue areas identified by radiologists.Results: Both simulation and real experiments confirmed that the optimized energy bins by the proposed method allow better material decomposition quality. Moreover, for the specimen thickness estimation errors up to 2 mm, the proposed method provides good reconstruction results in both simulation and real ex vivo breast phantom experiments compared to existing methods.Conclusions: The proposed statistical framework of PCXD has been successfully applied for the energy optimization and decomposition of three material in a mammographic environment. Experimental results using the physical breast phantom and ex vivo specimen support the practicality of the proposed algorithm

  18. A question of separation: disentangling tracer bias and gravitational non-linearity with counts-in-cells statistics

    Science.gov (United States)

    Uhlemann, C.; Feix, M.; Codis, S.; Pichon, C.; Bernardeau, F.; L'Huillier, B.; Kim, J.; Hong, S. E.; Laigle, C.; Park, C.; Shin, J.; Pogosyan, D.

    2018-02-01

    Starting from a very accurate model for density-in-cells statistics of dark matter based on large deviation theory, a bias model for the tracer density in spheres is formulated. It adopts a mean bias relation based on a quadratic bias model to relate the log-densities of dark matter to those of mass-weighted dark haloes in real and redshift space. The validity of the parametrized bias model is established using a parametrization-independent extraction of the bias function. This average bias model is then combined with the dark matter PDF, neglecting any scatter around it: it nevertheless yields an excellent model for densities-in-cells statistics of mass tracers that is parametrized in terms of the underlying dark matter variance and three bias parameters. The procedure is validated on measurements of both the one- and two-point statistics of subhalo densities in the state-of-the-art Horizon Run 4 simulation showing excellent agreement for measured dark matter variance and bias parameters. Finally, it is demonstrated that this formalism allows for a joint estimation of the non-linear dark matter variance and the bias parameters using solely the statistics of subhaloes. Having verified that galaxy counts in hydrodynamical simulations sampled on a scale of 10 Mpc h-1 closely resemble those of subhaloes, this work provides important steps towards making theoretical predictions for density-in-cells statistics applicable to upcoming galaxy surveys like Euclid or WFIRST.

  19. Relationship of long-term highly active antiretroviral therapy on salivary flow rate and CD4 Count among HIV-infected patients.

    Science.gov (United States)

    Kumar, J Vijay; Baghirath, P Venkat; Naishadham, P Parameswar; Suneetha, Sujai; Suneetha, Lavanya; Sreedevi, P

    2015-01-01

    To determine if long-term highly active antiretroviral therapy (HAART) therapy alters salivary flow rate and also to compare its relation of CD4 count with unstimulated and stimulated whole saliva. A cross-sectional study was performed on 150 individuals divided into three groups. Group I (50 human immunodeficiency virus (HIV) seropositive patients, but not on HAART therapy), Group II (50 HIV-infected subjects and on HAART for less than 3 years called short-term HAART), Group III (50 HIV-infected subjects and on HAART for more than or equal to 3 years called long-term HAART). Spitting method proposed by Navazesh and Kumar was used for the measurement of unstimulated and stimulated salivary flow rate. Chi-square test and analysis of variance (ANOVA) were used for statistical analysis. The mean CD4 count was 424.78 ± 187.03, 497.82 ± 206.11 and 537.6 ± 264.00 in the respective groups. Majority of the patients in all the groups had a CD4 count between 401 and 600. Both unstimulated and stimulated whole salivary (UWS and SWS) flow rates in Group I was found to be significantly higher than in Group II (P flow rate between Group II and III subjects were also found to be statistically significant (P relationship in Group II (P flow rates of HIV-infected individuals who are on long-term HAART.

  20. Statistical analysis of nematode counts from interlaboratory proficiency tests

    NARCIS (Netherlands)

    Berg, van den W.; Hartsema, O.; Nijs, Den J.M.F.

    2014-01-01

    A series of proficiency tests on potato cyst nematode (PCN; n=29) and free-living stages of Meloidogyne and Pratylenchus (n=23) were investigated to determine the accuracy and precision of the nematode counts and to gain insights into possible trends and potential improvements. In each test, each

  1. Study of mast cell count in skin tags

    Directory of Open Access Journals (Sweden)

    Zaher Hesham

    2007-01-01

    Full Text Available Background: Skin tags or acrochordons are common tumors of middle-aged and elderly subjects. They consist of loose fibrous tissue and occur mainly on the neck and major flexures as small, soft, pedunculated protrusions. Objectives: The aim was to compare the mast cells count in skin tags to adjacent normal skin in diabetic and nondiabetic participants in an attempt to elucidate the possible role of mast cells in the pathogenesis of skin tags. Participants and Methods: Thirty participants with skin tags were divided into group I (15 nondiabetic participants and group II (15 diabetic participants. Three biopsies were obtained from each participant: a large skin tag, a small skin tag and adjacent normal skin. Mast cell count from all the obtained sections was carried out, and the mast cell density was expressed as the average mast cell count/high power field (HPF. Results: A statistically significant increase in mast cells count in skin tags in comparison to normal skin was detected in group I and group II. There was no statistically significant difference between mast cell counts in skin tags of both the groups. Conclusion: Both the mast cell mediators and hyperinsulinemia are capable of inducing fibroblast proliferation and epidermal hyperplasia that are the main pathologic abnormalities seen in all types of skin tags. However, the presence of mast cells in all examined skin tags regardless of diabetes and obesity may point to the possible crucial role of mast cells in the etiogenesis of skin tags through its interaction with fibroblasts and keratinocytes.

  2. Andreev Bound States Formation and Quasiparticle Trapping in Quench Dynamics Revealed by Time-Dependent Counting Statistics.

    Science.gov (United States)

    Souto, R Seoane; Martín-Rodero, A; Yeyati, A Levy

    2016-12-23

    We analyze the quantum quench dynamics in the formation of a phase-biased superconducting nanojunction. We find that in the absence of an external relaxation mechanism and for very general conditions the system gets trapped in a metastable state, corresponding to a nonequilibrium population of the Andreev bound states. The use of the time-dependent full counting statistics analysis allows us to extract information on the asymptotic population of even and odd many-body states, demonstrating that a universal behavior, dependent only on the Andreev state energy, is reached in the quantum point contact limit. These results shed light on recent experimental observations on quasiparticle trapping in superconducting atomic contacts.

  3. High Channel Count Time-to-Digital Converter and Lasercom Processor, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — High-channel-count, high-precision, and high-throughput time-to-digital converters (TDC) are needed to support detector arrays used in deep-space optical...

  4. A high-throughput, multi-channel photon-counting detector with picosecond timing

    CERN Document Server

    Lapington, J S; Miller, G M; Ashton, T J R; Jarron, P; Despeisse, M; Powolny, F; Howorth, J; Milnes, J

    2009-01-01

    High-throughput photon counting with high time resolution is a niche application area where vacuum tubes can still outperform solid-state devices. Applications in the life sciences utilizing time-resolved spectroscopies, particularly in the growing field of proteomics, will benefit greatly from performance enhancements in event timing and detector throughput. The HiContent project is a collaboration between the University of Leicester Space Research Centre, the Microelectronics Group at CERN, Photek Ltd., and end-users at the Gray Cancer Institute and the University of Manchester. The goal is to develop a detector system specifically designed for optical proteomics, capable of high content (multi-parametric) analysis at high throughput. The HiContent detector system is being developed to exploit this niche market. It combines multi-channel, high time resolution photon counting in a single miniaturized detector system with integrated electronics. The combination of enabling technologies; small pore microchanne...

  5. A statistical analysis of count normalization methods used in positron-emission tomography

    International Nuclear Information System (INIS)

    Holmes, T.J.; Ficke, D.C.; Snyder, D.L.

    1984-01-01

    As part of the Positron-Emission Tomography (PET) reconstruction process, annihilation counts are normalized for photon absorption, detector efficiency and detector-pair duty-cycle. Several normalization methods of time-of-flight and conventional systems are analyzed mathematically for count bias and variance. The results of the study have some implications on hardware and software complexity and on image noise and distortion

  6. Mutacins and bacteriocins like genes in Streptococcus mutans isolated from participants with high, moderate, and low salivary count.

    Science.gov (United States)

    Soto, Carolina; Padilla, Carlos; Lobos, Olga

    2017-02-01

    To detect S. mutans producers of mutacins and bacteriocins like substances (BLIS) from saliva of participants with low, moderate, and high salivary counts. 123 strains of S. mutans were obtained from participants with low, moderate, and high salivary counts (age 18 and 20 years old) and their antibacterial capacity analyzed. By using PCR amplification, the expression levels of mutacins and BLIS genes were studied (expressed in arbitrary units/ml) in all three levels. S. mutans strains from participants with low salivary counts show high production of mutacins (63%). In contrast, participants with moderate and high salivary counts depict relatively low levels of mutacins (22 and 15%, respectively). Moreover, participants with low salivary counts showed high expression levels of genes encoding mutacins, a result that correlates with the strong antimicrobial activity of the group. Participants with moderate and high salivary counts however depict low expression levels of mutacin related genes, and little antimicrobial activity. No BLIS were detected in any of the groups studied. S. mutans isolated from the saliva of participants with low bacterial counts have significant antibacterial capacity compared to that of participants with moderate and high salivary counts. The superior lethality of S. mutans in participants with low salivary counts is likely due to the augmented expression of mutacin- related genes. Copyright © 2016 Elsevier Ltd. All rights reserved.

  7. Determining Gate Count Reliability in a Library Setting

    OpenAIRE

    Jeffrey Phillips

    2016-01-01

    Objective – Patron counts are a common form of measurement for library assessment. To develop accurate library statistics, it is necessary to determine any differences between various counting devices. A yearlong comparison between card reader turnstiles and laser gate counters in a university library sought to offer a standard percentage of variance and provide suggestions to increase the precision of counts. Methods – The collection of library exit counts identified the differences be...

  8. A high-throughput, multi-channel photon-counting detector with picosecond timing

    Science.gov (United States)

    Lapington, J. S.; Fraser, G. W.; Miller, G. M.; Ashton, T. J. R.; Jarron, P.; Despeisse, M.; Powolny, F.; Howorth, J.; Milnes, J.

    2009-06-01

    High-throughput photon counting with high time resolution is a niche application area where vacuum tubes can still outperform solid-state devices. Applications in the life sciences utilizing time-resolved spectroscopies, particularly in the growing field of proteomics, will benefit greatly from performance enhancements in event timing and detector throughput. The HiContent project is a collaboration between the University of Leicester Space Research Centre, the Microelectronics Group at CERN, Photek Ltd., and end-users at the Gray Cancer Institute and the University of Manchester. The goal is to develop a detector system specifically designed for optical proteomics, capable of high content (multi-parametric) analysis at high throughput. The HiContent detector system is being developed to exploit this niche market. It combines multi-channel, high time resolution photon counting in a single miniaturized detector system with integrated electronics. The combination of enabling technologies; small pore microchannel plate devices with very high time resolution, and high-speed multi-channel ASIC electronics developed for the LHC at CERN, provides the necessary building blocks for a high-throughput detector system with up to 1024 parallel counting channels and 20 ps time resolution. We describe the detector and electronic design, discuss the current status of the HiContent project and present the results from a 64-channel prototype system. In the absence of an operational detector, we present measurements of the electronics performance using a pulse generator to simulate detector events. Event timing results from the NINO high-speed front-end ASIC captured using a fast digital oscilloscope are compared with data taken with the proposed electronic configuration which uses the multi-channel HPTDC timing ASIC.

  9. A high-throughput, multi-channel photon-counting detector with picosecond timing

    International Nuclear Information System (INIS)

    Lapington, J.S.; Fraser, G.W.; Miller, G.M.; Ashton, T.J.R.; Jarron, P.; Despeisse, M.; Powolny, F.; Howorth, J.; Milnes, J.

    2009-01-01

    High-throughput photon counting with high time resolution is a niche application area where vacuum tubes can still outperform solid-state devices. Applications in the life sciences utilizing time-resolved spectroscopies, particularly in the growing field of proteomics, will benefit greatly from performance enhancements in event timing and detector throughput. The HiContent project is a collaboration between the University of Leicester Space Research Centre, the Microelectronics Group at CERN, Photek Ltd., and end-users at the Gray Cancer Institute and the University of Manchester. The goal is to develop a detector system specifically designed for optical proteomics, capable of high content (multi-parametric) analysis at high throughput. The HiContent detector system is being developed to exploit this niche market. It combines multi-channel, high time resolution photon counting in a single miniaturized detector system with integrated electronics. The combination of enabling technologies; small pore microchannel plate devices with very high time resolution, and high-speed multi-channel ASIC electronics developed for the LHC at CERN, provides the necessary building blocks for a high-throughput detector system with up to 1024 parallel counting channels and 20 ps time resolution. We describe the detector and electronic design, discuss the current status of the HiContent project and present the results from a 64-channel prototype system. In the absence of an operational detector, we present measurements of the electronics performance using a pulse generator to simulate detector events. Event timing results from the NINO high-speed front-end ASIC captured using a fast digital oscilloscope are compared with data taken with the proposed electronic configuration which uses the multi-channel HPTDC timing ASIC.

  10. Population-based CD4 counts in a rural area in South Africa with high HIV prevalence and high antiretroviral treatment coverage.

    Directory of Open Access Journals (Sweden)

    Abraham Malaza

    Full Text Available Little is known about the variability of CD4 counts in the general population of sub-Saharan Africa countries affected by the HIV epidemic. We investigated factors associated with CD4 counts in a rural area in South Africa with high HIV prevalence and high antiretroviral treatment (ART coverage.CD4 counts, health status, body mass index (BMI, demographic characteristics and HIV status were assessed in 4990 adult resident participants of a demographic surveillance in rural KwaZulu-Natal in South Africa; antiretroviral treatment duration was obtained from a linked clinical database. Multivariable regression analysis, overall and stratified by HIV status, was performed with CD4 count levels as outcome.Median CD4 counts were significantly higher in women than in men overall (714 vs. 630 cells/µl, p<0.0001, both in HIV-uninfected (833 vs. 683 cells/µl, p<0.0001 and HIV-infected adults (384.5 vs. 333 cells/µl, p<0.0001. In multivariable regression analysis, women had 19.4% (95% confidence interval (CI 16.1-22.9 higher CD4 counts than men, controlling for age, HIV status, urban/rural residence, household wealth, education, BMI, self-reported tuberculosis, high blood pressure, other chronic illnesses and sample processing delay. At ART initiation, HIV-infected adults had 21.7% (95% CI 14.6-28.2 lower CD4 counts than treatment-naive individuals; CD4 counts were estimated to increase by 9.2% (95% CI 6.2-12.4 per year of treatment.CD4 counts are primarily determined by sex in HIV-uninfected adults, and by sex, age and duration of antiretroviral treatment in HIV-infected adults. Lower CD4 counts at ART initiation in men could be a consequence of lower CD4 cell counts before HIV acquisition.

  11. High Triglycerides Are Associated with Low Thrombocyte Counts and High VEGF in Nephropathia Epidemica.

    Science.gov (United States)

    Martynova, Ekaterina V; Valiullina, Aygul H; Gusev, Oleg A; Davidyuk, Yuriy N; Garanina, Ekaterina E; Shakirova, Venera G; Khaertynova, Ilsiyar; Anokhin, Vladimir A; Rizvanov, Albert A; Khaiboullina, Svetlana F

    2016-01-01

    Nephropathia epidemica (NE) is a mild form of hemorrhagic fever with renal syndrome. Several reports have demonstrated a severe alteration in lipoprotein metabolism. However, little is known about changes in circulating lipids in NE. The objectives of this study were to evaluate changes in serum total cholesterol, high density cholesterol (HDCL), and triglycerides. In addition to evaluation of serum cytokine activation associations, changes in lipid profile and cytokine activation were determined for gender, thrombocyte counts, and VEGF. Elevated levels of triglycerides and decreased HDCL were observed in NE, while total cholesterol did not differ from controls. High triglycerides were associated with both the lowest thrombocyte counts and high serum VEGF, as well as a high severity score. Additionally, there were higher levels of triglycerides in male than female NE patients. Low triglycerides were associated with upregulation of IFN- γ and IL-12, suggesting activation of Th1 helper cells. Furthermore, levels of IFN- γ and IL-12 were increased in patients with lower severity scores, suggesting that a Th1 type immune response is playing protective role in NE. These combined data advance the understanding of NE pathogenesis and indicate a role for high triglycerides in disease severity.

  12. High rate 4π β-γ coincidence counting system

    International Nuclear Information System (INIS)

    Johnson, L.O.; Gehrke, R.J.

    1978-01-01

    A high count rate 4π β-γ coincidence counting system for the determination of absolute disintegration rates of short half-life radionuclides is described. With this system the dead time per pulse is minimized by not stretching any pulses beyond the width necessary to satisfy overlap coincidence requirements. The equations used to correct for the β, γ, and coincidence channel dead times and for accidental coincidences are presented but not rigorously developed. Experimental results are presented for a decaying source of 56 Mn initially at 2 x 10 6 d/s and a set of 60 Co sources of accurately known source strengths varying from 10 3 to 2 x 10 6 d/s. A check of the accidental coincidence equation for the case of two independent sources with varying source strengths is presented

  13. Exact Local Correlations and Full Counting Statistics for Arbitrary States of the One-Dimensional Interacting Bose Gas

    Science.gov (United States)

    Bastianello, Alvise; Piroli, Lorenzo; Calabrese, Pasquale

    2018-05-01

    We derive exact analytic expressions for the n -body local correlations in the one-dimensional Bose gas with contact repulsive interactions (Lieb-Liniger model) in the thermodynamic limit. Our results are valid for arbitrary states of the model, including ground and thermal states, stationary states after a quantum quench, and nonequilibrium steady states arising in transport settings. Calculations for these states are explicitly presented and physical consequences are critically discussed. We also show that the n -body local correlations are directly related to the full counting statistics for the particle-number fluctuations in a short interval, for which we provide an explicit analytic result.

  14. Optimization of counting time using count statistics on a diffraction beamline

    Energy Technology Data Exchange (ETDEWEB)

    Marais, D., E-mail: Deon.Marais@necsa.co.za [Research and Development Division, South African Nuclear Energy Corporation (Necsa) SOC Limited, PO Box 582, Pretoria 0001 (South Africa); School of Mechanical and Nuclear Engineering, North-West University, Potchefstroom 2520 (South Africa); Venter, A.M., E-mail: Andrew.Venter@necsa.co.za [Research and Development Division, South African Nuclear Energy Corporation (Necsa) SOC Limited, PO Box 582, Pretoria 0001 (South Africa); Faculty of Agriculture Science and Technology, North-West University, Mahikeng 2790 (South Africa); Markgraaff, J., E-mail: Johan.Markgraaff@nwu.ac.za [School of Mechanical and Nuclear Engineering, North-West University, Potchefstroom 2520 (South Africa)

    2016-05-11

    The feasibility of an alternative data acquisition strategy to improve the efficiency of beam time usage with neutron strain scanner instruments is demonstrated. By performing strain measurements against set statistical criteria, rather than time, not only leads to substantially reduced sample investigation time but also renders data of similar quality throughout.

  15. Medicaid Drug Claims Statistics

    Data.gov (United States)

    U.S. Department of Health & Human Services — The Medicaid Drug Claims Statistics CD is a useful tool that conveniently breaks up Medicaid claim counts and separates them by quarter and includes an annual count.

  16. An automated approach for annual layer counting in ice cores

    Science.gov (United States)

    Winstrup, M.; Svensson, A.; Rasmussen, S. O.; Winther, O.; Steig, E.; Axelrod, A.

    2012-04-01

    The temporal resolution of some ice cores is sufficient to preserve seasonal information in the ice core record. In such cases, annual layer counting represents one of the most accurate methods to produce a chronology for the core. Yet, manual layer counting is a tedious and sometimes ambiguous job. As reliable layer recognition becomes more difficult, a manual approach increasingly relies on human interpretation of the available data. Thus, much may be gained by an automated and therefore objective approach for annual layer identification in ice cores. We have developed a novel method for automated annual layer counting in ice cores, which relies on Bayesian statistics. It uses algorithms from the statistical framework of Hidden Markov Models (HMM), originally developed for use in machine speech recognition. The strength of this layer detection algorithm lies in the way it is able to imitate the manual procedures for annual layer counting, while being based on purely objective criteria for annual layer identification. With this methodology, it is possible to determine the most likely position of multiple layer boundaries in an entire section of ice core data at once. It provides a probabilistic uncertainty estimate of the resulting layer count, hence ensuring a proper treatment of ambiguous layer boundaries in the data. Furthermore multiple data series can be incorporated to be used at once, hence allowing for a full multi-parameter annual layer counting method similar to a manual approach. In this study, the automated layer counting algorithm has been applied to data from the NGRIP ice core, Greenland. The NGRIP ice core has very high temporal resolution with depth, and hence the potential to be dated by annual layer counting far back in time. In previous studies [Andersen et al., 2006; Svensson et al., 2008], manual layer counting has been carried out back to 60 kyr BP. A comparison between the counted annual layers based on the two approaches will be presented

  17. Atom-counting in High Resolution Electron Microscopy:TEM or STEM - That's the question.

    Science.gov (United States)

    Gonnissen, J; De Backer, A; den Dekker, A J; Sijbers, J; Van Aert, S

    2017-03-01

    In this work, a recently developed quantitative approach based on the principles of detection theory is used in order to determine the possibilities and limitations of High Resolution Scanning Transmission Electron Microscopy (HR STEM) and HR TEM for atom-counting. So far, HR STEM has been shown to be an appropriate imaging mode to count the number of atoms in a projected atomic column. Recently, it has been demonstrated that HR TEM, when using negative spherical aberration imaging, is suitable for atom-counting as well. The capabilities of both imaging techniques are investigated and compared using the probability of error as a criterion. It is shown that for the same incoming electron dose, HR STEM outperforms HR TEM under common practice standards, i.e. when the decision is based on the probability function of the peak intensities in HR TEM and of the scattering cross-sections in HR STEM. If the atom-counting decision is based on the joint probability function of the image pixel values, the dependence of all image pixel intensities as a function of thickness should be known accurately. Under this assumption, the probability of error may decrease significantly for atom-counting in HR TEM and may, in theory, become lower as compared to HR STEM under the predicted optimal experimental settings. However, the commonly used standard for atom-counting in HR STEM leads to a high performance and has been shown to work in practice. Copyright © 2017 Elsevier B.V. All rights reserved.

  18. Statistical precision of delayed-neutron nondestructive assay techniques

    International Nuclear Information System (INIS)

    Bayne, C.K.; McNeany, S.R.

    1979-02-01

    A theoretical analysis of the statistical precision of delayed-neutron nondestructive assay instruments is presented. Such instruments measure the fissile content of nuclear fuel samples by neutron irradiation and delayed-neutron detection. The precision of these techniques is limited by the statistical nature of the nuclear decay process, but the precision can be optimized by proper selection of system operating parameters. Our method is a three-part analysis. We first present differential--difference equations describing the fundamental physics of the measurements. We then derive and present complete analytical solutions to these equations. Final equations governing the expected number and variance of delayed-neutron counts were computer programmed to calculate the relative statistical precision of specific system operating parameters. Our results show that Poisson statistics do not govern the number of counts accumulated in multiple irradiation-count cycles and that, in general, maximum count precision does not correspond with maximum count as first expected. Covariance between the counts of individual cycles must be considered in determining the optimum number of irradiation-count cycles and the optimum irradiation-to-count time ratio. For the assay system in use at ORNL, covariance effects are small, but for systems with short irradiation-to-count transition times, covariance effects force the optimum number of irradiation-count cycles to be half those giving maximum count. We conclude that the equations governing the expected value and variance of delayed-neutron counts have been derived in closed form. These have been computerized and can be used to select optimum operating parameters for delayed-neutron assay devices

  19. Counts-in-Cylinders in the Sloan Digital Sky Survey with Comparisons to N-Body

    Energy Technology Data Exchange (ETDEWEB)

    Berrier, Heather D.; Barton, Elizabeth J.; /UC, Irvine; Berrier, Joel C.; /Arkansas U.; Bullock, James S.; /UC, Irvine; Zentner, Andrew R.; /Pittsburgh U.; Wechsler, Risa H. /KIPAC, Menlo Park /SLAC

    2010-12-16

    Environmental statistics provide a necessary means of comparing the properties of galaxies in different environments and a vital test of models of galaxy formation within the prevailing, hierarchical cosmological model. We explore counts-in-cylinders, a common statistic defined as the number of companions of a particular galaxy found within a given projected radius and redshift interval. Galaxy distributions with the same two-point correlation functions do not necessarily have the same companion count distributions. We use this statistic to examine the environments of galaxies in the Sloan Digital Sky Survey, Data Release 4. We also make preliminary comparisons to four models for the spatial distributions of galaxies, based on N-body simulations, and data from SDSS DR4 to study the utility of the counts-in-cylinders statistic. There is a very large scatter between the number of companions a galaxy has and the mass of its parent dark matter halo and the halo occupation, limiting the utility of this statistic for certain kinds of environmental studies. We also show that prevalent, empirical models of galaxy clustering that match observed two- and three-point clustering statistics well fail to reproduce some aspects of the observed distribution of counts-in-cylinders on 1, 3 and 6-h{sup -1}Mpc scales. All models that we explore underpredict the fraction of galaxies with few or no companions in 3 and 6-h{sup -1} Mpc cylinders. Roughly 7% of galaxies in the real universe are significantly more isolated within a 6 h{sup -1} Mpc cylinder than the galaxies in any of the models we use. Simple, phenomenological models that map galaxies to dark matter halos fail to reproduce high-order clustering statistics in low-density environments.

  20. A comparative study on the mast cells count in oral squamous cell carcinoma and normal oral mucosa

    Directory of Open Access Journals (Sweden)

    Mahsa Dastpak

    2015-03-01

    Full Text Available Introduction: Oral squamous cell carcinoma (OSCC is one of the 10 most common malignant tumors and SCC accounts 94% of all oral malignancies. Mast cells are regarded as complex and multifunctional cells, playing a significant role in immunopathology . The aim of this study is to evaluate the number of mast cells in tissue sections of oral squamous cell carcinoma (OSCC in comparison with normal mucosa. Materials & Methods: Sixty paraffin-embedded specimens were obtained from the archives of the Department of Oral and Maxillofacial Pathology,dental school of Babol university of medical science (15 high grade,15 low grade and 30 Iritation Fibroma. Classification of OSCC cases was according to the BRODER`S malignancy grading system. Hematoxylin and Eosin-stained slides were re-evaluated before entering the samples in our study. Toluidine blue(1% staining was used to identify Mast cells in samples . We used SPSS software version 18 and one way ANOVA test for analyzing data. Results: The highest mast cell count was seen in normal tissue and it was higher in low grade OSCC in comparison with high grade, but the differences between groups weren’t statistically significant. The Mean count of mast cell between OSCC and normal oral mucosa was statistically significant different(p=0.019.We didn’t observe any statistically significant difference between Mast cell counts of control group and low grade OSCC . The same result was seen between high garde and low grade OSCC . The Mean mast cell count difference between male and female groups weren’t statistically significant. The Mean mast cell count difference between high grade OSCC and control group was significant (p<0.05. Conclusion: According to the results, the average amount of mast cells decreased in OSCC in comparison with normal oral mucosa . It does not seem that mast cells play an important role in tumor progression, although further study is needed. 

  1. An Odyssey of Connecticut's Children: KIDS COUNT Data Book 2001.

    Science.gov (United States)

    Sampson, Amy E.

    This Kids Count Data Book provides state and regional trends in the well-being of Connecticut's children. The statistical portrait is based on 19 indicators of well-being: (1) children in families receiving welfare; (2) children receiving free or reduced-price meals; (3) high school employment; (4) births to teen mothers; (5) low birth weight; (6)…

  2. Compton suppression gamma-counting: The effect of count rate

    Science.gov (United States)

    Millard, H.T.

    1984-01-01

    Past research has shown that anti-coincidence shielded Ge(Li) spectrometers enhanced the signal-to-background ratios for gamma-photopeaks, which are situated on high Compton backgrounds. Ordinarily, an anti- or non-coincidence spectrum (A) and a coincidence spectrum (C) are collected simultaneously with these systems. To be useful in neutron activation analysis (NAA), the fractions of the photopeak counts routed to the two spectra must be constant from sample to sample to variations must be corrected quantitatively. Most Compton suppression counting has been done at low count rate, but in NAA applications, count rates may be much higher. To operate over the wider dynamic range, the effect of count rate on the ratio of the photopeak counts in the two spectra (A/C) was studied. It was found that as the count rate increases, A/C decreases for gammas not coincident with other gammas from the same decay. For gammas coincident with other gammas, A/C increases to a maximum and then decreases. These results suggest that calibration curves are required to correct photopeak areas so quantitative data can be obtained at higher count rates. ?? 1984.

  3. Relationship of milking rate to somatic cell count.

    Science.gov (United States)

    Brown, C A; Rischette, S J; Schultz, L H

    1986-03-01

    Information on milking rate, monthly bucket somatic cell counts, mastitis treatment, and milk production was obtained from 284 lactations of Holstein cows separated into three lactation groups. Significant correlations between somatic cell count (linear score) and other parameters included production in lactation 1 (-.185), production in lactation 2 (-.267), and percent 2-min milk in lactation 2 (.251). Somatic cell count tended to increase with maximum milking rate in all lactations, but correlations were not statistically significant. Twenty-nine percent of cows with milking rate measurements were treated for clinical mastitis. Treated cows in each lactation group produced less milk than untreated cows. In the second and third lactation groups, treated cows had a shorter total milking time and a higher percent 2-min milk than untreated cows, but differences were not statistically significant. Overall, the data support the concept that faster milking cows tend to have higher cell counts and more mastitis treatments, particularly beyond first lactation. However, the magnitude of the relationship was small.

  4. An integral whole circuit of amplifying and discriminating suited to high counting rate

    International Nuclear Information System (INIS)

    Dong Chengfu; Su Hong; Wu Ming; Li Xiaogang; Peng Yu; Qian Yi; Liu Yicai; Xu Sijiu; Ma Xiaoli

    2007-01-01

    A hybrid circuit consists of charge sensitive preamplifier, main amplifier, discriminator and shaping circuit was described. This instrument has characteristics of low power consumption, small volume, high sensitivity, potable and so on, and is convenient for use in field. The output pulse of this instrument may directly consist with CMOS or TTL logic level. This instrument was mainly used for count measurement, for example, for high sensitive 3 He neutron detector, meanwhile also may used for other heavy ion detectors, the highest counting rate can reach 10 6 /s. (authors)

  5. A multiwire proportional counter for very high counting rates

    International Nuclear Information System (INIS)

    Barbosa, A.F.; Guedes, G.P.; Tamura, E.; Pepe, I.M.; Oliveira, N.B.

    1997-12-01

    Preliminary measurements in a proportional counter with two independently counting wires showed that counting rates up to 10 6 counts/s per wire can be reached without critical loss in the true versus measured linearity relation. Results obtained with a detector containing 30 active wires (2 mm pitch) are presented. To each wire is associated a fast pre-amplifier and a discriminator channel. Global counting rates in excess to 10 7 events/s are reported. Data acquisition systems are described for 1D (real time) and 2D (off-line) position sensitive detection systems. (author)

  6. EcoCount

    Directory of Open Access Journals (Sweden)

    Phillip P. Allen

    2014-05-01

    Full Text Available Techniques that analyze biological remains from sediment sequences for environmental reconstructions are well established and widely used. Yet, identifying, counting, and recording biological evidence such as pollen grains remain a highly skilled, demanding, and time-consuming task. Standard procedure requires the classification and recording of between 300 and 500 pollen grains from each representative sample. Recording the data from a pollen count requires significant effort and focused resources from the palynologist. However, when an adaptation to the recording procedure is utilized, efficiency and time economy improve. We describe EcoCount, which represents a development in environmental data recording procedure. EcoCount is a voice activated fully customizable digital count sheet that allows the investigator to continuously interact with a field of view during the data recording. Continuous viewing allows the palynologist the opportunity to remain engaged with the essential task, identification, for longer, making pollen counting more efficient and economical. EcoCount is a versatile software package that can be used to record a variety of environmental evidence and can be installed onto different computer platforms, making the adoption by users and laboratories simple and inexpensive. The user-friendly format of EcoCount allows any novice to be competent and functional in a very short time.

  7. Intuitive introductory statistics

    CERN Document Server

    Wolfe, Douglas A

    2017-01-01

    This textbook is designed to give an engaging introduction to statistics and the art of data analysis. The unique scope includes, but also goes beyond, classical methodology associated with the normal distribution. What if the normal model is not valid for a particular data set? This cutting-edge approach provides the alternatives. It is an introduction to the world and possibilities of statistics that uses exercises, computer analyses, and simulations throughout the core lessons. These elementary statistical methods are intuitive. Counting and ranking features prominently in the text. Nonparametric methods, for instance, are often based on counts and ranks and are very easy to integrate into an introductory course. The ease of computation with advanced calculators and statistical software, both of which factor into this text, allows important techniques to be introduced earlier in the study of statistics. This book's novel scope also includes measuring symmetry with Walsh averages, finding a nonp...

  8. Mapping the layer count of few-layer hexagonal boron nitride at high lateral spatial resolutions

    Science.gov (United States)

    Mohsin, Ali; Cross, Nicholas G.; Liu, Lei; Watanabe, Kenji; Taniguchi, Takashi; Duscher, Gerd; Gu, Gong

    2018-01-01

    Layer count control and uniformity of two dimensional (2D) layered materials are critical to the investigation of their properties and to their electronic device applications, but methods to map 2D material layer count at nanometer-level lateral spatial resolutions have been lacking. Here, we demonstrate a method based on two complementary techniques widely available in transmission electron microscopes (TEMs) to map the layer count of multilayer hexagonal boron nitride (h-BN) films. The mass-thickness contrast in high-angle annular dark-field (HAADF) imaging in the scanning transmission electron microscope (STEM) mode allows for thickness determination in atomically clean regions with high spatial resolution (sub-nanometer), but is limited by surface contamination. To complement, another technique based on the boron K ionization edge in the electron energy loss spectroscopy spectrum (EELS) of h-BN is developed to quantify the layer count so that surface contamination does not cause an overestimate, albeit at a lower spatial resolution (nanometers). The two techniques agree remarkably well in atomically clean regions with discrepancies within  ±1 layer. For the first time, the layer count uniformity on the scale of nanometers is quantified for a 2D material. The methodology is applicable to layer count mapping of other 2D layered materials, paving the way toward the synthesis of multilayer 2D materials with homogeneous layer count.

  9. Hanford whole body counting manual

    International Nuclear Information System (INIS)

    Palmer, H.E.; Rieksts, G.A.; Lynch, T.P.

    1990-06-01

    This document describes the Hanford Whole Body Counting Program as it is administered by Pacific Northwest Laboratory (PNL) in support of the US Department of Energy--Richland Operations Office (DOE-RL) and its Hanford contractors. Program services include providing in vivo measurements of internally deposited radioactivity in Hanford employees (or visitors). Specific chapters of this manual deal with the following subjects: program operational charter, authority, administration, and practices, including interpreting applicable DOE Orders, regulations, and guidance into criteria for in vivo measurement frequency, etc., for the plant-wide whole body counting services; state-of-the-art facilities and equipment used to provide the best in vivo measurement results possible for the approximately 11,000 measurements made annually; procedures for performing the various in vivo measurements at the Whole Body Counter (WBC) and related facilities including whole body counts; operation and maintenance of counting equipment, quality assurance provisions of the program, WBC data processing functions, statistical aspects of in vivo measurements, and whole body counting records and associated guidance documents. 16 refs., 48 figs., 22 tabs

  10. Amplitude distributions of dark counts and photon counts in NbN superconducting single-photon detectors integrated with the HEMT readout

    Energy Technology Data Exchange (ETDEWEB)

    Kitaygorsky, J. [Kavli Institute of Nanoscience Delft, Delft University of Technology, 2600 GA Delft (Netherlands); Department of Electrical and Computer Engineering and Laboratory for Laser Energetics, University of Rochester, Rochester, NY 14627-0231 (United States); Słysz, W., E-mail: wslysz@ite.waw.pl [Institute of Electron Technology, PL-02 668 Warsaw (Poland); Shouten, R.; Dorenbos, S.; Reiger, E.; Zwiller, V. [Kavli Institute of Nanoscience Delft, Delft University of Technology, 2600 GA Delft (Netherlands); Sobolewski, Roman [Department of Electrical and Computer Engineering and Laboratory for Laser Energetics, University of Rochester, Rochester, NY 14627-0231 (United States)

    2017-01-15

    Highlights: • A new operation regime of NbN superconducting single-photon detectors (SSPDs). • A better understanding of the origin of dark counts generated by the detector. • A promise of PNR functionality in SSPD measurements. - Abstract: We present a new operation regime of NbN superconducting single-photon detectors (SSPDs) by integrating them with a low-noise cryogenic high-electron-mobility transistor and a high-load resistor. The integrated sensors are designed to get a better understanding of the origin of dark counts triggered by the detector, as our scheme allows us to distinguish the origin of dark pulses from the actual photon pulses in SSPDs. The presented approach is based on a statistical analysis of amplitude distributions of recorded trains of the SSPD photoresponse transients. It also enables to obtain information on energy of the incident photons, as well as demonstrates some photon-number-resolving capability of meander-type SSPDs.

  11. A multiwire proportional counter for very high counting rates

    Energy Technology Data Exchange (ETDEWEB)

    Barbosa, A F; Guedes, G P [Centro Brasileiro de Pesquisas Fisicas (CBPF), Rio de Janeiro, RJ (Brazil); Tamura, E [Laboratorio Nacional de Luz Sincrotron (LNLS), Campinas, SP (Brazil); Pepe, I M; Oliveira, N B [Bahia Univ., Salvador, BA (Brazil). Inst. de Fisica

    1997-12-01

    Preliminary measurements in a proportional counter with two independently counting wires showed that counting rates up to 10{sup 6} counts/s per wire can be reached without critical loss in the true versus measured linearity relation. Results obtained with a detector containing 30 active wires (2 mm pitch) are presented. To each wire is associated a fast pre-amplifier and a discriminator channel. Global counting rates in excess to 10{sup 7} events/s are reported. Data acquisition systems are described for 1D (real time) and 2D (off-line) position sensitive detection systems. (author) 13 refs., 6 figs.

  12. High-Throughput Quantification of Bacterial-Cell Interactions Using Virtual Colony Counts

    Directory of Open Access Journals (Sweden)

    Stefanie Hoffmann

    2018-02-01

    Full Text Available The quantification of bacteria in cell culture infection models is of paramount importance for the characterization of host-pathogen interactions and pathogenicity factors involved. The standard to enumerate bacteria in these assays is plating of a dilution series on solid agar and counting of the resulting colony forming units (CFU. In contrast, the virtual colony count (VCC method is a high-throughput compatible alternative with minimized manual input. Based on the recording of quantitative growth kinetics, VCC relates the time to reach a given absorbance threshold to the initial cell count using a series of calibration curves. Here, we adapted the VCC method using the model organism Salmonella enterica sv. Typhimurium (S. Typhimurium in combination with established cell culture-based infection models. For HeLa infections, a direct side-by-side comparison showed a good correlation of VCC with CFU counting after plating. For MDCK cells and RAW macrophages we found that VCC reproduced the expected phenotypes of different S. Typhimurium mutants. Furthermore, we demonstrated the use of VCC to test the inhibition of Salmonella invasion by the probiotic E. coli strain Nissle 1917. Taken together, VCC provides a flexible, label-free, automation-compatible methodology to quantify bacteria in in vitro infection assays.

  13. Use of Statistics for Data Evaluation in Environmental Radioactivity Measurements

    International Nuclear Information System (INIS)

    Sutarman

    2001-01-01

    Counting statistics will give a correction on environmental radioactivity measurement result. Statistics provides formulas to determine standard deviation (S B ) and minimum detectable concentration (MDC) according to the Poisson distribution. Both formulas depend on the background count rate, counting time, counting efficiency, gamma intensity, and sample size. A long time background counting results in relatively low S B and MDC that can present relatively accurate measurement results. (author)

  14. COUNTS-IN-CYLINDERS IN THE SLOAN DIGITAL SKY SURVEY WITH COMPARISONS TO N-BODY SIMULATIONS

    International Nuclear Information System (INIS)

    Berrier, Heather D.; Barton, Elizabeth J.; Bullock, James S.; Berrier, Joel C.; Zentner, Andrew R.; Wechsler, Risa H.

    2011-01-01

    Environmental statistics provide a necessary means of comparing the properties of galaxies in different environments, and a vital test of models of galaxy formation within the prevailing hierarchical cosmological model. We explore counts-in-cylinders, a common statistic defined as the number of companions of a particular galaxy found within a given projected radius and redshift interval. Galaxy distributions with the same two-point correlation functions do not necessarily have the same companion count distributions. We use this statistic to examine the environments of galaxies in the Sloan Digital Sky Survey Data Release 4 (SDSS DR4). We also make preliminary comparisons to four models for the spatial distributions of galaxies, based on N-body simulations and data from SDSS DR4, to study the utility of the counts-in-cylinders statistic. There is a very large scatter between the number of companions a galaxy has and the mass of its parent dark matter halo and the halo occupation, limiting the utility of this statistic for certain kinds of environmental studies. We also show that prevalent empirical models of galaxy clustering, that match observed two- and three-point clustering statistics well, fail to reproduce some aspects of the observed distribution of counts-in-cylinders on 1, 3, and 6 h -1 Mpc scales. All models that we explore underpredict the fraction of galaxies with few or no companions in 3 and 6 h -1 Mpc cylinders. Roughly 7% of galaxies in the real universe are significantly more isolated within a 6 h -1 Mpc cylinder than the galaxies in any of the models we use. Simple phenomenological models that map galaxies to dark matter halos fail to reproduce high-order clustering statistics in low-density environments.

  15. Comparing distribution models for small samples of overdispersed counts of freshwater fish

    Science.gov (United States)

    Vaudor, Lise; Lamouroux, Nicolas; Olivier, Jean-Michel

    2011-05-01

    The study of species abundance often relies on repeated abundance counts whose number is limited by logistic or financial constraints. The distribution of abundance counts is generally right-skewed (i.e. with many zeros and few high values) and needs to be modelled for statistical inference. We used an extensive dataset involving about 100,000 fish individuals of 12 freshwater fish species collected in electrofishing points (7 m 2) during 350 field surveys made in 25 stream sites, in order to compare the performance and the generality of four distribution models of counts (Poisson, negative binomial and their zero-inflated counterparts). The negative binomial distribution was the best model (Bayesian Information Criterion) for 58% of the samples (species-survey combinations) and was suitable for a variety of life histories, habitat, and sample characteristics. The performance of the models was closely related to samples' statistics such as total abundance and variance. Finally, we illustrated the consequences of a distribution assumption by calculating confidence intervals around the mean abundance, either based on the most suitable distribution assumption or on an asymptotical, distribution-free (Student's) method. Student's method generally corresponded to narrower confidence intervals, especially when there were few (≤3) non-null counts in the samples.

  16. VSRR Provisional Drug Overdose Death Counts

    Data.gov (United States)

    U.S. Department of Health & Human Services — This data contains provisional counts for drug overdose deaths based on a current flow of mortality data in the National Vital Statistics System. National...

  17. Calibration of the Accuscan II IN Vivo System for High Energy Lung Counting

    Energy Technology Data Exchange (ETDEWEB)

    Ovard R. Perry; David L. Georgeson

    2011-07-01

    This report describes the April 2011 calibration of the Accuscan II HpGe In Vivo system for high energy lung counting. The source used for the calibration was a NIST traceable lung set manufactured at the University of Cincinnati UCLL43AMEU & UCSL43AMEU containing Am-241 and Eu-152 with energies from 26 keV to 1408 keV. The lung set was used in conjunction with a Realistic Torso phantom. The phantom was placed on the RMC II counting table (with pins removed) between the v-ridges on the backwall of the Accuscan II counter. The top of the detector housing was positioned perpendicular to the junction of the phantom clavicle with the sternum. This position places the approximate center line of the detector housing with the center of the lungs. The energy and efficiency calibrations were performed using a Realistic Torso phantom (Appendix I) and the University of Cincinnati lung set. This report includes an overview introduction and records for the energy/FWHM and efficiency calibration including performance verification and validation counting. The Accuscan II system was successfully calibrated for high energy lung counting and verified in accordance with ANSI/HPS N13.30-1996 criteria.

  18. Bayesian inference from count data using discrete uniform priors.

    Directory of Open Access Journals (Sweden)

    Federico Comoglio

    Full Text Available We consider a set of sample counts obtained by sampling arbitrary fractions of a finite volume containing an homogeneously dispersed population of identical objects. We report a Bayesian derivation of the posterior probability distribution of the population size using a binomial likelihood and non-conjugate, discrete uniform priors under sampling with or without replacement. Our derivation yields a computationally feasible formula that can prove useful in a variety of statistical problems involving absolute quantification under uncertainty. We implemented our algorithm in the R package dupiR and compared it with a previously proposed Bayesian method based on a Gamma prior. As a showcase, we demonstrate that our inference framework can be used to estimate bacterial survival curves from measurements characterized by extremely low or zero counts and rather high sampling fractions. All in all, we provide a versatile, general purpose algorithm to infer population sizes from count data, which can find application in a broad spectrum of biological and physical problems.

  19. Photon Counting System for High-Sensitivity Detection of Bioluminescence at Optical Fiber End.

    Science.gov (United States)

    Iinuma, Masataka; Kadoya, Yutaka; Kuroda, Akio

    2016-01-01

    The technique of photon counting is widely used for various fields and also applicable to a high-sensitivity detection of luminescence. Thanks to recent development of single photon detectors with avalanche photodiodes (APDs), the photon counting system with an optical fiber has become powerful for a detection of bioluminescence at an optical fiber end, because it allows us to fully use the merits of compactness, simple operation, highly quantum efficiency of the APD detectors. This optical fiber-based system also has a possibility of improving the sensitivity to a local detection of Adenosine triphosphate (ATP) by high-sensitivity detection of the bioluminescence. In this chapter, we are introducing a basic concept of the optical fiber-based system and explaining how to construct and use this system.

  20. Comment on: 'A Poisson resampling method for simulating reduced counts in nuclear medicine images'.

    Science.gov (United States)

    de Nijs, Robin

    2015-07-21

    In order to be able to calculate half-count images from already acquired data, White and Lawson published their method based on Poisson resampling. They verified their method experimentally by measurements with a Co-57 flood source. In this comment their results are reproduced and confirmed by a direct numerical simulation in Matlab. Not only Poisson resampling, but also two direct redrawing methods were investigated. Redrawing methods were based on a Poisson and a Gaussian distribution. Mean, standard deviation, skewness and excess kurtosis half-count/full-count ratios were determined for all methods, and compared to the theoretical values for a Poisson distribution. Statistical parameters showed the same behavior as in the original note and showed the superiority of the Poisson resampling method. Rounding off before saving of the half count image had a severe impact on counting statistics for counts below 100. Only Poisson resampling was not affected by this, while Gaussian redrawing was less affected by it than Poisson redrawing. Poisson resampling is the method of choice, when simulating half-count (or less) images from full-count images. It simulates correctly the statistical properties, also in the case of rounding off of the images.

  1. Statistical distribution for generalized ideal gas of fractional-statistics particles

    International Nuclear Information System (INIS)

    Wu, Y.

    1994-01-01

    We derive the occupation-number distribution in a generalized ideal gas of particles obeying fractional statistics, including mutual statistics, by adopting a state-counting definition. When there is no mutual statistics, the statistical distribution interpolates between bosons and fermions, and respects a fractional exclusion principle (except for bosons). Anyons in a strong magnetic field at low temperatures constitute such a physical system. Applications to the thermodynamic properties of quasiparticle excitations in the Laughlin quantum Hall fluid are discussed

  2. Method of inspecting Raschig rings by neutron absorption counting

    International Nuclear Information System (INIS)

    Morris, R.N.; Murri, R.L.; Hume, M.W.

    1979-01-01

    A neutron counting method for inspecting borosilicate glass Raschig rings and an apparatus designed specifically for this method are discussed. The neutron count ratios for rings of a given thickness show a linear correlation to the boron oxide content of the rings. The count ratio also has a linear relationship to the thickness of rings of a given boron oxide content. Consequently, the experimentally-determined count ratio and physically-measured thickness of Raschig rings can be used to statistically predict their boron oxide content and determine whether or not they meet quality control acceptance criteria

  3. Photon-counting digital radiography using high-pressure xenon filled detectors

    CERN Document Server

    Li, Maozhen; Johns, P C

    2001-01-01

    Digital radiography overcomes many of the limitations of the traditional screen/film system. Further enhancements in the digital radiography image are possible if the X-ray image receptor could measure the energy of individual photons instead of simply integrating their energy, as is the case at present. A prototype photon counting scanned projection radiography system has been constructed, which combines a Gas Electron Multiplier (GEM) and a Gas Microstrip Detector (GMD) using Xe : CH sub 4 (90 : 10) at high pressure. With the gain contribution from the GEM, the GMD can be operated at lower and safer voltages making the imaging system more reliable. Good energy resolution, and spatial resolution comparable to that of screen/film, have been demonstrated for the GEM/GMD hybrid imaging system in photon counting mode for X-ray spectra up to 50 kV.

  4. Liver stiffness plus platelet count can be used to exclude high-risk oesophageal varices.

    Science.gov (United States)

    Ding, Nik S; Nguyen, Tin; Iser, David M; Hong, Thai; Flanagan, Emma; Wong, Avelyn; Luiz, Lauren; Tan, Jonathan Y C; Fulforth, James; Holmes, Jacinta; Ryan, Marno; Bell, Sally J; Desmond, Paul V; Roberts, Stuart K; Lubel, John; Kemp, William; Thompson, Alexander J

    2016-02-01

    Endoscopic screening for high-risk gastro-oesophageal varices (GOV) is recommended for compensated cirrhotic patients with transient elastography identifying increasing numbers of patients with cirrhosis without portal hypertension. Using liver stiffness measurement (LSM) ± platelet count, the aim was to develop a simple clinical rule to exclude the presence of high-risk GOV in patients with Child-Pugh A cirrhosis. A retrospective analysis of 71 patients with Child-Pugh A cirrhosis diagnosed by transient elastography (LSM >13.6 kPa) who underwent screening gastroscopy was conducted. A predictive model using LSM ± platelet count was assessed to exclude the presence of high-risk GOV (diameter >5 mm and/or the presence of high-risk stigmata) and validated using a second cohort of 200 patients from two independent centres. High-risk GOV were present in 10 (15%) and 16 (8%) of the training and validation cohorts, respectively, which was associated with LSM and Pl count (P < 0.05). A combined model based on LSM and Pl count was more accurate for excluding the presence of high-risk GOV than either alone (training cohort AUROC: 0.87 [0.77-0.96] vs. 0.78 [0.65-0.92] for LSM and 0.71 [0.52-0.90] for platelets) with the combination of LSM ≤25 kPa and Pl ≥100 having a NPV of 100% in both the training and validation cohorts. A total of 107 (39%) patients meet this criterion. The combination of LSM ≤25 kPa and Pl ≥100 can be used in clinical practice to exclude the presence of high-risk GOV in patients with Child-Pugh A cirrhosis. © 2015 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  5. Title V Permitting Statistics Inventory

    Data.gov (United States)

    U.S. Environmental Protection Agency — The Title V Permitting Statistics Inventory contains measured and estimated nationwide statistical data, consisting of counts of permitted sources, types of permits...

  6. Full-counting statistics of energy transport of molecular junctions in the polaronic regime

    International Nuclear Information System (INIS)

    Tang, Gaomin; Yu, Zhizhou; Wang, Jian

    2017-01-01

    We investigate the full-counting statistics (FCS) of energy transport carried by electrons in molecular junctions for the Anderson–Holstein model in the polaronic regime. Using the two-time quantum measurement scheme, the generating function (GF) for the energy transport is derived and expressed as a Fredholm determinant in terms of Keldysh nonequilibrium Green’s function in the time domain. Dressed tunneling approximation is used in decoupling the phonon cloud operator in the polaronic regime. This formalism enables us to analyze the time evolution of energy transport dynamics after a sudden switch-on of the coupling between the dot and the leads towards the stationary state. The steady state energy current cumulant GF in the long time limit is obtained in the energy domain as well. Universal relations for steady state energy current FCS are derived under a finite temperature gradient with zero bias and this enabled us to express the equilibrium energy current cumulant by a linear combination of lower order cumulants. The behaviors of energy current cumulants in steady state under temperature gradient and external bias are numerically studied and explained. The transient dynamics of energy current cumulants is numerically calculated and analyzed. Universal scaling of normalized transient energy cumulants is found under both temperature gradient and external bias. (paper)

  7. Fast imaging by photon counting application to long-baseline optical stellar interferometry

    International Nuclear Information System (INIS)

    Morel, Sebastien

    1998-01-01

    Image acquisition by photon counting in the visible spectrum with a high precision on photo-events dating is especially useful for ground-based observations. In the first part of this thesis, and after a review of several techniques for photon acquisition and processing, I introduce a new type of photon counting camera, noticeable for its high temporal resolution and its high maximum counting rate: the DELTA (Detector Enhancement by Linear-projections on Three Axes) camera. I describe the concept of this camera, and the engineering solutions (optics, electronics, computing) that could be used for its construction. The second part of my work regards fringe detection and tracking in ground-based and long- baseline optical stellar interferometry. After a statistical approach of the issue, I describe methods introducing a priori information in the data, in order to have a better detection efficiency. One of the proposed methods, using a priori information on the atmospheric piston, requires a precise photo-event dating, and therefore uses methods described in the first part. (author) [fr

  8. Time and position resolution of high granularity, high counting rate MRPC for the inner zone of the CBM-TOF wall

    CERN Document Server

    Petriş, M.

    2016-09-13

    Multi-gap RPC prototypes with readout on a multi-strip electrode were developed for the small polar angle region of the CBM-TOF subdetector, the most demanding zone in terms of granularity and counting rate. The prototypes are based on low resistivity ($\\sim$10$^{10}$ $\\Omega$cm) glass electrodes for performing in high counting rate environment. The strip width/pitch size was chosen such to fulfill the impedance matching with the front-end electronics and the granularity requirements of the innermost zone of the CBM-TOF wall. The in-beam tests using secondary particles produced in heavy ion collisions on a Pb target at SIS18 - GSI Darmstadt and SPS - CERN were focused on the performance of the prototype in conditions similar to the ones expected at SIS100/FAIR. An efficiency larger than 98\\% and a system time resolution in the order of 70~-~80~ps were obtained in high counting rate and high multiplicity environment.

  9. What every radiochemist should know about statistics

    International Nuclear Information System (INIS)

    Nicholson, W.L.

    1994-04-01

    Radionuclide decay and measurement with appropriate counting instruments is one of the few physical processes for which exact mathematical/probabilistic models are available. This paper discusses statistical procedures associated with display and analysis of radionuclide counting data that derive from these exact models. For low count situations the attractiveness of fixed-count-random-time procedures is discussed

  10. Study of the effects of photoelectron statistics on Thomson scattering data

    International Nuclear Information System (INIS)

    Hart, G.W.; Levinton, F.M.; McNeill, D.H.

    1986-01-01

    A computer code has been developed which simulates a Thomson scattering measurement, from the counting statistics of the input channels through the mathematical analysis of the data. The scattered and background signals in each of the wavelength channels are assumed to obey Poisson statistics, and the spectral data are fitted to a Gaussian curve using a nonlinear least-squares fitting algorithm. This method goes beyond the usual calculation of the signal-to-noise ratio for the hardware and gives a quantitative measure of the effect of the noise on the final measurement. This method is applicable to Thomson scattering measurements in which the signal-to-noise ratio is low due to either low signal or high background. Thomson scattering data from the S-1 spheromak have been compared to this simulation, and they have been found to be in good agreement. This code has proven to be useful in assessing the effects of counting statistics relative to shot-to-shot variability in producing the observed spread in the data. It was also useful for designing improvements for the S-1 Thomson scattering system, and this method would be applicable to any measurement affected by counting statistics

  11. Statistical evaluation and measuring strategy for extremely small line shifts

    International Nuclear Information System (INIS)

    Hansen, P.G.

    1978-01-01

    For a measuring situation limited by counting statistics, but where the level of precision is such that possible systematic errors are a major concern, it is proposed to determine the position of a spectral line from a measured line segment by applying a bias correction to the centre of gravity of the segment. This procedure is statistically highly efficient and not sensitive to small errors in assumptions about the line shape. The counting strategy for an instrument that takes data point by point is also considered. It is shown that an optimum (''two-point'') strategy exists; a scan of the central part of the line is 68% efficient by this standard. (Auth.)

  12. Improved Yield, Performance and Reliability of High-Actuator-Count Deformable Mirrors, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — The project team will conduct processing and design research aimed at improving yield, performance, and reliability of high-actuator-count micro-electro-mechanical...

  13. Development of bonded semiconductor device for high counting rate high efficiency photon detectors

    International Nuclear Information System (INIS)

    Kanno, Ikuo

    2008-01-01

    We are trying to decrease dose exposure in medical diagnosis by way of measuring the energy of X-rays. For this purpose, radiation detectors for X-ray energy measurement with high counting rate should be developed. Direct bonding of Si wafers was carried out to make a radiation detector, which had separated X-ray absorber and detector. The resistivity of bonding interface was estimated with the results of four-probe measurements and model calculations. Direct bonding of high resistivity p and n-Si wafers was also performed. The resistance of the pn bonded diode was 0.7 MΩ. The resistance should be increased in the future. (author)

  14. Scalable Intersample Interpolation Architecture for High-channel-count Beamformers

    DEFF Research Database (Denmark)

    Tomov, Borislav Gueorguiev; Nikolov, Svetoslav I; Jensen, Jørgen Arendt

    2011-01-01

    Modern ultrasound scanners utilize digital beamformers that operate on sampled and quantized echo signals. Timing precision is of essence for achieving good focusing. The direct way to achieve it is through the use of high sampling rates, but that is not economical, so interpolation between echo...... samples is used. This paper presents a beamformer architecture that combines a band-pass filter-based interpolation algorithm with the dynamic delay-and-sum focusing of a digital beamformer. The reduction in the number of multiplications relative to a linear perchannel interpolation and band-pass per......-channel interpolation architecture is respectively 58 % and 75 % beamformer for a 256-channel beamformer using 4-tap filters. The approach allows building high channel count beamformers while maintaining high image quality due to the use of sophisticated intersample interpolation....

  15. Gain reduction due to space charge at high counting rates in multiwire proportional chambers

    International Nuclear Information System (INIS)

    Smith, G.C.; Mathieson, E.

    1986-10-01

    Measurements with a small MWPC of gas gain reduction, due to ion space charge at high counting rates, have been compared with theoretical predictions. The quantity ln(q/q 0 )/(q/q 0 ), where (q/q 0 ) is the relative reduced avalanche charge, has been found to be closely proportional to count rate, as predicted. The constant of proportionality is in good agreement with calculations made with a modified version of the original, simplified theory

  16. High counting rate resistive-plate chamber

    International Nuclear Information System (INIS)

    Peskov, V.; Anderson, D.F.; Kwan, S.

    1993-05-01

    Parallel-plate avalanche chambers (PPAC) are widely used in physics experiments because they are fast ( 5 counts/mm 2 . A resistive-plate chamber (RPC) is similar to the PPAC in construction except that one or both of the electrodes are made from high resistivity (≥10 10 Ω·cm) materials. In practice RPCs are usually used in the spark mode. Resistive electrodes are charged by sparks, locally reducing the actual electric field in the gap. The size of the charged surface is about 10 mm 2 , leaving the rest of the detector unaffected. Therefore, the rate capability of such detectors in the spark mode is considerably higher than conventional spark counters. Among the different glasses tested the best results were obtained with electron type conductive glasses, which obey Ohm's law. Most of the work with such glasses was done with high pressure parallel-plate chambers (10 atm) for time-of-flight measurements. Resistive glasses have been expensive and produced only in small quantities. Now resistive glasses are commercially available, although they are still expensive in small scale production. From the positive experience of different groups working with the resistive glasses, it was decided to review the old idea to use this glass for the RPC. This work has investigated the possibility of using the RPC at 1 atm and in the avalanche mode. This has several advantages: simplicity of construction, high rate capability, low voltage operation, and the ability to work with non-flammable gases

  17. Effect of finite Coulomb interaction on full counting statistics of electronic transport through single-molecule magnet

    Energy Technology Data Exchange (ETDEWEB)

    Xue Haibin, E-mail: xhb98326110@163.co [Institute of Theoretical Physics, Shanxi University, Taiyuan, Shanxi 030006 (China); Nie, Y.-H., E-mail: nieyh@sxu.edu.c [Institute of Theoretical Physics, Shanxi University, Taiyuan, Shanxi 030006 (China); Li, Z.-J.; Liang, J.-Q. [Institute of Theoretical Physics, Shanxi University, Taiyuan, Shanxi 030006 (China)

    2011-01-17

    We study the full counting statistics (FCS) in a single-molecule magnet (SMM) with finite Coulomb interaction U. For finite U the FCS, differing from U{yields}{infinity}, shows a symmetric gate-voltage-dependence when the coupling strengths with two electrodes are interchanged, which can be observed experimentally just by reversing the bias-voltage. Moreover, we find that the effect of finite U on shot noise depends on the internal level structure of the SMM and the coupling asymmetry of the SMM with two electrodes as well. When the coupling of the SMM with the incident-electrode is stronger than that with the outgoing-electrode, the super-Poissonian shot noise in the sequential tunneling regime appears under relatively small gate-voltage and relatively large finite U, and dose not for U{yields}{infinity}; while it occurs at relatively large gate-voltage for the opposite coupling case. The formation mechanism of super-Poissonian shot noise can be qualitatively attributed to the competition between fast and slow transport channels.

  18. Effect of finite Coulomb interaction on full counting statistics of electronic transport through single-molecule magnet

    International Nuclear Information System (INIS)

    Xue Haibin; Nie, Y.-H.; Li, Z.-J.; Liang, J.-Q.

    2011-01-01

    We study the full counting statistics (FCS) in a single-molecule magnet (SMM) with finite Coulomb interaction U. For finite U the FCS, differing from U→∞, shows a symmetric gate-voltage-dependence when the coupling strengths with two electrodes are interchanged, which can be observed experimentally just by reversing the bias-voltage. Moreover, we find that the effect of finite U on shot noise depends on the internal level structure of the SMM and the coupling asymmetry of the SMM with two electrodes as well. When the coupling of the SMM with the incident-electrode is stronger than that with the outgoing-electrode, the super-Poissonian shot noise in the sequential tunneling regime appears under relatively small gate-voltage and relatively large finite U, and dose not for U→∞; while it occurs at relatively large gate-voltage for the opposite coupling case. The formation mechanism of super-Poissonian shot noise can be qualitatively attributed to the competition between fast and slow transport channels.

  19. Low-Noise Free-Running High-Rate Photon-Counting for Space Communication and Ranging

    Science.gov (United States)

    Lu, Wei; Krainak, Michael A.; Yang, Guan; Sun, Xiaoli; Merritt, Scott

    2016-01-01

    We present performance data for low-noise free-running high-rate photon counting method for space optical communication and ranging. NASA GSFC is testing the performance of two types of novel photon-counting detectors 1) a 2x8 mercury cadmium telluride (HgCdTe) avalanche array made by DRS Inc., and a 2) a commercial 2880-element silicon avalanche photodiode (APD) array. We successfully measured real-time communication performance using both the 2 detected-photon threshold and logic AND-gate coincidence methods. Use of these methods allows mitigation of dark count, after-pulsing and background noise effects without using other method of Time Gating The HgCdTe APD array routinely demonstrated very high photon detection efficiencies (50) at near infrared wavelength. The commercial silicon APD array exhibited a fast output with rise times of 300 ps and pulse widths of 600 ps. On-chip individually filtered signals from the entire array were multiplexed onto a single fast output. NASA GSFC has tested both detectors for their potential application for space communications and ranging. We developed and compare their performances using both the 2 detected photon threshold and coincidence methods.

  20. Bulk tank somatic cell counts analyzed by statistical process control tools to identify and monitor subclinical mastitis incidence.

    Science.gov (United States)

    Lukas, J M; Hawkins, D M; Kinsel, M L; Reneau, J K

    2005-11-01

    The objective of this study was to examine the relationship between monthly Dairy Herd Improvement (DHI) subclinical mastitis and new infection rate estimates and daily bulk tank somatic cell count (SCC) summarized by statistical process control tools. Dairy Herd Improvement Association test-day subclinical mastitis and new infection rate estimates along with daily or every other day bulk tank SCC data were collected for 12 mo of 2003 from 275 Upper Midwest dairy herds. Herds were divided into 5 herd production categories. A linear score [LNS = ln(BTSCC/100,000)/0.693147 + 3] was calculated for each individual bulk tank SCC. For both the raw SCC and the transformed data, the mean and sigma were calculated using the statistical quality control individual measurement and moving range chart procedure of Statistical Analysis System. One hundred eighty-three herds of the 275 herds from the study data set were then randomly selected and the raw (method 1) and transformed (method 2) bulk tank SCC mean and sigma were used to develop models for predicting subclinical mastitis and new infection rate estimates. Herd production category was also included in all models as 5 dummy variables. Models were validated by calculating estimates of subclinical mastitis and new infection rates for the remaining 92 herds and plotting them against observed values of each of the dependents. Only herd production category and bulk tank SCC mean were significant and remained in the final models. High R2 values (0.83 and 0.81 for methods 1 and 2, respectively) indicated a strong correlation between the bulk tank SCC and herd's subclinical mastitis prevalence. The standard errors of the estimate were 4.02 and 4.28% for methods 1 and 2, respectively, and decreased with increasing herd production. As a case study, Shewhart Individual Measurement Charts were plotted from the bulk tank SCC to identify shifts in mastitis incidence. Four of 5 charts examined signaled a change in bulk tank SCC before

  1. Photon Counting Energy Dispersive Detector Arrays for X-ray Imaging.

    Science.gov (United States)

    Iwanczyk, Jan S; Nygård, Einar; Meirav, Oded; Arenson, Jerry; Barber, William C; Hartsough, Neal E; Malakhov, Nail; Wessel, Jan C

    2009-01-01

    The development of an innovative detector technology for photon-counting in X-ray imaging is reported. This new generation of detectors, based on pixellated cadmium telluride (CdTe) and cadmium zinc telluride (CZT) detector arrays electrically connected to application specific integrated circuits (ASICs) for readout, will produce fast and highly efficient photon-counting and energy-dispersive X-ray imaging. There are a number of applications that can greatly benefit from these novel imagers including mammography, planar radiography, and computed tomography (CT). Systems based on this new detector technology can provide compositional analysis of tissue through spectroscopic X-ray imaging, significantly improve overall image quality, and may significantly reduce X-ray dose to the patient. A very high X-ray flux is utilized in many of these applications. For example, CT scanners can produce ~100 Mphotons/mm(2)/s in the unattenuated beam. High flux is required in order to collect sufficient photon statistics in the measurement of the transmitted flux (attenuated beam) during the very short time frame of a CT scan. This high count rate combined with a need for high detection efficiency requires the development of detector structures that can provide a response signal much faster than the transit time of carriers over the whole detector thickness. We have developed CdTe and CZT detector array structures which are 3 mm thick with 16×16 pixels and a 1 mm pixel pitch. These structures, in the two different implementations presented here, utilize either a small pixel effect or a drift phenomenon. An energy resolution of 4.75% at 122 keV has been obtained with a 30 ns peaking time using discrete electronics and a (57)Co source. An output rate of 6×10(6) counts per second per individual pixel has been obtained with our ASIC readout electronics and a clinical CT X-ray tube. Additionally, the first clinical CT images, taken with several of our prototype photon-counting and

  2. Development of a high-count-rate neutron detector with position sensitivity and high efficiency

    International Nuclear Information System (INIS)

    Nelson, R.; Sandoval, J.

    1996-01-01

    While the neutron scattering community is bombarded with hints of new technologies that may deliver detectors with high-count-rate capability, high efficiency, gamma-ray insensitivity, and high resolution across large areas, only the time-tested, gas-filled 3 He and scintillation detectors are in widespread use. Future spallation sources with higher fluxes simply must exploit some of the advanced detector schemes that are as yet unproved as production systems. Technologies indicating promise as neutron detectors include pixel arrays of amorphous silicon, silicon microstrips, microstrips with gas, and new scintillation materials. This project sought to study the competing neutron detector technologies and determine which or what combination will lead to a production detector system well suited for use at a high-intensity neutron scattering source

  3. Low power ion spectrometer for high counting rates

    International Nuclear Information System (INIS)

    Klein, J.W.; Dullenkopf, P.; Glasmachers, A.; Melbert, J.; Winkelnkemper, W.

    1980-01-01

    This report describes in detail the electronic concept for a time-of-flight (TOF) ion spectrometer for high counting rates and high dynamic range which can be used as a satellite instrument. The detection principle of the spectrometer is based on a time-of-flight and energy measurement for each incident ion. The ionmass is related to these two quantities by a simple equation. The described approach for the mass identification systems is using an analog fast-slow concept: The fast TOF-signal preselects the gainstep in the much slower energy channel. The conversion time of the mass identifier is approximately 10 -6 s and the dynamic range of the energy channel is better than 10 3 (20 keV to 25 MeV). The purpose of this study was to demonstrate the feasibility of a TOF-spectrometer capable to measure the ion composition in planetary magnetospheres. (orig.) [de

  4. Assessment of the effect of probiotic curd consumption on salivary pH and streptococcus mutans counts.

    Science.gov (United States)

    Sudhir, R; Praveen, P; Anantharaj, A; Venkataraghavan, Karthik

    2012-07-01

    Antimicrobial methods of controlling dental caries that include probiotic agents can play a valuable role in establishing caries control in children at moderate to high risk for developing dental caries. Several studies have demonstrated the beneficial effects of use of various Probiotic products including curd. The objective of this study was to compare the effect of short-term consumption of probiotic curd containing Lactobacillus acidophilus and normal curd on salivary Streptococcus Mutans counts, as well as salivary pH. Forty, caries-free, 10-12 years old children were selected and randomly allocated to two groups. Test Group consisted of 20 children who consumed 200ml of probiotic curd daily for 30 days. Control Group consisted of 20 children who were given 200ml of regular curd for 30 days. Salivary pH and salivary Streptococcus Mutans counts were recorded at baseline and after 30 days and statistically compared using the Student's t-test. Consumption of probiotic curd resulted in a statistically significant reduction in S. Mutans colony counts (PpH (P>0.05) in both the groups. Short-term consumption of probiotic curds can reduce oral S. Mutans counts. However, this caused a slight reduction in salivary pH.

  5. An Adaptive Smoother for Counting Measurements

    International Nuclear Information System (INIS)

    Kondrasovs Vladimir; Coulon Romain; Normand Stephane

    2013-06-01

    Counting measurements associated with nuclear instruments are tricky to carry out due to the stochastic process of the radioactivity. Indeed events counting have to be processed and filtered in order to display a stable count rate value and to allow variations monitoring in the measured activity. Smoothers (as the moving average) are adjusted by a time constant defined as a compromise between stability and response time. A new approach has been developed and consists in improving the response time while maintaining count rate stability. It uses the combination of a smoother together with a detection filter. A memory of counting data is processed to calculate several count rate estimates using several integration times. These estimates are then sorted into the memory from short to long integration times. A measurement position, in terms of integration time, is then chosen into this memory after a detection test. An inhomogeneity into the Poisson counting process is detected by comparison between current position estimate and the other estimates contained into the memory in respect with the associated statistical variance calculated with homogeneous assumption. The measurement position (historical time) and the ability to forget an obsolete data or to keep in memory a useful data are managed using the detection test result. The proposed smoother is then an adaptive and a learning algorithm allowing an optimization of the response time while maintaining measurement counting stability and converging efficiently to the best counting estimate after an effective change in activity. This algorithm has also the specificity to be low recursive and thus easily embedded into DSP electronics based on FPGA or micro-controllers meeting 'real life' time requirements. (authors)

  6. Collecting operational event data for statistical analysis

    International Nuclear Information System (INIS)

    Atwood, C.L.

    1994-09-01

    This report gives guidance for collecting operational data to be used for statistical analysis, especially analysis of event counts. It discusses how to define the purpose of the study, the unit (system, component, etc.) to be studied, events to be counted, and demand or exposure time. Examples are given of classification systems for events in the data sources. A checklist summarizes the essential steps in data collection for statistical analysis

  7. Statistical Engine Knock Control

    DEFF Research Database (Denmark)

    Stotsky, Alexander A.

    2008-01-01

    A new statistical concept of the knock control of a spark ignition automotive engine is proposed . The control aim is associated with the statistical hy pothesis test which compares the threshold value to the average value of the max imal amplitud e of the knock sensor signal at a given freq uency....... C ontrol algorithm which is used for minimization of the regulation error realizes a simple count-up-count-d own logic. A new ad aptation algorithm for the knock d etection threshold is also d eveloped . C onfi d ence interval method is used as the b asis for ad aptation. A simple statistical mod el...... which includ es generation of the amplitud e signals, a threshold value d etermination and a knock sound mod el is d eveloped for evaluation of the control concept....

  8. Ancestry informative markers and complete blood count parameters in Brazilian blood donors

    Directory of Open Access Journals (Sweden)

    Gabriela E. S. Felix

    Full Text Available A complete blood count is very useful in clinical diagnoses when reference ranges are well established for the population. Complete blood counts and allele frequencies of Ancestry Informative Markers (AIMs were analyzed in Brazilians with the aim of characterizing the hematological values of an admixed population. Positive associations were observed between gender and neutrophils, monocytes, eosinophils, erythrocytes, hemoglobin, hematocrit, MCV, MCHC and platelet counts. No significant differences were found for age, alcohol consumption, educational status, ethnicity, smoking in respect to the complete blood count values. In general, men had higher red blood cell values, while women had higher values for white blood cells and platelets. The study of the population was highly heterogeneous with mean proportions (± SE of African, European and Amerindian ancestry being 49.0 ± 3.0%, 44.0 ± 9.0% and 7.0 ± 9.0%, respectively. Amerindian ancestry showed limited contribution to the makeup of the population, but estimated ancestral proportions were statistically significant (r = 0.9838; P<0.001. These hematologic values are similar to Afro-Americans, another admixed population.

  9. High energy behaviour of particles and unified statistics

    International Nuclear Information System (INIS)

    Chang, Y.

    1984-01-01

    Theories and experiments suggest that particles at high energy appear to possess a new statistics unifying Bose-Einstein and Fermi-Dirac statistics via the GAMMA distribution. This hypothesis can be obtained from many models, and agrees quantitatively with scaling, the multiplicty, large transverse momentum, the mass spectrum, and other data. It may be applied to scatterings at high energy, and agrees with experiments and known QED's results. The Veneziano model and other theories have implied new statistics, such as, the B distribution and the Polya distribution. They revert to the GAMMA distribution at high energy. The possible inapplicability of Pauli's exclusion principle within the unified statistics is considered and associated to the quark constituents

  10. Statistical learning in high energy and astrophysics

    International Nuclear Information System (INIS)

    Zimmermann, J.

    2005-01-01

    This thesis studies the performance of statistical learning methods in high energy and astrophysics where they have become a standard tool in physics analysis. They are used to perform complex classification or regression by intelligent pattern recognition. This kind of artificial intelligence is achieved by the principle ''learning from examples'': The examples describe the relationship between detector events and their classification. The application of statistical learning methods is either motivated by the lack of knowledge about this relationship or by tight time restrictions. In the first case learning from examples is the only possibility since no theory is available which would allow to build an algorithm in the classical way. In the second case a classical algorithm exists but is too slow to cope with the time restrictions. It is therefore replaced by a pattern recognition machine which implements a fast statistical learning method. But even in applications where some kind of classical algorithm had done a good job, statistical learning methods convinced by their remarkable performance. This thesis gives an introduction to statistical learning methods and how they are applied correctly in physics analysis. Their flexibility and high performance will be discussed by showing intriguing results from high energy and astrophysics. These include the development of highly efficient triggers, powerful purification of event samples and exact reconstruction of hidden event parameters. The presented studies also show typical problems in the application of statistical learning methods. They should be only second choice in all cases where an algorithm based on prior knowledge exists. Some examples in physics analyses are found where these methods are not used in the right way leading either to wrong predictions or bad performance. Physicists also often hesitate to profit from these methods because they fear that statistical learning methods cannot be controlled in a

  11. Statistical learning in high energy and astrophysics

    Energy Technology Data Exchange (ETDEWEB)

    Zimmermann, J.

    2005-06-16

    This thesis studies the performance of statistical learning methods in high energy and astrophysics where they have become a standard tool in physics analysis. They are used to perform complex classification or regression by intelligent pattern recognition. This kind of artificial intelligence is achieved by the principle ''learning from examples'': The examples describe the relationship between detector events and their classification. The application of statistical learning methods is either motivated by the lack of knowledge about this relationship or by tight time restrictions. In the first case learning from examples is the only possibility since no theory is available which would allow to build an algorithm in the classical way. In the second case a classical algorithm exists but is too slow to cope with the time restrictions. It is therefore replaced by a pattern recognition machine which implements a fast statistical learning method. But even in applications where some kind of classical algorithm had done a good job, statistical learning methods convinced by their remarkable performance. This thesis gives an introduction to statistical learning methods and how they are applied correctly in physics analysis. Their flexibility and high performance will be discussed by showing intriguing results from high energy and astrophysics. These include the development of highly efficient triggers, powerful purification of event samples and exact reconstruction of hidden event parameters. The presented studies also show typical problems in the application of statistical learning methods. They should be only second choice in all cases where an algorithm based on prior knowledge exists. Some examples in physics analyses are found where these methods are not used in the right way leading either to wrong predictions or bad performance. Physicists also often hesitate to profit from these methods because they fear that statistical learning methods cannot

  12. Is total lymphocyte count related to nutritional markers in hospitalized older adults?

    Directory of Open Access Journals (Sweden)

    Vânia Aparecida LEANDRO-MERHI

    Full Text Available ABSTRACT BACKGROUND Older patients are commonly malnourished during hospital stay, and a high prevalence of malnutrition is found in hospitalized patients aged more than 65 years. OBJECTIVE To investigate whether total lymphocyte count is related to other nutritional markers in hospitalized older adults. METHODS Hospitalized older adults (N=131 were recruited for a cross-sectional study. Their nutritional status was assessed by the Nutritional Risk Screening (NRS, anthropometry, and total lymphocyte count. The statistical analyses included the chi-square test, Fisher's exact test, and Mann-Whitney test. Spearman's linear correlation coefficient determined whether total lymphocyte count was correlated with the nutritional markers. Multiple linear regression determined the parameters associated with lymphocyte count. The significance level was set at 5%. RESULTS According to the NRS, 41.2% of the patients were at nutritional risk, and 36% had mild or moderate depletion according to total lymphocyte count. Total lymphocyte count was weakly correlated with mid-upper arm circumference (r=0.20507; triceps skinfold thickness (r=0.29036, and length of hospital stay (r= -0.21518. Total lymphocyte count in different NRS categories differed significantly: older adults who were not at nutritional risk had higher mean and median total lymphocyte count ( P =0.0245. Multiple regression analysis showed that higher lymphocyte counts were associated with higher triceps skinfold thicknesses and no nutritional risk according to the NRS. CONCLUSION Total lymphocyte count was correlated with mid-upper arm circumference, triceps skinfold thickness, and nutritional risk according to the NRS. In multiple regression the combined factors that remained associated with lymphocyte count were NRS and triceps skinfold thickness. Therefore, total lymphocyte count may be considered a nutritional marker. Other studies should confirm these findings.

  13. Time-over-threshold readout to enhance the high flux capabilities of single-photon-counting detectors

    International Nuclear Information System (INIS)

    Bergamaschi, Anna; Dinapoli, Roberto; Greiffenberg, Dominic; Henrich, Beat; Johnson, Ian; Mozzanica, Aldo; Radicci, Valeria; Schmitt, Bernd; Shi, Xintian; Stoppani, Laura

    2011-01-01

    The MYTHEN photon-counting ASIC operated in time-over-threshold mode shows an innovative approach towards the development of a detector operating with very high photon intensities while maintaining the single-photon sensitivity for synchrotron radiation experiments. The MYTHEN single-photon-counting (SPC) detector has been characterized using the time-over-threshold (ToT) readout method, i.e. measuring the time that the signal produced by the detected X-rays remains above the comparator threshold. In the following it is shown that the ToT readout preserves the sensitivity, dynamic range and capability of background suppression of the SPC mode, while enhancing the count-rate capability, which is the main limitation of state-of-the-art SPC systems

  14. Time-over-threshold readout to enhance the high flux capabilities of single-photon-counting detectors

    Energy Technology Data Exchange (ETDEWEB)

    Bergamaschi, Anna, E-mail: anna.bergamaschi@psi.ch; Dinapoli, Roberto; Greiffenberg, Dominic; Henrich, Beat; Johnson, Ian; Mozzanica, Aldo; Radicci, Valeria; Schmitt, Bernd; Shi, Xintian; Stoppani, Laura [Paul Scherrer Institut, CH-5232 Villigen (Switzerland)

    2011-11-01

    The MYTHEN photon-counting ASIC operated in time-over-threshold mode shows an innovative approach towards the development of a detector operating with very high photon intensities while maintaining the single-photon sensitivity for synchrotron radiation experiments. The MYTHEN single-photon-counting (SPC) detector has been characterized using the time-over-threshold (ToT) readout method, i.e. measuring the time that the signal produced by the detected X-rays remains above the comparator threshold. In the following it is shown that the ToT readout preserves the sensitivity, dynamic range and capability of background suppression of the SPC mode, while enhancing the count-rate capability, which is the main limitation of state-of-the-art SPC systems.

  15. Short communication: Repeatability of differential goat bulk milk culture and associations with somatic cell count, total bacterial count, and standard plate count.

    Science.gov (United States)

    Koop, G; Dik, N; Nielen, M; Lipman, L J A

    2010-06-01

    The aims of this study were to assess how different bacterial groups in bulk milk are related to bulk milk somatic cell count (SCC), bulk milk total bacterial count (TBC), and bulk milk standard plate count (SPC) and to measure the repeatability of bulk milk culturing. On 53 Dutch dairy goat farms, 3 bulk milk samples were collected at intervals of 2 wk. The samples were cultured for SPC, coliform count, and staphylococcal count and for the presence of Staphylococcus aureus. Furthermore, SCC (Fossomatic 5000, Foss, Hillerød, Denmark) and TBC (BactoScan FC 150, Foss) were measured. Staphylococcal count was correlated to SCC (r=0.40), TBC (r=0.51), and SPC (r=0.53). Coliform count was correlated to TBC (r=0.33), but not to any of the other variables. Staphylococcus aureus did not correlate to SCC. The contribution of the staphylococcal count to the SPC was 31%, whereas the coliform count comprised only 1% of the SPC. The agreement of the repeated measurements was low. This study indicates that staphylococci in goat bulk milk are related to SCC and make a significant contribution to SPC. Because of the high variation in bacterial counts, repeated sampling is necessary to draw valid conclusions from bulk milk culturing. 2010 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.

  16. Introduction to high-dimensional statistics

    CERN Document Server

    Giraud, Christophe

    2015-01-01

    Ever-greater computing technologies have given rise to an exponentially growing volume of data. Today massive data sets (with potentially thousands of variables) play an important role in almost every branch of modern human activity, including networks, finance, and genetics. However, analyzing such data has presented a challenge for statisticians and data analysts and has required the development of new statistical methods capable of separating the signal from the noise.Introduction to High-Dimensional Statistics is a concise guide to state-of-the-art models, techniques, and approaches for ha

  17. Pixel-Cluster Counting Luminosity Measurement in ATLAS

    CERN Document Server

    McCormack, William Patrick; The ATLAS collaboration

    2016-01-01

    A precision measurement of the delivered luminosity is a key component of the ATLAS physics program at the Large Hadron Collider (LHC). A fundamental ingredient of the strategy to control the systematic uncertainties affecting the absolute luminosity has been to compare the measurements of several luminometers, most of which use more than one counting technique. The level of consistency across the various methods provides valuable cross-checks as well as an estimate of the detector-related systematic uncertainties. This poster describes the development of a luminosity algorithm based on pixel-cluster counting in the recently installed ATLAS inner b-layer (IBL), using data recorded during the 2015 pp run at the LHC. The noise and background contamination of the luminosity-associated cluster count is minimized by a multi-component fit to the measured cluster-size distribution in the forward pixel modules of the IBL. The linearity, long-term stability and statistical precision of the cluster-counting method are ...

  18. Pixel-Cluster Counting Luminosity Measurement In ATLAS

    CERN Document Server

    AUTHOR|(SzGeCERN)782710; The ATLAS collaboration

    2017-01-01

    A precision measurement of the delivered luminosity is a key component of the ATLAS physics program at the Large Hadron Collider (LHC). A fundamental ingredient of the strategy to control the systematic uncertainties affecting the absolute luminosity has been to compare the measure- ments of several luminometers, most of which use more than one counting technique. The level of consistency across the various methods provides valuable cross-checks as well as an estimate of the detector-related systematic uncertainties. This poster describes the development of a luminosity algorithm based on pixel-cluster counting in the recently installed ATLAS inner b-layer (IBL), using data recorded during the 2015 pp run at the LHC. The noise and background contamination of the luminosity-associated cluster count is minimized by a multi-component fit to the measured cluster-size distribution in the forward pixel modules of the IBL. The linearity, long-term stability and statistical precision of the cluster- counting method a...

  19. High energy X-ray photon counting imaging using linear accelerator and silicon strip detectors

    International Nuclear Information System (INIS)

    Tian, Y.; Shimazoe, K.; Yan, X.; Ueda, O.; Ishikura, T.; Fujiwara, T.; Uesaka, M.; Ohno, M.; Tomita, H.; Yoshihara, Y.; Takahashi, H.

    2016-01-01

    A photon counting imaging detector system for high energy X-rays is developed for on-site non-destructive testing of thick objects. One-dimensional silicon strip (1 mm pitch) detectors are stacked to form a two-dimensional edge-on module. Each detector is connected to a 48-channel application specific integrated circuit (ASIC). The threshold-triggered events are recorded by a field programmable gate array based counter in each channel. The detector prototype is tested using 950 kV linear accelerator X-rays. The fast CR shaper (300 ns pulse width) of the ASIC makes it possible to deal with the high instant count rate during the 2 μs beam pulse. The preliminary imaging results of several metal and concrete samples are demonstrated.

  20. High energy X-ray photon counting imaging using linear accelerator and silicon strip detectors

    Energy Technology Data Exchange (ETDEWEB)

    Tian, Y., E-mail: cycjty@sophie.q.t.u-tokyo.ac.jp [Department of Bioengineering, the University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656 (Japan); Shimazoe, K.; Yan, X. [Department of Nuclear Engineering and Management, the University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656 (Japan); Ueda, O.; Ishikura, T. [Fuji Electric Co., Ltd., Fuji, Hino, Tokyo 191-8502 (Japan); Fujiwara, T. [National Institute of Advanced Industrial Science and Technology, 1-1-1 Umezono, Tsukuba, Ibaraki 305-8568 (Japan); Uesaka, M.; Ohno, M. [Nuclear Professional School, the University of Tokyo, 2-22 Shirakata-shirane, Tokai, Ibaraki 319-1188 (Japan); Tomita, H. [Department of Quantum Engineering, Nagoya University, Furo, Chikusa, Nagoya 464-8603 (Japan); Yoshihara, Y. [Department of Nuclear Engineering and Management, the University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656 (Japan); Takahashi, H. [Department of Bioengineering, the University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656 (Japan); Department of Nuclear Engineering and Management, the University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656 (Japan)

    2016-09-11

    A photon counting imaging detector system for high energy X-rays is developed for on-site non-destructive testing of thick objects. One-dimensional silicon strip (1 mm pitch) detectors are stacked to form a two-dimensional edge-on module. Each detector is connected to a 48-channel application specific integrated circuit (ASIC). The threshold-triggered events are recorded by a field programmable gate array based counter in each channel. The detector prototype is tested using 950 kV linear accelerator X-rays. The fast CR shaper (300 ns pulse width) of the ASIC makes it possible to deal with the high instant count rate during the 2 μs beam pulse. The preliminary imaging results of several metal and concrete samples are demonstrated.

  1. High impact  =  high statistical standards? Not necessarily so.

    Science.gov (United States)

    Tressoldi, Patrizio E; Giofré, David; Sella, Francesco; Cumming, Geoff

    2013-01-01

    What are the statistical practices of articles published in journals with a high impact factor? Are there differences compared with articles published in journals with a somewhat lower impact factor that have adopted editorial policies to reduce the impact of limitations of Null Hypothesis Significance Testing? To investigate these questions, the current study analyzed all articles related to psychological, neuropsychological and medical issues, published in 2011 in four journals with high impact factors: Science, Nature, The New England Journal of Medicine and The Lancet, and three journals with relatively lower impact factors: Neuropsychology, Journal of Experimental Psychology-Applied and the American Journal of Public Health. Results show that Null Hypothesis Significance Testing without any use of confidence intervals, effect size, prospective power and model estimation, is the prevalent statistical practice used in articles published in Nature, 89%, followed by articles published in Science, 42%. By contrast, in all other journals, both with high and lower impact factors, most articles report confidence intervals and/or effect size measures. We interpreted these differences as consequences of the editorial policies adopted by the journal editors, which are probably the most effective means to improve the statistical practices in journals with high or low impact factors.

  2. High Impact = High Statistical Standards? Not Necessarily So

    Science.gov (United States)

    Tressoldi, Patrizio E.; Giofré, David; Sella, Francesco; Cumming, Geoff

    2013-01-01

    What are the statistical practices of articles published in journals with a high impact factor? Are there differences compared with articles published in journals with a somewhat lower impact factor that have adopted editorial policies to reduce the impact of limitations of Null Hypothesis Significance Testing? To investigate these questions, the current study analyzed all articles related to psychological, neuropsychological and medical issues, published in 2011 in four journals with high impact factors: Science, Nature, The New England Journal of Medicine and The Lancet, and three journals with relatively lower impact factors: Neuropsychology, Journal of Experimental Psychology-Applied and the American Journal of Public Health. Results show that Null Hypothesis Significance Testing without any use of confidence intervals, effect size, prospective power and model estimation, is the prevalent statistical practice used in articles published in Nature, 89%, followed by articles published in Science, 42%. By contrast, in all other journals, both with high and lower impact factors, most articles report confidence intervals and/or effect size measures. We interpreted these differences as consequences of the editorial policies adopted by the journal editors, which are probably the most effective means to improve the statistical practices in journals with high or low impact factors. PMID:23418533

  3. Comment on: 'A Poisson resampling method for simulating reduced counts in nuclear medicine images'

    DEFF Research Database (Denmark)

    de Nijs, Robin

    2015-01-01

    In order to be able to calculate half-count images from already acquired data, White and Lawson published their method based on Poisson resampling. They verified their method experimentally by measurements with a Co-57 flood source. In this comment their results are reproduced and confirmed...... by a direct numerical simulation in Matlab. Not only Poisson resampling, but also two direct redrawing methods were investigated. Redrawing methods were based on a Poisson and a Gaussian distribution. Mean, standard deviation, skewness and excess kurtosis half-count/full-count ratios were determined for all...... methods, and compared to the theoretical values for a Poisson distribution. Statistical parameters showed the same behavior as in the original note and showed the superiority of the Poisson resampling method. Rounding off before saving of the half count image had a severe impact on counting statistics...

  4. Bias in iterative reconstruction of low-statistics PET data: benefits of a resolution model

    Energy Technology Data Exchange (ETDEWEB)

    Walker, M D; Asselin, M-C; Julyan, P J; Feldmann, M; Matthews, J C [School of Cancer and Enabling Sciences, Wolfson Molecular Imaging Centre, MAHSC, University of Manchester, Manchester M20 3LJ (United Kingdom); Talbot, P S [Mental Health and Neurodegeneration Research Group, Wolfson Molecular Imaging Centre, MAHSC, University of Manchester, Manchester M20 3LJ (United Kingdom); Jones, T, E-mail: matthew.walker@manchester.ac.uk [Academic Department of Radiation Oncology, Christie Hospital, University of Manchester, Manchester M20 4BX (United Kingdom)

    2011-02-21

    Iterative image reconstruction methods such as ordered-subset expectation maximization (OSEM) are widely used in PET. Reconstructions via OSEM are however reported to be biased for low-count data. We investigated this and considered the impact for dynamic PET. Patient listmode data were acquired in [{sup 11}C]DASB and [{sup 15}O]H{sub 2}O scans on the HRRT brain PET scanner. These data were subsampled to create many independent, low-count replicates. The data were reconstructed and the images from low-count data were compared to the high-count originals (from the same reconstruction method). This comparison enabled low-statistics bias to be calculated for the given reconstruction, as a function of the noise-equivalent counts (NEC). Two iterative reconstruction methods were tested, one with and one without an image-based resolution model (RM). Significant bias was observed when reconstructing data of low statistical quality, for both subsampled human and simulated data. For human data, this bias was substantially reduced by including a RM. For [{sup 11}C]DASB the low-statistics bias in the caudate head at 1.7 M NEC (approx. 30 s) was -5.5% and -13% with and without RM, respectively. We predicted biases in the binding potential of -4% and -10%. For quantification of cerebral blood flow for the whole-brain grey- or white-matter, using [{sup 15}O]H{sub 2}O and the PET autoradiographic method, a low-statistics bias of <2.5% and <4% was predicted for reconstruction with and without the RM. The use of a resolution model reduces low-statistics bias and can hence be beneficial for quantitative dynamic PET.

  5. Problems and precision of the alpha scintillation radon counting system

    International Nuclear Information System (INIS)

    Lucas, H.F.; Markuu, F.

    1985-01-01

    Variations in efficiency as large as 3% have been found for radon scintillation counting systems in which the photomultiplier tubes are sensitive to the thermoluminescent photons emitted by the scintillator after exposure to light or for which the resolution has deteriorated. The additional standard deviation caused by counting a radon chamber on multiple counting systems has been evaluated and the effect, if present, did not exceed about 0.1%. The chambers have been calibrated for the measurement of radon in air, and the standard deviation was equal to statistical counting error combined with a systematic error of 1.1%. 3 references, 2 figures, 2 tables

  6. Unaccounted Workload Factor: Game-Day Pitch Counts in High School Baseball Pitchers—An Observational Study

    Science.gov (United States)

    Zaremski, Jason L.; Zeppieri, Giorgio; Jones, Deborah L.; Tripp, Brady L.; Bruner, Michelle; Vincent, Heather K.; Horodyski, MaryBeth

    2018-01-01

    Background: Throwing injuries are common in high school baseball. Known risk factors include excessive pitch counts, year-round pitching, and pitching with arm pain and fatigue. Despite the evidence, the prevalence of pitching injuries among high school players has not decreased. One possibility to explain this pattern is that players accumulate unaccounted pitch volume during warm-up and bullpen activity, but this has not yet been examined. Hypotheses: Our primary hypothesis was that approximately 30% to 40% of pitches thrown off a mound by high school pitchers during a game-day outing are unaccounted for in current data but will be revealed when bullpen sessions and warm-up pitches are included. Our secondary hypothesis was that there is wide variability among players in the number of bullpen pitches thrown per outing. Study Design: Cross-sectional study; Level of evidence, 3. Methods: Researchers counted all pitches thrown off a mound during varsity high school baseball games played by 34 high schools in North Central Florida during the 2017 season. Results: We recorded 13,769 total pitches during 115 varsity high school baseball starting pitcher outings. The mean ± SD pitch numbers per game were calculated for bullpen activity (27.2 ± 9.4), warm-up (23.6 ±8.0), live games (68.9 ±19.7), and total pitches per game (119.7 ± 27.8). Thus, 42.4% of the pitches performed were not accounted for in the pitch count monitoring of these players. The number of bullpen pitches thrown varied widely among players, with 25% of participants in our data set throwing fewer than 22 pitches and 25% throwing more than 33 pitches per outing. Conclusion: In high school baseball players, pitch count monitoring does not account for the substantial volume of pitching that occurs during warm-up and bullpen activity during the playing season. These extra pitches should be closely monitored to help mitigate the risk of overuse injury. PMID:29662911

  7. Unaccounted Workload Factor: Game-Day Pitch Counts in High School Baseball Pitchers-An Observational Study.

    Science.gov (United States)

    Zaremski, Jason L; Zeppieri, Giorgio; Jones, Deborah L; Tripp, Brady L; Bruner, Michelle; Vincent, Heather K; Horodyski, MaryBeth

    2018-04-01

    Throwing injuries are common in high school baseball. Known risk factors include excessive pitch counts, year-round pitching, and pitching with arm pain and fatigue. Despite the evidence, the prevalence of pitching injuries among high school players has not decreased. One possibility to explain this pattern is that players accumulate unaccounted pitch volume during warm-up and bullpen activity, but this has not yet been examined. Our primary hypothesis was that approximately 30% to 40% of pitches thrown off a mound by high school pitchers during a game-day outing are unaccounted for in current data but will be revealed when bullpen sessions and warm-up pitches are included. Our secondary hypothesis was that there is wide variability among players in the number of bullpen pitches thrown per outing. Cross-sectional study; Level of evidence, 3. Researchers counted all pitches thrown off a mound during varsity high school baseball games played by 34 high schools in North Central Florida during the 2017 season. We recorded 13,769 total pitches during 115 varsity high school baseball starting pitcher outings. The mean ± SD pitch numbers per game were calculated for bullpen activity (27.2 ± 9.4), warm-up (23.6 ±8.0), live games (68.9 ±19.7), and total pitches per game (119.7 ± 27.8). Thus, 42.4% of the pitches performed were not accounted for in the pitch count monitoring of these players. The number of bullpen pitches thrown varied widely among players, with 25% of participants in our data set throwing fewer than 22 pitches and 25% throwing more than 33 pitches per outing. In high school baseball players, pitch count monitoring does not account for the substantial volume of pitching that occurs during warm-up and bullpen activity during the playing season. These extra pitches should be closely monitored to help mitigate the risk of overuse injury.

  8. The distribution of controlled drugs on banknotes via counting machines.

    Science.gov (United States)

    Carter, James F; Sleeman, Richard; Parry, Joanna

    2003-03-27

    Bundles of paper, similar to sterling banknotes, were counted in banks in England and Wales. Subsequent analysis showed that the counting process, both by machine and by hand, transferred nanogram amounts of cocaine to the paper. Crystalline material, similar to cocaine hydrochloride, could be observed on the surface of the paper following counting. The geographical distribution of contamination broadly followed Government statistics for cocaine usage within the UK. Diacetylmorphine, Delta(9)-tetrahydrocannabinol (THC) and 3,4-methylenedioxymethylamphetamine (MDMA) were not detected during this study.

  9. Generalized interpolative quantum statistics

    International Nuclear Information System (INIS)

    Ramanathan, R.

    1992-01-01

    A generalized interpolative quantum statistics is presented by conjecturing a certain reordering of phase space due to the presence of possible exotic objects other than bosons and fermions. Such an interpolation achieved through a Bose-counting strategy predicts the existence of an infinite quantum Boltzmann-Gibbs statistics akin to the one discovered by Greenberg recently

  10. ChromAIX2: A large area, high count-rate energy-resolving photon counting ASIC for a Spectral CT Prototype

    Science.gov (United States)

    Steadman, Roger; Herrmann, Christoph; Livne, Amir

    2017-08-01

    Spectral CT based on energy-resolving photon counting detectors is expected to deliver additional diagnostic value at a lower dose than current state-of-the-art CT [1]. The capability of simultaneously providing a number of spectrally distinct measurements not only allows distinguishing between photo-electric and Compton interactions but also discriminating contrast agents that exhibit a K-edge discontinuity in the absorption spectrum, referred to as K-edge Imaging [2]. Such detectors are based on direct converting sensors (e.g. CdTe or CdZnTe) and high-rate photon counting electronics. To support the development of Spectral CT and show the feasibility of obtaining rates exceeding 10 Mcps/pixel (Poissonian observed count-rate), the ChromAIX ASIC has been previously reported showing 13.5 Mcps/pixel (150 Mcps/mm2 incident) [3]. The ChromAIX has been improved to offer the possibility of a large area coverage detector, and increased overall performance. The new ASIC is called ChromAIX2, and delivers count-rates exceeding 15 Mcps/pixel with an rms-noise performance of approximately 260 e-. It has an isotropic pixel pitch of 500 μm in an array of 22×32 pixels and is tile-able on three of its sides. The pixel topology consists of a two stage amplifier (CSA and Shaper) and a number of test features allowing to thoroughly characterize the ASIC without a sensor. A total of 5 independent thresholds are also available within each pixel, allowing to acquire 5 spectrally distinct measurements simultaneously. The ASIC also incorporates a baseline restorer to eliminate excess currents induced by the sensor (e.g. dark current and low frequency drifts) which would otherwise cause an energy estimation error. In this paper we report on the inherent electrical performance of the ChromAXI2 as well as measurements obtained with CZT (CdZnTe)/CdTe sensors and X-rays and radioactive sources.

  11. Use of a mixture statistical model in studying malaria vectors density.

    Directory of Open Access Journals (Sweden)

    Olayidé Boussari

    Full Text Available Vector control is a major step in the process of malaria control and elimination. This requires vector counts and appropriate statistical analyses of these counts. However, vector counts are often overdispersed. A non-parametric mixture of Poisson model (NPMP is proposed to allow for overdispersion and better describe vector distribution. Mosquito collections using the Human Landing Catches as well as collection of environmental and climatic data were carried out from January to December 2009 in 28 villages in Southern Benin. A NPMP regression model with "village" as random effect is used to test statistical correlations between malaria vectors density and environmental and climatic factors. Furthermore, the villages were ranked using the latent classes derived from the NPMP model. Based on this classification of the villages, the impacts of four vector control strategies implemented in the villages were compared. Vector counts were highly variable and overdispersed with important proportion of zeros (75%. The NPMP model had a good aptitude to predict the observed values and showed that: i proximity to freshwater body, market gardening, and high levels of rain were associated with high vector density; ii water conveyance, cattle breeding, vegetation index were associated with low vector density. The 28 villages could then be ranked according to the mean vector number as estimated by the random part of the model after adjustment on all covariates. The NPMP model made it possible to describe the distribution of the vector across the study area. The villages were ranked according to the mean vector density after taking into account the most important covariates. This study demonstrates the necessity and possibility of adapting methods of vector counting and sampling to each setting.

  12. Alabama Kids Count 2002 Data Book.

    Science.gov (United States)

    Curtis, Apreill; Bogie, Don

    This Kids Count data book examines statewide trends in well-being of Alabamas children. The statistical portrait is based on 18 indicators in the areas of child health, education, safety, and security: (1) infant mortality rate; (2) low weight births; (3) child health index; (4) births to unmarried teens; (5) first grade retention; (6) school…

  13. Evaluation of Deterministic and Stochastic Components of Traffic Counts

    Directory of Open Access Journals (Sweden)

    Ivan Bošnjak

    2012-10-01

    Full Text Available Traffic counts or statistical evidence of the traffic processare often a characteristic of time-series data. In this paper fundamentalproblem of estimating deterministic and stochasticcomponents of a traffic process are considered, in the context of"generalised traffic modelling". Different methods for identificationand/or elimination of the trend and seasonal componentsare applied for concrete traffic counts. Further investigationsand applications of ARIMA models, Hilbert space formulationsand state-space representations are suggested.

  14. Playing at Statistical Mechanics

    Science.gov (United States)

    Clark, Paul M.; And Others

    1974-01-01

    Discussed are the applications of counting techniques of a sorting game to distributions and concepts in statistical mechanics. Included are the following distributions: Fermi-Dirac, Bose-Einstein, and most probable. (RH)

  15. Is it safe to use Poisson statistics in nuclear spectrometry?

    International Nuclear Information System (INIS)

    Pomme, S.; Robouch, P.; Arana, G.; Eguskiza, M.; Maguregui, M.I.

    2000-01-01

    The boundary conditions in which Poisson statistics can be applied in nuclear spectrometry are investigated. Improved formulas for the uncertainty of nuclear counting with deadtime and pulse pileup are presented. A comparison is made between the expected statistical uncertainty for loss-free counting, fixed live-time and fixed real-time measurements. (author)

  16. Disseminated HIV-Associated Kaposi’s Sarcoma With High CD4 Cell Count And Low Viral Load

    Directory of Open Access Journals (Sweden)

    Diana Pereira Anjos

    2017-12-01

    Full Text Available Kaposi’s sarcoma is considered an acquired immunodeficiency syndrome-defining illness and is caused by human herpesvirus 8. It has been associated with patients infected with human immunodeficiency virus (HIV who have CD4 T lymphocytes <200 cells/uL and high viral loads. We report a case of a 23-year old woman infected with HIV-1 and receiving antiretroviral treatment since diagnosis, with high CD4 cell count and low viral load that presented with disseminated Kaposi’s sarcoma. Clinicians should be aware of the occurrence of Kaposi’s sarcoma despite robust CD4 cell counts.

  17. Identification of cotton properties to improve yarn count quality by using regression analysis

    International Nuclear Information System (INIS)

    Amin, M.; Ullah, M.; Akbar, A.

    2014-01-01

    Identification of raw material characteristics towards yarn count variation was studied by using statistical techniques. Regression analysis is used to meet the objective. Stepwise regression is used for mode) selection, and coefficient of determination and mean squared error (MSE) criteria are used to identify the contributing factors of cotton properties for yam count. Statistical assumptions of normality, autocorrelation and multicollinearity are evaluated by using probability plot, Durbin Watson test, variance inflation factor (VIF), and then model fitting is carried out. It is found that, invisible (INV), nepness (Nep), grayness (RD), cotton trash (TR) and uniformity index (VI) are the main contributing cotton properties for yarn count variation. The results are also verified by Pareto chart. (author)

  18. High sensitivity neutron activation analysis using coincidence counting method

    International Nuclear Information System (INIS)

    Suzuki, Shogo; Okada, Yukiko; Hirai, Shoji

    1999-01-01

    Four kinds of standard samples such as river sediment (NIES CRM No.16), Typical Japanese Diet, otoliths and river water were irradiated by TRIGA-II (100 kW, 3.7x10 12 n cm -2 s -1 ) for 6 h. After irradiation and cooling, they were analyzed by the coincidence counting method and a conventional γ-ray spectrometry. Se, Ba and Hf were determined by 75 Se 265 keV, 131 Ba 496 keV and 181 Hf 482 keV. On the river sediment sample, Ba and Hf showed the same values by two methods, but Se value contained Ta by the conventional method, although the coincidence counting method could analyze Se. On Typical Japanese Diet and otoliths, Se could be determined by two methods and Ba and Hf determined by the coincidence counting method but not determined by the conventional method. Se value in the river water agreed with the authorization value. (S.Y.)

  19. Alabama Kids Count 2001 Data Book.

    Science.gov (United States)

    Curtis, Apreill; Bogie, Don

    This Kids Count data book examines statewide trends in well-being for Alabama's children. The statistical portrait is based on 17 indicators in the areas of health, education, safety, and security. The indicators are: (1) infant mortality rate; (2) low weight births; (3) child health index; (4) births to unmarried teens; (5) first grade retention;…

  20. Nevada Kids Count Data Book, 1997.

    Science.gov (United States)

    We Can, Inc., Las Vegas, NV.

    This Kids Count data book is the first to examine statewide indicators of the well being of Nevada's children. The statistical portrait is based on 15 indicators of child well being: (1) percent low birth-weight babies; (2) infant mortality rate; (3) percent of children in poverty; (4) percent of children in single-parent families; (5) percent of…

  1. Determination of Np, Pu and Am in high level radioactive waste with extraction-liquid scintillation counting

    International Nuclear Information System (INIS)

    Yang Dazhu; Zhu Yongjun; Jiao Rongzhou

    1994-01-01

    A new method for the determination of transuranium elements, Np, Pu and Am with extraction-liquid scintillation counting has been studied systematically. Procedures for the separation of Pu and Am by HDEHP-TRPO extraction and for the separation of Np by TTA-TiOA extraction have been developed, by which the recovery of Np, Pu and Am is 97%, 99% and 99%, respectively, and the decontamination factors for the major fission products ( 90 Sr, 137 Cs etc.) are 10 4 -10 6 . Pulse shape discrimination (PSD) technique has been introduced to liquid scintillation counting, by which the counting efficiency of α-activity is >99% and the rejection of β-counts is >99.95%. This new method, combining extraction and pulse shape discrimination with liquid scintillation technique, has been successfully applied to the assay of Np, Pu and Am in high level radioactive waste. (author) 7 refs.; 7 figs.; 4 tabs

  2. The Kruskal Count

    OpenAIRE

    Lagarias, Jeffrey C.; Rains, Eric; Vanderbei, Robert J.

    2001-01-01

    The Kruskal Count is a card trick invented by Martin J. Kruskal in which a magician "guesses" a card selected by a subject according to a certain counting procedure. With high probability the magician can correctly "guess" the card. The success of the trick is based on a mathematical principle related to coupling methods for Markov chains. This paper analyzes in detail two simplified variants of the trick and estimates the probability of success. The model predictions are compared with simula...

  3. Evaluation of bias and variance in low-count OSEM list mode reconstruction

    International Nuclear Information System (INIS)

    Jian, Y; Carson, R E; Planeta, B

    2015-01-01

    Statistical algorithms have been widely used in PET image reconstruction. The maximum likelihood expectation maximization reconstruction has been shown to produce bias in applications where images are reconstructed from a relatively small number of counts. In this study, image bias and variability in low-count OSEM reconstruction are investigated on images reconstructed with MOLAR (motion-compensation OSEM list-mode algorithm for resolution-recovery reconstruction) platform. A human brain ([ 11 C]AFM) and a NEMA phantom are used in the simulation and real experiments respectively, for the HRRT and Biograph mCT. Image reconstructions were repeated with different combinations of subsets and iterations. Regions of interest were defined on low-activity and high-activity regions to evaluate the bias and noise at matched effective iteration numbers (iterations × subsets). Minimal negative biases and no positive biases were found at moderate count levels and less than 5% negative bias was found using extremely low levels of counts (0.2 M NEC). At any given count level, other factors, such as subset numbers and frame-based scatter correction may introduce small biases (1–5%) in the reconstructed images. The observed bias was substantially lower than that reported in the literature, perhaps due to the use of point spread function and/or other implementation methods in MOLAR. (paper)

  4. Multivariate statistics high-dimensional and large-sample approximations

    CERN Document Server

    Fujikoshi, Yasunori; Shimizu, Ryoichi

    2010-01-01

    A comprehensive examination of high-dimensional analysis of multivariate methods and their real-world applications Multivariate Statistics: High-Dimensional and Large-Sample Approximations is the first book of its kind to explore how classical multivariate methods can be revised and used in place of conventional statistical tools. Written by prominent researchers in the field, the book focuses on high-dimensional and large-scale approximations and details the many basic multivariate methods used to achieve high levels of accuracy. The authors begin with a fundamental presentation of the basic

  5. Comparison of Drive Counts and Mark-Resight As Methods of Population Size Estimation of Highly Dense Sika Deer (Cervus nippon Populations.

    Directory of Open Access Journals (Sweden)

    Kazutaka Takeshita

    Full Text Available Assessing temporal changes in abundance indices is an important issue in the management of large herbivore populations. The drive counts method has been frequently used as a deer abundance index in mountainous regions. However, despite an inherent risk for observation errors in drive counts, which increase with deer density, evaluations of the utility of drive counts at a high deer density remain scarce. We compared the drive counts and mark-resight (MR methods in the evaluation of a highly dense sika deer population (MR estimates ranged between 11 and 53 individuals/km2 on Nakanoshima Island, Hokkaido, Japan, between 1999 and 2006. This deer population experienced two large reductions in density; approximately 200 animals in total were taken from the population through a large-scale population removal and a separate winter mass mortality event. Although the drive counts tracked temporal changes in deer abundance on the island, they overestimated the counts for all years in comparison to the MR method. Increased overestimation in drive count estimates after the winter mass mortality event may be due to a double count derived from increased deer movement and recovery of body condition secondary to the mitigation of density-dependent food limitations. Drive counts are unreliable because they are affected by unfavorable factors such as bad weather, and they are cost-prohibitive to repeat, which precludes the calculation of confidence intervals. Therefore, the use of drive counts to infer the deer abundance needs to be reconsidered.

  6. Analysis of counting data: Development of the SATLAS Python package

    Science.gov (United States)

    Gins, W.; de Groote, R. P.; Bissell, M. L.; Granados Buitrago, C.; Ferrer, R.; Lynch, K. M.; Neyens, G.; Sels, S.

    2018-01-01

    For the analysis of low-statistics counting experiments, a traditional nonlinear least squares minimization routine may not always provide correct parameter and uncertainty estimates due to the assumptions inherent in the algorithm(s). In response to this, a user-friendly Python package (SATLAS) was written to provide an easy interface between the data and a variety of minimization algorithms which are suited for analyzinglow, as well as high, statistics data. The advantage of this package is that it allows the user to define their own model function and then compare different minimization routines to determine the optimal parameter values and their respective (correlated) errors. Experimental validation of the different approaches in the package is done through analysis of hyperfine structure data of 203Fr gathered by the CRIS experiment at ISOLDE, CERN.

  7. Deep 3 GHz number counts from a P(D) fluctuation analysis

    Science.gov (United States)

    Vernstrom, T.; Scott, Douglas; Wall, J. V.; Condon, J. J.; Cotton, W. D.; Fomalont, E. B.; Kellermann, K. I.; Miller, N.; Perley, R. A.

    2014-05-01

    Radio source counts constrain galaxy populations and evolution, as well as the global star formation history. However, there is considerable disagreement among the published 1.4-GHz source counts below 100 μJy. Here, we present a statistical method for estimating the μJy and even sub-μJy source count using new deep wide-band 3-GHz data in the Lockman Hole from the Karl G. Jansky Very Large Array. We analysed the confusion amplitude distribution P(D), which provides a fresh approach in the form of a more robust model, with a comprehensive error analysis. We tested this method on a large-scale simulation, incorporating clustering and finite source sizes. We discuss in detail our statistical methods for fitting using Markov chain Monte Carlo, handling correlations, and systematic errors from the use of wide-band radio interferometric data. We demonstrated that the source count can be constrained down to 50 nJy, a factor of 20 below the rms confusion. We found the differential source count near 10 μJy to have a slope of -1.7, decreasing to about -1.4 at fainter flux densities. At 3 GHz, the rms confusion in an 8-arcsec full width at half-maximum beam is ˜ 1.2 μJy beam-1, and a radio background temperature ˜14 mK. Our counts are broadly consistent with published evolutionary models. With these results, we were also able to constrain the peak of the Euclidean normalized differential source count of any possible new radio populations that would contribute to the cosmic radio background down to 50 nJy.

  8. Reliability of joint count assessment in rheumatoid arthritis: a systematic literature review.

    Science.gov (United States)

    Cheung, Peter P; Gossec, Laure; Mak, Anselm; March, Lyn

    2014-06-01

    Joint counts are central to the assessment of rheumatoid arthritis (RA) but reliability is an issue. To evaluate the reliability and agreement of joint counts (intra-observer and inter-observer) by health care professionals (physicians, nurses, and metrologists) and patients in RA, and the impact of training and standardization on joint count reliability through a systematic literature review. Articles reporting joint count reliability or agreement in RA in PubMed, EMBase, and the Cochrane library between 1960 and 2012 were selected. Data were extracted regarding tender joint counts (TJCs) and swollen joint counts (SJCs) derived by physicians, metrologists, or patients for intra-observer and inter-observer reliability. In addition, methods and effects of training or standardization were extracted. Statistics expressing reliability such as intraclass correlation coefficients (ICCs) were extracted. Data analysis was primarily descriptive due to high heterogeneity. Twenty-eight studies on health care professionals (HCP) and 20 studies on patients were included. Intra-observer reliability for TJCs and SJCs was good for HCPs and patients (range of ICC: 0.49-0.98). Inter-observer reliability between HCPs for TJCs was higher than for SJCs (range of ICC: 0.64-0.88 vs. 0.29-0.98). Patient inter-observer reliability with HCPs as comparators was better for TJCs (range of ICC: 0.31-0.91) compared to SJCs (0.16-0.64). Nine studies (7 with HCPs and 2 with patients) evaluated consensus or training, with improvement in reliability of TJCs but conflicting evidence for SJCs. Intra- and inter-observer reliability was high for TJCs for HCPs and patients: among all groups, reliability was better for TJCs than SJCs. Inter-observer reliability of SJCs was poorer for patients than HCPs. Data were inconclusive regarding the potential for training to improve SJC reliability. Overall, the results support further evaluation for patient-reported joint counts as an outcome measure. © 2013

  9. Improvements in the energy resolution and high-count-rate performance of bismuth germanate

    International Nuclear Information System (INIS)

    Koehler, P.E.; Wender, S.A.; Kapustinsky, J.S.

    1985-01-01

    Several methods for improving the energy resolution of bismuth germanate (BGO) have been investigated. It is shown that some of these methods resulted in a substantial improvement in the energy resolution. In addition, a method to improve the performance of BGO at high counting rates has been systematically studied. The results of this study are presented and discussed

  10. Nutritional status and CD4 cell counts in patients with HIV/AIDS receiving antiretroviral therapy

    Directory of Open Access Journals (Sweden)

    Ana Celia Oliveira dos Santos

    2013-12-01

    Full Text Available Introduction Even with current highly active antiretroviral therapy, individuals with AIDS continue to exhibit important nutritional deficits and reduced levels of albumin and hemoglobin, which may be directly related to their cluster of differentiation 4 (CD4 cell counts. The aim of this study was to characterize the nutritional status of individuals with human immunodeficiency virus/acquired immunodeficiency syndrome (HIV/AIDS and relate the findings to the albumin level, hemoglobin level and CD4 cell count. Methods Patients over 20 years of age with AIDS who were hospitalized in a university hospital and were receiving antiretroviral therapy were studied with regard to clinical, anthropometric, biochemical and sociodemographic characteristics. Body mass index, percentage of weight loss, arm circumference, triceps skinfold and arm muscle circumference were analyzed. Data on albumin, hemoglobin, hematocrit and CD4 cell count were obtained from patient charts. Statistical analysis was performed using Fisher's exact test, Student's t-test for independent variables and the Mann-Whitney U-test. The level of significance was set to 0.05 (α = 5%. Statistical analysis was performed using Statistical Package for the Social Sciences (SPSS 17.0 software for Windows. Results Of the 50 patients evaluated, 70% were male. The prevalence of malnutrition was higher when the definition was based on arm circumference and triceps skinfold measurement. The concentrations of all biochemical variables were significantly lower among patients with a body mass index of less than 18.5kg/m2. The CD4 cell count, albumin, hemoglobin and hematocrit anthropometric measures were directly related to each other. Conclusions These findings underscore the importance of nutritional follow-up for underweight patients with AIDS, as nutritional status proved to be related to important biochemical alterations.

  11. Metal ion levels and lymphocyte counts

    DEFF Research Database (Denmark)

    Penny, Jeannette Ø; Varmarken, Jens-Erik; Ovesen, Ole

    2013-01-01

    BACKGROUND AND PURPOSE: Wear particles from metal-on-metal arthroplasties are under suspicion for adverse effects both locally and systemically, and the DePuy ASR Hip Resurfacing System (RHA) has above-average failure rates. We compared lymphocyte counts in RHA and total hip arthroplasty (THA) an....../ppb. INTERPRETATION: Circulating T-lymphocyte levels may decline after surgery, regardless of implant type. Metal ions-particularly cobalt-may have a general depressive effect on T- and B-lymphocyte levels. Registered with ClinicalTrials.gov under # NCT01113762.......BACKGROUND AND PURPOSE: Wear particles from metal-on-metal arthroplasties are under suspicion for adverse effects both locally and systemically, and the DePuy ASR Hip Resurfacing System (RHA) has above-average failure rates. We compared lymphocyte counts in RHA and total hip arthroplasty (THA....... RESULTS: The T-lymphocyte counts for both implant types declined over the 2-year period. This decline was statistically significant for CD3(+)CD8(+) in the THA group, with a regression coefficient of -0.04 × 10(9)cells/year (95% CI: -0.08 to -0.01). Regression analysis indicated a depressive effect...

  12. Pattern statistics on Markov chains and sensitivity to parameter estimation

    Directory of Open Access Journals (Sweden)

    Nuel Grégory

    2006-10-01

    Full Text Available Abstract Background: In order to compute pattern statistics in computational biology a Markov model is commonly used to take into account the sequence composition. Usually its parameter must be estimated. The aim of this paper is to determine how sensitive these statistics are to parameter estimation, and what are the consequences of this variability on pattern studies (finding the most over-represented words in a genome, the most significant common words to a set of sequences,.... Results: In the particular case where pattern statistics (overlap counting only computed through binomial approximations we use the delta-method to give an explicit expression of σ, the standard deviation of a pattern statistic. This result is validated using simulations and a simple pattern study is also considered. Conclusion: We establish that the use of high order Markov model could easily lead to major mistakes due to the high sensitivity of pattern statistics to parameter estimation.

  13. High Resolution Gamma Ray Spectroscopy at MHz Counting Rates With LaBr3 Scintillators for Fusion Plasma Applications

    Science.gov (United States)

    Nocente, M.; Tardocchi, M.; Olariu, A.; Olariu, S.; Pereira, R. C.; Chugunov, I. N.; Fernandes, A.; Gin, D. B.; Grosso, G.; Kiptily, V. G.; Neto, A.; Shevelev, A. E.; Silva, M.; Sousa, J.; Gorini, G.

    2013-04-01

    High resolution γ-ray spectroscopy measurements at MHz counting rates were carried out at nuclear accelerators, combining a LaBr 3(Ce) detector with dedicated hardware and software solutions based on digitization and off-line analysis. Spectra were measured at counting rates up to 4 MHz, with little or no degradation of the energy resolution, adopting a pile up rejection algorithm. The reported results represent a step forward towards the final goal of high resolution γ-ray spectroscopy measurements on a burning plasma device.

  14. Modelling the Covariance Structure in Marginal Multivariate Count Models

    DEFF Research Database (Denmark)

    Bonat, W. H.; Olivero, J.; Grande-Vega, M.

    2017-01-01

    The main goal of this article is to present a flexible statistical modelling framework to deal with multivariate count data along with longitudinal and repeated measures structures. The covariance structure for each response variable is defined in terms of a covariance link function combined...... be used to indicate whether there was statistical evidence of a decline in blue duikers and other species hunted during the study period. Determining whether observed drops in the number of animals hunted are indeed true is crucial to assess whether species depletion effects are taking place in exploited...... with a matrix linear predictor involving known matrices. In order to specify the joint covariance matrix for the multivariate response vector, the generalized Kronecker product is employed. We take into account the count nature of the data by means of the power dispersion function associated with the Poisson...

  15. Microbial counts of food contact surfaces at schools depending on a feeding scheme

    Directory of Open Access Journals (Sweden)

    Nthabiseng Nhlapo

    2014-11-01

    Full Text Available The prominence of disease transmission between individuals in confined environments is a concern, particularly in the educational environment. With respect to school feeding schemes, food contact surfaces have been shown to be potential vehicles of foodborne pathogens. The aim of this study was to assess the cleanliness of the surfaces that come into contact with food that is provided to children through the National School Nutrition Programme in central South Africa. In each school under study, microbiological samples were collected from the preparation surface and the dominant hand and apron of the food handler. The samples were analysed for total viable counts, coliforms, Escherichia coli, Staphylococcus aureus and yeasts and moulds. The criteria specified in the British Columbia Guide for Environmental Health Officers were used to evaluate the results. Total viable counts were high for all surfaces, with the majority of colonies being too numerous to count (over 100 colonies per plate. Counts of organisms were relatively low, with 20% of the surfaces producing unsatisfactory enumeration of S. aureus and E. coli and 30% unsatisfactory for coliforms. Yeast and mould produced 50% and 60% unsatisfactory counts from preparation surfaces and aprons, respectively. Statistically significant differences could not be established amongst microbial counts of the surfaces, which suggests cross-contamination may have occurred. Contamination may be attributed to foodstuffs and animals in the vicinity of the preparation area rather than to the food handlers, because hands had the lowest counts of enumerated organisms amongst the analysed surfaces.

  16. Statistics II essentials

    CERN Document Server

    Milewski, Emil G

    2012-01-01

    REA's Essentials provide quick and easy access to critical information in a variety of different fields, ranging from the most basic to the most advanced. As its name implies, these concise, comprehensive study guides summarize the essentials of the field covered. Essentials are helpful when preparing for exams, doing homework and will remain a lasting reference source for students, teachers, and professionals. Statistics II discusses sampling theory, statistical inference, independent and dependent variables, correlation theory, experimental design, count data, chi-square test, and time se

  17. Loss-Free Counting with Digital Signal Processors

    International Nuclear Information System (INIS)

    Markku Koskelo; Dave Hall; Martin Moslinger

    2000-01-01

    Loss-free-counting (LFC) techniques have frequently been used with traditional analog pulse processing systems to compensate for the time or pulses lost when a spectroscopy system is unavailable (busy) for processing an accepted pulse. With the availability of second-generation digital signal processing (DSP) electronics that offer a significantly improved performance for both high and low count rate applications, the LFC technique has been revisited. Specific attention was given to the high and ultra-high count rate behavior, using high-purity germanium (HPGe) detectors with both transistor reset preamplifiers (TRP) and conventional RC preamplifiers. The experiments conducted for this work show that the known LFC techniques further benefit when combined with modern DSP pulse shaping

  18. HIGH-RESOLUTION IMAGING OF THE ATLBS REGIONS: THE RADIO SOURCE COUNTS

    Energy Technology Data Exchange (ETDEWEB)

    Thorat, K.; Subrahmanyan, R.; Saripalli, L.; Ekers, R. D., E-mail: kshitij@rri.res.in [Raman Research Institute, C. V. Raman Avenue, Sadashivanagar, Bangalore 560080 (India)

    2013-01-01

    The Australia Telescope Low-brightness Survey (ATLBS) regions have been mosaic imaged at a radio frequency of 1.4 GHz with 6'' angular resolution and 72 {mu}Jy beam{sup -1} rms noise. The images (centered at R.A. 00{sup h}35{sup m}00{sup s}, decl. -67 Degree-Sign 00'00'' and R.A. 00{sup h}59{sup m}17{sup s}, decl. -67 Degree-Sign 00'00'', J2000 epoch) cover 8.42 deg{sup 2} sky area and have no artifacts or imaging errors above the image thermal noise. Multi-resolution radio and optical r-band images (made using the 4 m CTIO Blanco telescope) were used to recognize multi-component sources and prepare a source list; the detection threshold was 0.38 mJy in a low-resolution radio image made with beam FWHM of 50''. Radio source counts in the flux density range 0.4-8.7 mJy are estimated, with corrections applied for noise bias, effective area correction, and resolution bias. The resolution bias is mitigated using low-resolution radio images, while effects of source confusion are removed by using high-resolution images for identifying blended sources. Below 1 mJy the ATLBS counts are systematically lower than the previous estimates. Showing no evidence for an upturn down to 0.4 mJy, they do not require any changes in the radio source population down to the limit of the survey. The work suggests that automated image analysis for counts may be dependent on the ability of the imaging to reproduce connecting emission with low surface brightness and on the ability of the algorithm to recognize sources, which may require that source finding algorithms effectively work with multi-resolution and multi-wavelength data. The work underscores the importance of using source lists-as opposed to component lists-and correcting for the noise bias in order to precisely estimate counts close to the image noise and determine the upturn at sub-mJy flux density.

  19. Track counting in radon dosimetry

    International Nuclear Information System (INIS)

    Fesenbeck, Ingo; Koehler, Bernd; Reichert, Klaus-Martin

    2013-01-01

    The newly developed, computer-controlled track counting system is capable of imaging and analyzing the entire area of nuclear track detectors. The high optical resolution allows a new analysis approach for the process of automated counting using digital image processing technologies. This way, higher exposed detectors can be evaluated reliably by an automated process as well. (orig.)

  20. Lower white blood cell counts in elite athletes training for highly aerobic sports.

    Science.gov (United States)

    Horn, P L; Pyne, D B; Hopkins, W G; Barnes, C J

    2010-11-01

    White cell counts at rest might be lower in athletes participating in selected endurance-type sports. Here, we analysed blood tests of elite athletes collected over a 10-year period. Reference ranges were established for 14 female and 14 male sports involving 3,679 samples from 937 females and 4,654 samples from 1,310 males. Total white blood cell counts and counts of neutrophils, lymphocytes and monocytes were quantified. Each sport was scaled (1-5) for its perceived metabolic stress (aerobic-anaerobic) and mechanical stress (concentric-eccentric) by 13 sports physiologists. Substantially lower total white cell and neutrophil counts were observed in aerobic sports of cycling and triathlon (~16% of test results below the normal reference range) compared with team or skill-based sports such as water polo, cricket and volleyball. Mechanical stress of sports had less effect on the distribution of cell counts. The lower white cell counts in athletes in aerobic sports probably represent an adaptive response, not underlying pathology.

  1. A cell-based high-throughput screening assay for radiation susceptibility using automated cell counting

    International Nuclear Information System (INIS)

    Hodzic, Jasmina; Dingjan, Ilse; Maas, Mariëlle JP; Meulen-Muileman, Ida H van der; Menezes, Renee X de; Heukelom, Stan; Verheij, Marcel; Gerritsen, Winald R; Geldof, Albert A; Triest, Baukelien van; Beusechem, Victor W van

    2015-01-01

    Radiotherapy is one of the mainstays in the treatment for cancer, but its success can be limited due to inherent or acquired resistance. Mechanisms underlying radioresistance in various cancers are poorly understood and available radiosensitizers have shown only modest clinical benefit. There is thus a need to identify new targets and drugs for more effective sensitization of cancer cells to irradiation. Compound and RNA interference high-throughput screening technologies allow comprehensive enterprises to identify new agents and targets for radiosensitization. However, the gold standard assay to investigate radiosensitivity of cancer cells in vitro, the colony formation assay (CFA), is unsuitable for high-throughput screening. We developed a new high-throughput screening method for determining radiation susceptibility. Fast and uniform irradiation of batches up to 30 microplates was achieved using a Perspex container and a clinically employed linear accelerator. The readout was done by automated counting of fluorescently stained nuclei using the Acumen eX3 laser scanning cytometer. Assay performance was compared to that of the CFA and the CellTiter-Blue homogeneous uniform-well cell viability assay. The assay was validated in a whole-genome siRNA library screening setting using PC-3 prostate cancer cells. On 4 different cancer cell lines, the automated cell counting assay produced radiation dose response curves that followed a linear-quadratic equation and that exhibited a better correlation to the results of the CFA than did the cell viability assay. Moreover, the cell counting assay could be used to detect radiosensitization by silencing DNA-PKcs or by adding caffeine. In a high-throughput screening setting, using 4 Gy irradiated and control PC-3 cells, the effects of DNA-PKcs siRNA and non-targeting control siRNA could be clearly discriminated. We developed a simple assay for radiation susceptibility that can be used for high-throughput screening. This will aid

  2. High Sensitivity Detection of Xe Isotopes Via Beta-Gamma Coincidence Counting

    International Nuclear Information System (INIS)

    Bowyer, Ted W.; McIntyre, Justin I.; Reeder, Paul L.

    1999-01-01

    Measurement of xenon fission product isotopes is a key element in the global network being established to monitor the Comprehensive Nuclear-Test-Ban Treaty. Pacific Northwest National Laboratory has developed an automated system for separating Xe from air which includes a beta-gamma counting system for 131mXe, 133mXe, 133Xe, and 135Xe. Betas and conversion electrons are detected in a plastic scintillation cell containing the Xe sample. The counting geometry is nearly 100% for beta and conversion electrons. The resolution in the pulse height spectrum from the plastic scintillator is sufficient to observe distinct peaks for specific conversion electrons. Gamma and X-rays are detected in a NaI(Tl) scintillation detector which surrounds the plastic scintillator sample cell. Two-dimensional pulse height spectra of gamma energy versus beta energy are obtained. Each of the four xenon isotopes has a distinctive signature in the two-dimensional energy array. The details of the counting system, examples of two-dimensional beta-gamma data, and operational experience with this counting system will be described

  3. Haematological changes in HIV infection with correlation to CD4 cell count

    Directory of Open Access Journals (Sweden)

    SS Parinitha

    2012-03-01

    Full Text Available BackgroundHIV infection is associated with a wide range of haematological abnormalities.Methods and ObjectivesThe objectives in this study were to study haematological changes in HIV patients and to correlate them with CD4 cell counts. Two hundred and fifty HIV positive patients referred to the haematology laboratory section for complete haemogram in whom CD4 count was done were included in the study. Haematologic parameters and CD4 counts were studied in each of these patients.Descriptive statistics were applied. Association between two attributes was calculated by chi-square test and p value less than 0.05 was considered statistically significant.ResultsAmong 250 patients, anaemia was seen in 210 (84% cases. The most common type was normocytic normochromic (40.4%. Lymphopenia was seen in 163 (65.2% cases and thrombocytopenia in 45 (18% cases. The majority of cases (70% had CD4 cell counts below 200 cells/mm3. Fifty-four cases (21.6% had CD4 counts between 200 to 499 cells/mm3 and 21 (8.4% cases had CD4 counts more than 500 cells/ mm3.In patients with CD4 counts less than 200 cells/mm3, anaemia was seen in 91.4% cases, leucopenia in 26.8%cases, lymphopenia in 80% cases and thrombocytopenia in 21.7% cases.ConclusionHaematologic manifestations of HIV infection are common and more frequent with progression of disease. The present study revealed a significant increase in the number of cases of anaemia, and lymphopenia, with decreasing CD4 cell counts. Thrombocytopenia is also seen but does not show significant increase with disease progression. The study also highlights the importance of simultaneously treating HIV patients for haematologic manifestations to reduce morbidity.

  4. High levels of viral suppression among East African HIV-infected women and men in serodiscordant partnerships initiating antiretroviral therapy with high CD4 counts and during pregnancy.

    Science.gov (United States)

    Mujugira, Andrew; Baeten, Jared; Kidoguchi, Lara; Haberer, Jessica; Celum, Connie; Donnell, Deborah; Ngure, Kenneth; Bukusi, Elizabeth; Mugo, Nelly; Asiimwe, Stephen; Odoyo, Josephine; Tindimwebwa, Edna; Bulya, Nulu; Katabira, Elly; Heffron, Renee

    2017-09-13

    People who are asymptomatic and feel healthy, including pregnant women, may be less motivated to initiate ART or achieve high adherence. We assessed whether ART initiation, and viral suppression 6, 12 and 24-months after ART initiation, were lower in HIV-infected members of serodiscordant couples who initiated during pregnancy or with higher CD4 counts. We used data from the Partners Demonstration Project, an open-label study of the delivery of integrated PrEP and ART (at any CD4 count) for HIV prevention among high-risk HIV serodiscordant couples in Kenya and Uganda. Differences in viral suppression (HIV RNA 500 cells/mm3) and during pregnancy were estimated using Poisson regression. Of 865 HIV-infected participants retained after becoming eligible for ART during study follow-up, 95% initiated ART. Viral suppression 24-months after ART initiation was high overall (97%), and comparable among those initiating ART at CD4 counts >500, 351-500 and ≤350 cells/mm3 (96% vs 97% vs 97%; relative risk [RR] 0.98; 95% CI: 0.93-1.03 for CD4 >500 vs <350 and RR 0.99; 95% CI: (0.93-1.06) for CD4 351-500 vs ≤350). Viral suppression was as likely among women initiating ART primarily to prevent perinatal transmission as ART initiation for other reasons (p=0.9 at 6 months and p=0.5 at 12 months). Nearly all HIV-infected partners initiating ART were virally suppressed by 24 months, irrespective of CD4 count or pregnancy status. These findings suggest that people initiating ART at high CD4 counts or due to pregnancy can adhere to ART as well as those starting treatment with symptomatic HIV disease or low CD4 counts.

  5. Students’ Learning Obstacles and Alternative Solution in Counting Rules Learning Levels Senior High School

    Directory of Open Access Journals (Sweden)

    M A Jatmiko

    2017-12-01

    Full Text Available The counting rules is a topic in mathematics senior high school. In the learning process, teachers often find students who have difficulties in learning this topic. Knowing the characteristics of students' learning difficulties and analyzing the causes is important for the teacher, as an effort in trying to reflect the learning process and as a reference in constructing alternative learning solutions which appropriate to anticipate students’ learning obstacles. This study uses qualitative methods and involves 70 students of class XII as research subjects. The data collection techniques used in this study is diagnostic test instrument about learning difficulties in counting rules, observation, and interview. The data used to know the learning difficulties experienced by students, the causes of learning difficulties, and to develop alternative learning solutions. From the results of data analysis, the results of diagnostic tests researcher found some obstacles faced by students, such as students get confused in describing the definition, students difficulties in understanding the procedure of solving multiplication rules. Based on those problems, researcher analyzed the causes of these difficulties and make hypothetical learning trajectory as an alternative solution in counting rules learning.

  6. Statistical method evaluation for differentially methylated CpGs in base resolution next-generation DNA sequencing data.

    Science.gov (United States)

    Zhang, Yun; Baheti, Saurabh; Sun, Zhifu

    2018-05-01

    High-throughput bisulfite methylation sequencing such as reduced representation bisulfite sequencing (RRBS), Agilent SureSelect Human Methyl-Seq (Methyl-seq) or whole-genome bisulfite sequencing is commonly used for base resolution methylome research. These data are represented either by the ratio of methylated cytosine versus total coverage at a CpG site or numbers of methylated and unmethylated cytosines. Multiple statistical methods can be used to detect differentially methylated CpGs (DMCs) between conditions, and these methods are often the base for the next step of differentially methylated region identification. The ratio data have a flexibility of fitting to many linear models, but the raw count data take consideration of coverage information. There is an array of options in each datatype for DMC detection; however, it is not clear which is an optimal statistical method. In this study, we systematically evaluated four statistic methods on methylation ratio data and four methods on count-based data and compared their performances with regard to type I error control, sensitivity and specificity of DMC detection and computational resource demands using real RRBS data along with simulation. Our results show that the ratio-based tests are generally more conservative (less sensitive) than the count-based tests. However, some count-based methods have high false-positive rates and should be avoided. The beta-binomial model gives a good balance between sensitivity and specificity and is preferred method. Selection of methods in different settings, signal versus noise and sample size estimation are also discussed.

  7. Some Statistics for Measuring Large-Scale Structure

    OpenAIRE

    Brandenberger, Robert H.; Kaplan, David M.; A, Stephen; Ramsey

    1993-01-01

    Good statistics for measuring large-scale structure in the Universe must be able to distinguish between different models of structure formation. In this paper, two and three dimensional ``counts in cell" statistics and a new ``discrete genus statistic" are applied to toy versions of several popular theories of structure formation: random phase cold dark matter model, cosmic string models, and global texture scenario. All three statistics appear quite promising in terms of differentiating betw...

  8. Technical feasibility proof for high-resolution low-dose photon-counting CT of the breast

    Energy Technology Data Exchange (ETDEWEB)

    Kalender, Willi A.; Kolditz, Daniel; Lueck, Ferdinand [University of Erlangen-Nuernberg, Institute of Medical Physics (IMP), Erlangen (Germany); CT Imaging GmbH, Erlangen (Germany); Steiding, Christian [University of Erlangen-Nuernberg, Institute of Medical Physics (IMP), Erlangen (Germany); CT Imaging GmbH, Erlangen (Germany); University Hospital of Erlangen, Institute of Radiology, Erlangen (Germany); Ruth, Veikko; Roessler, Ann-Christin [University of Erlangen-Nuernberg, Institute of Medical Physics (IMP), Erlangen (Germany); Wenkel, Evelyn [University Hospital of Erlangen, Institute of Radiology, Erlangen (Germany)

    2017-03-15

    X-ray computed tomography (CT) has been proposed and evaluated multiple times as a potentially alternative method for breast imaging. All efforts shown so far have been criticized and partly disapproved because of their limited spatial resolution and higher patient dose when compared to mammography. Our concept for a dedicated breast CT (BCT) scanner therefore aimed at novel apparatus and detector design to provide high spatial resolution of about 100 μm and average glandular dose (AGD) levels of 5 mGy or below. Photon-counting technology was considered as a solution to reach these goals. The complete concept was previously evaluated and confirmed by simulations and basic experiments on laboratory setups. We here present measurements of dose, technical image quality parameters and surgical specimen results on such a scanner. For comparison purposes, the specimens were also imaged with digital mammography (DM) and breast tomosynthesis (BT) apparatus. Results show that photon-counting BCT (pcBCT) at 5 mGy AGD offers sufficiently high 3D spatial resolution for reliable detectability of calcifications and soft tissue delineation. (orig.)

  9. Performance of Drift-Tube Detectors at High Counting Rates for High-Luminosity LHC Upgrades

    CERN Document Server

    Bittner, Bernhard; Kortner, Oliver; Kroha, Hubert; Manfredini, Alessandro; Nowak, Sebastian; Ott, Sebastian; Richter, Robert; Schwegler, Philipp; Zanzi, Daniele; Biebel, Otmar; Hertenberger, Ralf; Ruschke, Alexander; Zibell, Andre

    2016-01-01

    The performance of pressurized drift-tube detectors at very high background rates has been studied at the Gamma Irradiation Facility (GIF) at CERN and in an intense 20 MeV proton beam at the Munich Van-der-Graaf tandem accelerator for applications in large-area precision muon tracking at high-luminosity upgrades of the Large Hadron Collider (LHC). The ATLAS muon drifttube (MDT) chambers with 30 mm tube diameter have been designed to cope with and neutron background hit rates of up to 500 Hz/square cm. Background rates of up to 14 kHz/square cm are expected at LHC upgrades. The test results with standard MDT readout electronics show that the reduction of the drift-tube diameter to 15 mm, while leaving the operating parameters unchanged, vastly increases the rate capability well beyond the requirements. The development of new small-diameter muon drift-tube (sMDT) chambers for LHC upgrades is completed. Further improvements of tracking e?ciency and spatial resolution at high counting rates will be achieved with ...

  10. Comprehensive nuclear counting and detector characterisation system for the radiochemistry laboratory

    International Nuclear Information System (INIS)

    Parthasarathy, R.; Saisubalakshmi, D.; Mishra, G.K.; Srinivas, K.C.; Venkatasubramani, C.R.

    2004-01-01

    The paper describes a comprehensive nuclear pulse counting system that can cater to up to seven nuclear detector set-ups located in different places in the laboratory. Each detector set up has an interfacing module that conditions the amplifier pulses and transmits them to a common counting system. The microcontroller-based system receives these pulses through a multiplexer and counts the pulses for a user specified preset time. The system has a routine to determine detector plateau characteristics and fix the detector operating voltage. In this mode, the system collects the EHT-versus- counts data in a EHT programmed sequence and plots the profile. The system conducts the counting routine for a stipulated number of times and does all necessary statistical tests to ensure the proper functioning of the detector under test. The system also includes a test routine that checks the performance of the counting system by connecting it to a local pulse generator. The microcontroller based system interacts with a PC through RS232 communication for user interaction and reporting. (author)

  11. Correction for decay during counting in gamma spectrometry

    International Nuclear Information System (INIS)

    Nir-El, Y.

    2013-01-01

    A basic result in gamma spectrometry is the count rate of a relevant peak. Correction for decay during counting and expressing the count rate at the beginning of the measurement can be done by a multiplicative factor that is derived from integrating the count rate over time. The counting time substituted in this factor must be the live time, whereas the use of the real-time is an error that underestimates the count rate by about the dead-time (DT) (in percentage). This error of underestimation of the count rate is corroborated in the measurement of a nuclide with a high DT. The present methodology is not applicable in systems that include a zero DT correction function. (authors)

  12. Analysis of Parasite and Other Skewed Counts

    Science.gov (United States)

    Alexander, Neal

    2012-01-01

    Objective To review methods for the statistical analysis of parasite and other skewed count data. Methods Statistical methods for skewed count data are described and compared, with reference to those used over a ten year period of Tropical Medicine and International Health. Two parasitological datasets are used for illustration. Results Ninety papers were identified, 89 with descriptive and 60 with inferential analysis. A lack of clarity is noted in identifying measures of location, in particular the Williams and geometric mean. The different measures are compared, emphasizing the legitimacy of the arithmetic mean for skewed data. In the published papers, the t test and related methods were often used on untransformed data, which is likely to be invalid. Several approaches to inferential analysis are described, emphasizing 1) non-parametric methods, while noting that they are not simply comparisons of medians, and 2) generalized linear modelling, in particular with the negative binomial distribution. Additional methods, such as the bootstrap, with potential for greater use are described. Conclusions Clarity is recommended when describing transformations and measures of location. It is suggested that non-parametric methods and generalized linear models are likely to be sufficient for most analyses. PMID:22943299

  13. Alpha scintillation radon counting

    International Nuclear Information System (INIS)

    Lucas, H.F. Jr.

    1977-01-01

    Radon counting chambers which utilize the alpha-scintillation properties of silver activated zinc sulfide are simple to construct, have a high efficiency, and, with proper design, may be relatively insensitive to variations in the pressure or purity of the counter filling. Chambers which were constructed from glass, metal, or plastic in a wide variety of shapes and sizes were evaluated for the accuracy and the precision of the radon counting. The principles affecting the alpha-scintillation radon counting chamber design and an analytic system suitable for a large scale study of the 222 Rn and 226 Ra content of either air or other environmental samples are described. Particular note is taken of those factors which affect the accuracy and the precision of the method for monitoring radioactivity around uranium mines

  14. Statistical learning methods in high-energy and astrophysics analysis

    Energy Technology Data Exchange (ETDEWEB)

    Zimmermann, J. [Forschungszentrum Juelich GmbH, Zentrallabor fuer Elektronik, 52425 Juelich (Germany) and Max-Planck-Institut fuer Physik, Foehringer Ring 6, 80805 Munich (Germany)]. E-mail: zimmerm@mppmu.mpg.de; Kiesling, C. [Max-Planck-Institut fuer Physik, Foehringer Ring 6, 80805 Munich (Germany)

    2004-11-21

    We discuss several popular statistical learning methods used in high-energy- and astro-physics analysis. After a short motivation for statistical learning we present the most popular algorithms and discuss several examples from current research in particle- and astro-physics. The statistical learning methods are compared with each other and with standard methods for the respective application.

  15. Statistical learning methods in high-energy and astrophysics analysis

    International Nuclear Information System (INIS)

    Zimmermann, J.; Kiesling, C.

    2004-01-01

    We discuss several popular statistical learning methods used in high-energy- and astro-physics analysis. After a short motivation for statistical learning we present the most popular algorithms and discuss several examples from current research in particle- and astro-physics. The statistical learning methods are compared with each other and with standard methods for the respective application

  16. High Throughput, High Yield Fabrication of High Quantum Efficiency Back-Illuminated Photon Counting, Far UV, UV, and Visible Detector Arrays

    Science.gov (United States)

    Nikzad, Shouleh; Hoenk, M. E.; Carver, A. G.; Jones, T. J.; Greer, F.; Hamden, E.; Goodsall, T.

    2013-01-01

    In this paper we discuss the high throughput end-to-end post fabrication processing of high performance delta-doped and superlattice-doped silicon imagers for UV, visible, and NIR applications. As an example, we present our results on far ultraviolet and ultraviolet quantum efficiency (QE) in a photon counting, detector array. We have improved the QE by nearly an order of magnitude over microchannel plates (MCPs) that are the state-of-the-art UV detectors for many NASA space missions as well as defense applications. These achievements are made possible by precision interface band engineering of Molecular Beam Epitaxy (MBE) and Atomic Layer Deposition (ALD).

  17. Statistical analysis of the count and profitability of air conditioners.

    Science.gov (United States)

    Rady, El Houssainy A; Mohamed, Salah M; Abd Elmegaly, Alaa A

    2018-08-01

    This article presents the statistical analysis of the number and profitability of air conditioners in an Egyptian company. Checking the same distribution for each categorical variable has been made using Kruskal-Wallis test.

  18. RBC count

    Science.gov (United States)

    ... by kidney disease) RBC destruction ( hemolysis ) due to transfusion, blood vessel injury, or other cause Leukemia Malnutrition Bone ... slight risk any time the skin is broken) Alternative Names Erythrocyte count; Red blood cell count; Anemia - RBC count Images Blood test ...

  19. Poisson statistics application in modelling of neutron detection

    International Nuclear Information System (INIS)

    Avdic, S.; Marinkovic, P.

    1996-01-01

    The main purpose of this study is taking into account statistical analysis of the experimental data which were measured by 3 He neutron spectrometer. The unfolding method based on principle of maximum likelihood incorporates the Poisson approximation of counting statistics applied (aithor)

  20. Effects of quantum coherence on work statistics

    Science.gov (United States)

    Xu, Bao-Ming; Zou, Jian; Guo, Li-Sha; Kong, Xiang-Mu

    2018-05-01

    In the conventional two-point measurement scheme of quantum thermodynamics, quantum coherence is destroyed by the first measurement. But as we know the coherence really plays an important role in the quantum thermodynamics process, and how to describe the work statistics for a quantum coherent process is still an open question. In this paper, we use the full counting statistics method to investigate the effects of quantum coherence on work statistics. First, we give a general discussion and show that for a quantum coherent process, work statistics is very different from that of the two-point measurement scheme, specifically the average work is increased or decreased and the work fluctuation can be decreased by quantum coherence, which strongly depends on the relative phase, the energy level structure, and the external protocol. Then, we concretely consider a quenched one-dimensional transverse Ising model and show that quantum coherence has a more significant influence on work statistics in the ferromagnetism regime compared with that in the paramagnetism regime, so that due to the presence of quantum coherence the work statistics can exhibit the critical phenomenon even at high temperature.

  1. TasselNet: counting maize tassels in the wild via local counts regression network.

    Science.gov (United States)

    Lu, Hao; Cao, Zhiguo; Xiao, Yang; Zhuang, Bohan; Shen, Chunhua

    2017-01-01

    performance, with a mean absolute error of 6.6 and a mean squared error of 9.6 averaged over 8 test sequences. TasselNet can achieve robust in-field counting of maize tassels with a relatively high degree of accuracy. Our experimental evaluations also suggest several good practices for practitioners working on maize-tassel-like counting problems. It is worth noting that, though the counting errors have been greatly reduced by TasselNet, in-field counting of maize tassels remains an open and unsolved problem.

  2. TasselNet: counting maize tassels in the wild via local counts regression network

    Directory of Open Access Journals (Sweden)

    Hao Lu

    2017-11-01

    margins and achieves the overall best counting performance, with a mean absolute error of 6.6 and a mean squared error of 9.6 averaged over 8 test sequences. Conclusions TasselNet can achieve robust in-field counting of maize tassels with a relatively high degree of accuracy. Our experimental evaluations also suggest several good practices for practitioners working on maize-tassel-like counting problems. It is worth noting that, though the counting errors have been greatly reduced by TasselNet, in-field counting of maize tassels remains an open and unsolved problem.

  3. SCOTCH: Secure Counting Of encrypTed genomiC data using a Hybrid approach.

    Science.gov (United States)

    Chenghong, Wang; Jiang, Yichen; Mohammed, Noman; Chen, Feng; Jiang, Xiaoqian; Al Aziz, Md Momin; Sadat, Md Nazmus; Wang, Shuang

    2017-01-01

    As genomic data are usually at large scale and highly sensitive, it is essential to enable both efficient and secure analysis, by which the data owner can securely delegate both computation and storage on untrusted public cloud. Counting query of genotypes is a basic function for many downstream applications in biomedical research (e.g., computing allele frequency, calculating chi-squared statistics, etc.). Previous solutions show promise on secure counting of outsourced data but the efficiency is still a big limitation for real world applications. In this paper, we propose a novel hybrid solution to combine a rigorous theoretical model (homomorphic encryption) and the latest hardware-based infrastructure (i.e., Software Guard Extensions) to speed up the computation while preserving the privacy of both data owners and data users. Our results demonstrated efficiency by using the real data from the personal genome project.

  4. Use of the lymphocyte count as a diagnostic screen in adults with suspected Epstein-Barr virus infectious mononucleosis.

    Science.gov (United States)

    Biggs, Timothy C; Hayes, Stephen M; Bird, Jonathan H; Harries, Philip G; Salib, Rami J

    2013-10-01

    To evaluate the predictive diagnostic accuracy of the lymphocyte count in Epstein-Barr virus-related infectious mononucleosis (IM). Retrospective case note and blood results review within a university-affiliated teaching hospital. A retrospective review of 726 patients undergoing full blood count and Monospot testing was undertaken. Monospot testing outcomes were compared with the lymphocyte count, examining for significant statistical correlations. With a lymphocyte count of ≤4 × 10(9) /L, 99% of patients had an associated negative Monospot result (sensitivity of 84% and specificity of 94%). A group subanalysis of the population older than 18 years with a lymphocyte count ≤4 × 10(9) /L revealed that 100% were Monospot negative (sensitivity of 100% and specificity of 97%). A lymphocyte count of ≤4 × 10(9) /L correlated significantly with a negative Monospot result. A lymphocyte count of ≤4 × 10(9) /L appears to be a highly reliable predictor of a negative Monospot result, particularly in the population aged >18 years. Pediatric patients, and adults with strongly suggestive symptoms and signs of IM, should still undergo Monospot testing. However, in adults with more subtle symptoms and signs, representing the vast majority, Monospot testing should be restricted to those with a lymphocyte count >4 × 10(9) /L. NA Copyright © 2013 The American Laryngological, Rhinological and Otological Society, Inc.

  5. Multichannel analyzer with real-time correction of counting losses based on a fast 16/32 bit microprocessor

    International Nuclear Information System (INIS)

    Westphal, G.P.; Kasa, T.

    1984-01-01

    It is demonstrated that from a modern microprocessor with 32 bit architecture and from standard VLSI peripheral chips a multichannel analyzer with real-time correction of counting losses may be designed in a very flexible yet cost-effective manner. Throughput rates of 100,000 events/second are a good match even for high-rate spectroscopy systems and may be further enhanced by the use of already available CPU chips with higher clock frequency. Low power consumption and a very compact form factor make the design highly recommendable for portable applications. By means of a simple and easily reproducible rotating sample device the dynamic response of the VPG counting loss correction method have been tested and found to be more than sufficient for conceivable real-time applications. Enhanced statistical accuracy of correction factors may be traded against speed of response by the mere change of one preset value which lends itself to the simple implementation of self-adapting systems. Reliability as well as user convenience is improved by self-calibration of pulse evolution time in the VPG counting loss correction unit

  6. Overdispersion in allelic counts and θ-correction in forensic genetics

    DEFF Research Database (Denmark)

    Tvedebrink, Torben

    2010-01-01

    We present a statistical model for incorporating the extra variability in allelic counts due to subpopulation structures. In forensic genetics, this effect is modelled by the identical-by-descent parameter θ, which measures the relationship between pairs of alleles within a population relative...... with computation of the profile log-likelihood, confidence intervals and hypothesis testing. In order to compare our method with existing methods, we reanalysed FBI data from Budowle and Moretti (1999) with allele counts in six US subpopulations. Furthermore, we investigate properties of our methodology from...

  7. Intraoperative detection of 18F-FDG-avid tissue sites using the increased probe counting efficiency of the K-alpha probe design and variance-based statistical analysis with the three-sigma criteria

    International Nuclear Information System (INIS)

    Povoski, Stephen P; Chapman, Gregg J; Murrey, Douglas A; Lee, Robert; Martin, Edward W; Hall, Nathan C

    2013-01-01

    Intraoperative detection of 18 F-FDG-avid tissue sites during 18 F-FDG-directed surgery can be very challenging when utilizing gamma detection probes that rely on a fixed target-to-background (T/B) ratio (ratiometric threshold) for determination of probe positivity. The purpose of our study was to evaluate the counting efficiency and the success rate of in situ intraoperative detection of 18 F-FDG-avid tissue sites (using the three-sigma statistical threshold criteria method and the ratiometric threshold criteria method) for three different gamma detection probe systems. Of 58 patients undergoing 18 F-FDG-directed surgery for known or suspected malignancy using gamma detection probes, we identified nine 18 F-FDG-avid tissue sites (from amongst seven patients) that were seen on same-day preoperative diagnostic PET/CT imaging, and for which each 18 F-FDG-avid tissue site underwent attempted in situ intraoperative detection concurrently using three gamma detection probe systems (K-alpha probe, and two commercially-available PET-probe systems), and then were subsequently surgical excised. The mean relative probe counting efficiency ratio was 6.9 (± 4.4, range 2.2–15.4) for the K-alpha probe, as compared to 1.5 (± 0.3, range 1.0–2.1) and 1.0 (± 0, range 1.0–1.0), respectively, for two commercially-available PET-probe systems (P < 0.001). Successful in situ intraoperative detection of 18 F-FDG-avid tissue sites was more frequently accomplished with each of the three gamma detection probes tested by using the three-sigma statistical threshold criteria method than by using the ratiometric threshold criteria method, specifically with the three-sigma statistical threshold criteria method being significantly better than the ratiometric threshold criteria method for determining probe positivity for the K-alpha probe (P = 0.05). Our results suggest that the improved probe counting efficiency of the K-alpha probe design used in conjunction with the three

  8. A High-Speed, Event-Driven, Active Pixel Sensor Readout for Photon-Counting Microchannel Plate Detectors

    Science.gov (United States)

    Kimble, Randy A.; Pain, Bedabrata; Norton, Timothy J.; Haas, J. Patrick; Oegerle, William R. (Technical Monitor)

    2002-01-01

    Silicon array readouts for microchannel plate intensifiers offer several attractive features. In this class of detector, the electron cloud output of the MCP intensifier is converted to visible light by a phosphor; that light is then fiber-optically coupled to the silicon array. In photon-counting mode, the resulting light splashes on the silicon array are recognized and centroided to fractional pixel accuracy by off-chip electronics. This process can result in very high (MCP-limited) spatial resolution while operating at a modest MCP gain (desirable for dynamic range and long term stability). The principal limitation of intensified CCD systems of this type is their severely limited local dynamic range, as accurate photon counting is achieved only if there are not overlapping event splashes within the frame time of the device. This problem can be ameliorated somewhat by processing events only in pre-selected windows of interest of by using an addressable charge injection device (CID) for the readout array. We are currently pursuing the development of an intriguing alternative readout concept based on using an event-driven CMOS Active Pixel Sensor. APS technology permits the incorporation of discriminator circuitry within each pixel. When coupled with suitable CMOS logic outside the array area, the discriminator circuitry can be used to trigger the readout of small sub-array windows only when and where an event splash has been detected, completely eliminating the local dynamic range problem, while achieving a high global count rate capability and maintaining high spatial resolution. We elaborate on this concept and present our progress toward implementing an event-driven APS readout.

  9. Total bacterial count and somatic cell count in refrigerated raw milk stored in communal tanks

    Directory of Open Access Journals (Sweden)

    Edmar da Costa Alves

    2014-09-01

    Full Text Available The current industry demand for dairy products with extended shelf life has resulted in new challenges for milk quality maintenance. The processing of milk with high bacterial counts compromises the quality and performance of industrial products. The study aimed to evaluate the total bacteria counts (TBC and somatic cell count (SCC in 768 samples of refrigerated raw milk, from 32 communal tanks. Samples were collected in the first quarter of 2010, 2011, 2012 and 2013 and analyzed by the Laboratory of Milk Quality - LQL. Results showed that 62.5%, 37.5%, 15.6% and 27.1% of the means for TBC in 2010, 2011, 2012 and 2013, respectively, were above the values established by legislation. However, we observed a significant reduction in the levels of total bacterial count (TBC in the studied periods. For somatic cell count, 100% of the means indicated values below 600.000 cells/mL, complying with the actual Brazilian legislation. The values found for the somatic cell count suggests the adoption of effective measures for the sanitary control of the herd. However, the results must be considered with caution as it highlights the need for quality improvements of the raw material until it achieves reliable results effectively.

  10. Effect of recirculation and regional counting rate on reliability of noninvasive bicompartmental CBF measurements

    International Nuclear Information System (INIS)

    Herholz, K.

    1985-01-01

    Based on data from routine intravenous Xe133-rCBF studies in 50 patients, using Obrist's algorithm the effect of counting rate statistics and amount of recirculating activity on reproducibility of results was investigated at five simulated counting rate levels. Dependence of the standard deviation of compartmental and noncompartmental flow parameters on recirculation and counting rate was determined by multiple linear regression analysis. Those regression equations permit determination of the optimum accuracy that may be expected from individual flow measurements. Mainly due to a delay of the start-of-fit time an exponential increase in standard deviation of flow measurements was observed as recirculation increased. At constant start-of-fit, however, a linear increase in standard deviation of compartmental flow parameters only was found, while noncompartmental results remained constant. Therefore, and in regard to other studies of potential sources of error, an upper limit of 2.5 min for the start-of-fit time and usage of noncompartmental flow parameters for measurements affected by high recirculation are suggested

  11. Proposals of counting method for bubble detectors and their intercomparisons

    International Nuclear Information System (INIS)

    Ramalho, Eduardo; Silva, Ademir X.; Bellido, Luis F.; Facure, Alessandro; Pereira, Mario

    2009-01-01

    The study of neutron's spectrometry and dosimetry has become significantly easier due to relatively new devices called bubble detectors. Insensitive to gamma rays and composed by superheated emulsions, they still are subjects of many researches in Radiation Physics and Nuclear Engineering. In bubble detectors, either exposed to more intense neutron fields or for a long time, when more bubbles are produced, the statistical uncertainty during the dosimetric and spectrometric processes is reduced. A proposal of this nature is set up in this work, which presents ways to perform counting processes for bubble detectors and an updated proceeding to get the irradiated detectors' images in order to make the manual counting easier. Twelve BDS detectors were irradiated by RDS111 cyclotron from IEN's (Instituto de Engenharia Nuclear) and photographed using an assembly specially designed for this experiment. Counting was proceeded manually in a first moment; simultaneously, ImagePro was used in order to perform counting automatically. The bubble counting values, either manual or automatic, were compared and the time to get them and their difficult levels as well. After the bubble counting, the detectors' standardizes responses were calculated in both cases, according to BDS's manual and they were also compared. Among the results, the counting on these devices really becomes very hard at a large number of bubbles, besides higher variations in counting of many bubbles. Because of the good agreement between manual counting and the custom program, the last one revealed a good alternative in practical and economical levels. Despite the good results, the custom program needs of more adjustments in order to achieve more accuracy on higher counting on bubble detectors for neutron measurement applications. (author)

  12. Statistical behavior of high doses in medical radiodiagnosis

    International Nuclear Information System (INIS)

    Barboza, Adriana Elisa

    2014-01-01

    This work has as main purpose statistically estimating occupational exposure in medical diagnostic radiology in cases of high doses recorded in 2011 at national level. For statistical survey of this study, doses of 372 IOE's diagnostic radiology in different Brazilian states were evaluated. Data were extracted from the work of monograph (Research Methodology Of High Doses In Medical Radiodiagnostic) that contains the database's information Sector Management doses of IRD/CNEN-RJ, Brazil. The identification of these states allows the Sanitary Surveillance (VISA) responsible, becomes aware of events and work with programs to reduce these events. (author)

  13. Soft X ray spectrometry at high count rates

    International Nuclear Information System (INIS)

    Blanc, P.; Brouquet, P.; Uhre, N.

    1978-06-01

    Two modifications of the classical method of X-ray spectrometry by a semi-conductor diode permit a count rate of 10 5 c/s with an energy resolution of 350 eV. With a specially constructed pulse height analyzer, this detector can measure four spectra of 5 ms each, in the range of 1-30 keV, during a plasma shot

  14. Numerical reconstruction of photon-number statistics from photocounting statistics: Regularization of an ill-posed problem

    International Nuclear Information System (INIS)

    Starkov, V. N.; Semenov, A. A.; Gomonay, H. V.

    2009-01-01

    We demonstrate a practical possibility of loss compensation in measured photocounting statistics in the presence of dark counts and background radiation noise. It is shown that satisfactory results are obtained even in the case of low detection efficiency and large experimental errors.

  15. Delta count-rate monitoring system

    International Nuclear Information System (INIS)

    Van Etten, D.; Olsen, W.A.

    1985-01-01

    A need for a more effective way to rapidly search for gamma-ray contamination over large areas led to the design and construction of a very sensitive gamma detection system. The delta count-rate monitoring system was installed in a four-wheel-drive van instrumented for environmental surveillance and accident response. The system consists of four main sections: (1) two scintillation detectors, (2) high-voltage power supply amplifier and single-channel analyzer, (3) delta count-rate monitor, and (4) count-rate meter and recorder. The van's 6.5-kW generator powers the standard nuclear instrument modular design system. The two detectors are mounted in the rear corners of the van and can be run singly or jointly. A solid-state bar-graph count-rate meter mounted on the dashboard can be read easily by both the driver and passenger. A solid-state strip chart recorder shows trends and provides a permanent record of the data. An audible alarm is sounded at the delta monitor and at the dashboard count-rate meter if a detected radiation level exceeds the set background level by a predetermined amount

  16. Study on the behaviour of timing photomultipliers at a high counting rate

    International Nuclear Information System (INIS)

    Gladyshev, D.A.; Li, B.N.; Yunusov, Kh.R.

    1978-01-01

    Variations in the amplification factor K of a photomultiplier (PMU) with the accuracy of 1% in a pulse mode are studied. Measurements were performed by means of a light pulse generator based on a light diode which generates pulses at the repetition rate of 250-10 5 pulse/s. Relative variations in K were determined by the position of the peak gravity centre from the light diode using a pulse analyzer and a frequency meter. Results of PM testing show that, at a sudden counting rate increase, the amplification increases during the time period less than, the measurement time (less than 1 s) and returns to the stationary value. When the counting rate returns from 10 5 pulse/s to the initial value of 250 pulse/s, the amplification decreases and than increases to stationary value. The total time of K stabilization after counting rate applying constitutes 10-70 min. Restoration of K after counting rate removal occurs to be much slower, during 3 hr. 40 min. K values varied from 1 to 12%

  17. An automated approach for annual layer counting in ice cores

    DEFF Research Database (Denmark)

    Winstrup, Mai; Svensson, A. M.; Rasmussen, S. O.

    2012-01-01

    A novel method for automated annual layer counting in seasonally-resolved paleoclimate records has been developed. It relies on algorithms from the statistical framework of Hidden Markov Models (HMMs), which originally was developed for use in machine speech-recognition. The strength of the layer...

  18. An automated approach for annual layer counting in ice cores

    DEFF Research Database (Denmark)

    Winstrup, Mai; Svensson, A. M.; Rasmussen, S. O.

    2012-01-01

    A novel method for automated annual layer counting in seasonally-resolved paleoclimate records has been developed. It relies on algorithms from the statistical framework of hidden Markov models (HMMs), which originally was developed for use in machine speech recognition. The strength of the layer...

  19. Nondestructive detection of total viable count changes of chilled pork in high oxygen storage condition based on hyperspectral technology

    Science.gov (United States)

    Zheng, Xiaochun; Peng, Yankun; Li, Yongyu; Chao, Kuanglin; Qin, Jianwei

    2017-05-01

    The plate count method is commonly used to detect the total viable count (TVC) of bacteria in pork, which is timeconsuming and destructive. It has also been used to study the changes of the TVC in pork under different storage conditions. In recent years, many scholars have explored the non-destructive methods on detecting TVC by using visible near infrared (VIS/NIR) technology and hyperspectral technology. The TVC in chilled pork was monitored under high oxygen condition in this study by using hyperspectral technology in order to evaluate the changes of total bacterial count during storage, and then evaluate advantages and disadvantages of the storage condition. The VIS/NIR hyperspectral images of samples stored in high oxygen condition was acquired by a hyperspectral system in range of 400 1100nm. The actual reference value of total bacteria was measured by standard plate count method, and the results were obtained in 48 hours. The reflection spectra of the samples are extracted and used for the establishment of prediction model for TVC. The spectral preprocessing methods of standard normal variate transformation (SNV), multiple scatter correction (MSC) and derivation was conducted to the original reflectance spectra of samples. Partial least squares regression (PLSR) of TVC was performed and optimized to be the prediction model. The results show that the near infrared hyperspectral technology based on 400-1100nm combined with PLSR model can describe the growth pattern of the total bacteria count of the chilled pork under the condition of high oxygen very vividly and rapidly. The results obtained in this study demonstrate that the nondestructive method of TVC based on NIR hyperspectral has great potential in monitoring of edible safety in processing and storage of meat.

  20. Counting probe

    International Nuclear Information System (INIS)

    Matsumoto, Haruya; Kaya, Nobuyuki; Yuasa, Kazuhiro; Hayashi, Tomoaki

    1976-01-01

    Electron counting method has been devised and experimented for the purpose of measuring electron temperature and density, the most fundamental quantities to represent plasma conditions. Electron counting is a method to count the electrons in plasma directly by equipping a probe with the secondary electron multiplier. It has three advantages of adjustable sensitivity, high sensitivity of the secondary electron multiplier, and directional property. Sensitivity adjustment is performed by changing the size of collecting hole (pin hole) on the incident front of the multiplier. The probe is usable as a direct reading thermometer of electron temperature because it requires to collect very small amount of electrons, thus it doesn't disturb the surrounding plasma, and the narrow sweep width of the probe voltage is enough. Therefore it can measure anisotropy more sensitively than a Langmuir probe, and it can be used for very low density plasma. Though many problems remain on anisotropy, computer simulation has been carried out. Also it is planned to provide a Helmholtz coil in the vacuum chamber to eliminate the effect of earth magnetic field. In practical experiments, the measurement with a Langmuir probe and an emission probe mounted to the movable structure, the comparison with the results obtained in reverse magnetic field by using a Helmholtz coil, and the measurement of ionic sound wave are scheduled. (Wakatsuki, Y.)

  1. Statistical analysis of angular correlation measurements

    International Nuclear Information System (INIS)

    Oliveira, R.A.A.M. de.

    1986-01-01

    Obtaining the multipole mixing ratio, δ, of γ transitions in angular correlation measurements is a statistical problem characterized by the small number of angles in which the observation is made and by the limited statistic of counting, α. The inexistence of a sufficient statistics for the estimator of δ, is shown. Three different estimators for δ were constructed and their properties of consistency, bias and efficiency were tested. Tests were also performed in experimental results obtained in γ-γ directional correlation measurements. (Author) [pt

  2. Exponential decay and exponential recovery of modal gains in high count rate channel electron multipliers

    International Nuclear Information System (INIS)

    Hahn, S.F.; Burch, J.L.

    1980-01-01

    A series of data on high count rate channel electron multipliers revealed an initial drop and subsequent recovery of gains in exponential fashion. The FWHM of the pulse height distribution at the initial stage of testing can be used as a good criterion for the selection of operating bias voltage of the channel electron multiplier

  3. Immature germ cells in semen - correlation with total sperm count and sperm motility

    Directory of Open Access Journals (Sweden)

    Priya S Patil

    2013-01-01

    Conclusions: Round cells in semen can be differentiated into immature germ cells and leucocytes using simple staining methods. The differential counts mentioned in a semen report give valuable and clinically relevant information. In this study, we observed a negative correlation between total count and immature germ cells, as well as sperm motility and shedding of immature germ cells. The latter was statistically significant with a P value 0.000.

  4. Intraoperative detection of ¹⁸F-FDG-avid tissue sites using the increased probe counting efficiency of the K-alpha probe design and variance-based statistical analysis with the three-sigma criteria.

    Science.gov (United States)

    Povoski, Stephen P; Chapman, Gregg J; Murrey, Douglas A; Lee, Robert; Martin, Edward W; Hall, Nathan C

    2013-03-04

    Intraoperative detection of (18)F-FDG-avid tissue sites during 18F-FDG-directed surgery can be very challenging when utilizing gamma detection probes that rely on a fixed target-to-background (T/B) ratio (ratiometric threshold) for determination of probe positivity. The purpose of our study was to evaluate the counting efficiency and the success rate of in situ intraoperative detection of (18)F-FDG-avid tissue sites (using the three-sigma statistical threshold criteria method and the ratiometric threshold criteria method) for three different gamma detection probe systems. Of 58 patients undergoing (18)F-FDG-directed surgery for known or suspected malignancy using gamma detection probes, we identified nine (18)F-FDG-avid tissue sites (from amongst seven patients) that were seen on same-day preoperative diagnostic PET/CT imaging, and for which each (18)F-FDG-avid tissue site underwent attempted in situ intraoperative detection concurrently using three gamma detection probe systems (K-alpha probe, and two commercially-available PET-probe systems), and then were subsequently surgical excised. The mean relative probe counting efficiency ratio was 6.9 (± 4.4, range 2.2-15.4) for the K-alpha probe, as compared to 1.5 (± 0.3, range 1.0-2.1) and 1.0 (± 0, range 1.0-1.0), respectively, for two commercially-available PET-probe systems (P < 0.001). Successful in situ intraoperative detection of 18F-FDG-avid tissue sites was more frequently accomplished with each of the three gamma detection probes tested by using the three-sigma statistical threshold criteria method than by using the ratiometric threshold criteria method, specifically with the three-sigma statistical threshold criteria method being significantly better than the ratiometric threshold criteria method for determining probe positivity for the K-alpha probe (P = 0.05). Our results suggest that the improved probe counting efficiency of the K-alpha probe design used in conjunction with the three-sigma statistical

  5. Overdispersion in allelic counts and θ-correction in forensic genetics

    DEFF Research Database (Denmark)

    Tvedebrink, Torben

    2009-01-01

    A statistical model for incorporating the extra variability in allelic counts due to subpopulation structures is presented. In forensic genetics, this effect is modelled by the identical-by-decent-parameter, θ . It is shown, that θ may be defined as an overdispersion parameter capturing the extra...

  6. Heterogeneous counting on filter support media

    International Nuclear Information System (INIS)

    Long, E.; Kohler, V.; Kelly, M.J.

    1976-01-01

    Many investigators in the biomedical research area have used filter paper as the support for radioactive samples. This means that a heterogeneous counting of sample sometimes results. The count rate of a sample on a filter will be affected by positioning, degree of dryness, sample application procedure, the type of filter, and the type of cocktail used. Positioning of the filter (up or down) in the counting vial can cause a variation of 35% or more when counting tritiated samples on filter paper. Samples of varying degrees of dryness when added to the counting cocktail can cause nonreproducible counts if handled improperly. Count rates starting at 2400 CPM initially can become 10,000 CPM in 24 hours for 3 H-DNA (deoxyribonucleic acid) samples dried on standard cellulose acetate membrane filters. Data on cellulose nitrate filters show a similar trend. Sample application procedures in which the sample is applied to the filter in a small spot or on a large amount of the surface area can cause nonreproducible or very low counting rates. A tritiated DNA sample, when applied topically, gives a count rate of 4,000 CPM. When the sample is spread over the whole filter, 13,400 CPM are obtained with a much better coefficient of variation (5% versus 20%). Adding protein carrier (bovine serum albumin-BSA) to the sample to trap more of the tritiated DNA on the filter during the filtration process causes a serious beta absorption problem. Count rates which are one-fourth the count rate applied to the filter are obtained on calibrated runs. Many of the problems encountered can be alleviated by a proper choice of filter and the use of a liquid scintillation cocktail which dissolves the filter. Filter-Solv has been used to dissolve cellulose nitrate filters and filters which are a combination of cellulose nitrate and cellulose acetate. Count rates obtained for these dissolved samples are very reproducible and highly efficient

  7. Determining random counts in liquid scintillation counting

    International Nuclear Information System (INIS)

    Horrocks, D.L.

    1979-01-01

    During measurements involving coincidence counting techniques, errors can arise due to the detection of chance or random coincidences in the multiple detectors used. A method and the electronic circuits necessary are here described for eliminating this source of error in liquid scintillation detectors used in coincidence counting. (UK)

  8. Statistical properties of laser light scattering in Brownian medium

    International Nuclear Information System (INIS)

    Suwono; Santoso, Budi; Baiquni, A.

    1983-01-01

    Relationship between statistical properties of laser light scattering in Brownian medium and photon-counting distributions are described in detail. A coherence optical detection has been constructed and by using photon-counting technique the ensemble distribution of the scattered field within space and time coherence has been measured. Good agreement between theory and experiment is shown. (author)

  9. Immature germ cells in semen - correlation with total sperm count and sperm motility.

    Science.gov (United States)

    Patil, Priya S; Humbarwadi, Rajendra S; Patil, Ashalata D; Gune, Anita R

    2013-07-01

    Current data regarding infertility suggests that male factor contributes up to 30% of the total cases of infertility. Semen analysis reveals the presence of spermatozoa as well as a number of non-sperm cells, presently being mentioned in routine semen report as "round cells" without further differentiating them into leucocytes or immature germ cells. The aim of this work was to study a simple, cost-effective, and convenient method for differentiating the round cells in semen into immature germ cells and leucocytes and correlating them with total sperm counts and motility. Semen samples from 120 males, who had come for investigation for infertility, were collected, semen parameters recorded, and stained smears studied for different round cells. Statistical analysis of the data was done to correlate total sperm counts and sperm motility with the occurrence of immature germ cells and leucocytes. The average shedding of immature germ cells in different groups with normal and low sperm counts was compared. The clinical significance of "round cells" in semen and their differentiation into leucocytes and immature germ cells are discussed. Round cells in semen can be differentiated into immature germ cells and leucocytes using simple staining methods. The differential counts mentioned in a semen report give valuable and clinically relevant information. In this study, we observed a negative correlation between total count and immature germ cells, as well as sperm motility and shedding of immature germ cells. The latter was statistically significant with a P value 0.000.

  10. Radio source counts: comments on their convergence and assessment of the contribution to fluctuations of the microwave background

    International Nuclear Information System (INIS)

    Danese, L.; De Zotti, G.; Mandolesi, N.

    1982-01-01

    We point out that statistically estimated high frequency counts at milli-Jansky levels exhibit a slower convergence than expected on the basis of extrapolations of counts at higher flux densities and at longer wavelengths. This seems to demand a substantial cosmological evolution for at least a sub-population of flat-spectrum sources different from QSO's, a fact that might have important implications also in connection with the problem of the origin of the X-ray background. We also compute the discrete source contributions to small scale fluctuations in the Rayleigh-Jeans region of the cosmic microwave background and we show that they set a serious limit to the searches for truly primordial anisotropies using conventional radio-astronomical techniques

  11. A low-cost, scalable, current-sensing digital headstage for high channel count μECoG

    Science.gov (United States)

    Trumpis, Michael; Insanally, Michele; Zou, Jialin; Elsharif, Ashraf; Ghomashchi, Ali; Sertac Artan, N.; Froemke, Robert C.; Viventi, Jonathan

    2017-04-01

    Objective. High channel count electrode arrays allow for the monitoring of large-scale neural activity at high spatial resolution. Implantable arrays featuring many recording sites require compact, high bandwidth front-end electronics. In the present study, we investigated the use of a small, light weight, and low cost digital current-sensing integrated circuit for acquiring cortical surface signals from a 61-channel micro-electrocorticographic (μECoG) array. Approach. We recorded both acute and chronic μECoG signal from rat auditory cortex using our novel digital current-sensing headstage. For direct comparison, separate recordings were made in the same anesthetized preparations using an analog voltage headstage. A model of electrode impedance explained the transformation between current- and voltage-sensed signals, and was used to reconstruct cortical potential. We evaluated the digital headstage using several metrics of the baseline and response signals. Main results. The digital current headstage recorded neural signal with similar spatiotemporal statistics and auditory frequency tuning compared to the voltage signal. The signal-to-noise ratio of auditory evoked responses (AERs) was significantly stronger in the current signal. Stimulus decoding based on true and reconstructed voltage signals were not significantly different. Recordings from an implanted system showed AERs that were detectable and decodable for 52 d. The reconstruction filter mitigated the thermal current noise of the electrode impedance and enhanced overall SNR. Significance. We developed and validated a novel approach to headstage acquisition that used current-input circuits to independently digitize 61 channels of μECoG measurements of the cortical field. These low-cost circuits, intended to measure photo-currents in digital imaging, not only provided a signal representing the local cortical field with virtually the same sensitivity and specificity as a traditional voltage headstage but

  12. A computer simulation used to investigate optimization in low level counting

    International Nuclear Information System (INIS)

    Brown, R.C.; Kephart, G.S.

    1984-01-01

    The differential form of the interval distribution for randomly spaced events such as radioactive decay is represented as dP/sub t/=ae - /supat/dt, the Poisson distribution. As applied to radioactive decay, this states that the probability (dP/sub t/) of the duration of a particular interval (elapsed time between counts) will be between t and t+dt as a function of the count rate (a). Thus a logarithmic transformation of this probability distribution results in a linear function whose slope and intercept are defined by the count rate. The effort expended in defining the interval distribution of a given radiation measurement equates in the laboratory to measuring and accumulating discrete time intervals between events rather than the usual approach of counting events per unit time. It follows from basic information theory that this greater effort should result in an improved statistical confidence in determinations of the ''true'' count rate (a). Using a random number generator as an analog of the discrete decay event, the authors have devised a Monte Carlo approach to investigate application of the above theory to the low level counting situation. This investigative approach is well suited to sensitivity analyses such that any constraints on proposed optimization techniques can be well defined prior to introducing these methods into the counting requirements in the laboratory

  13. A Framework for Assessing High School Students' Statistical Reasoning.

    Science.gov (United States)

    Chan, Shiau Wei; Ismail, Zaleha; Sumintono, Bambang

    2016-01-01

    Based on a synthesis of literature, earlier studies, analyses and observations on high school students, this study developed an initial framework for assessing students' statistical reasoning about descriptive statistics. Framework descriptors were established across five levels of statistical reasoning and four key constructs. The former consisted of idiosyncratic reasoning, verbal reasoning, transitional reasoning, procedural reasoning, and integrated process reasoning. The latter include describing data, organizing and reducing data, representing data, and analyzing and interpreting data. In contrast to earlier studies, this initial framework formulated a complete and coherent statistical reasoning framework. A statistical reasoning assessment tool was then constructed from this initial framework. The tool was administered to 10 tenth-grade students in a task-based interview. The initial framework was refined, and the statistical reasoning assessment tool was revised. The ten students then participated in the second task-based interview, and the data obtained were used to validate the framework. The findings showed that the students' statistical reasoning levels were consistent across the four constructs, and this result confirmed the framework's cohesion. Developed to contribute to statistics education, this newly developed statistical reasoning framework provides a guide for planning learning goals and designing instruction and assessments.

  14. A high count rate position decoding and energy measuring method for nuclear cameras using Anger logic detectors

    International Nuclear Information System (INIS)

    Wong, W.H.; Li, H.; Uribe, J.

    1998-01-01

    A new method for processing signals from Anger position-sensitive detectors used in gamma cameras and PET is proposed for very high count-rate imaging where multiple-event pileups are the norm. This method is designed to sort out and recover every impinging event from multiple-event pileups while maximizing the collection of scintillation signal for every event to achieve optimal accuracy in the measurement of energy and position. For every detected event, this method cancels the remnant signals from previous events, and excludes the pileup of signals from following events. The remnant subtraction is exact even for multiple pileup events. A prototype circuit for energy recovery demonstrated that the maximum count rates can be increased by more than 10 times comparing to the pulse-shaping method, and the energy resolution is as good as pulse shaping (or fixed integration) at low count rates. At 2 x 10 6 events/sec on NaI(Tl), the true counts acquired with this method is 3.3 times more than the delay-line clipping method (256 ns clipping) due to events recovered from pileups. Pulse-height spectra up to 3.5 x 10 6 events/sec have been studied. Monte Carlo simulation studies have been performed for image-quality comparisons between different processing methods

  15. Tower counts

    Science.gov (United States)

    Woody, Carol Ann; Johnson, D.H.; Shrier, Brianna M.; O'Neal, Jennifer S.; Knutzen, John A.; Augerot, Xanthippe; O'Neal, Thomas A.; Pearsons, Todd N.

    2007-01-01

    Counting towers provide an accurate, low-cost, low-maintenance, low-technology, and easily mobilized escapement estimation program compared to other methods (e.g., weirs, hydroacoustics, mark-recapture, and aerial surveys) (Thompson 1962; Siebel 1967; Cousens et al. 1982; Symons and Waldichuk 1984; Anderson 2000; Alaska Department of Fish and Game 2003). Counting tower data has been found to be consistent with that of digital video counts (Edwards 2005). Counting towers do not interfere with natural fish migration patterns, nor are fish handled or stressed; however, their use is generally limited to clear rivers that meet specific site selection criteria. The data provided by counting tower sampling allow fishery managers to determine reproductive population size, estimate total return (escapement + catch) and its uncertainty, evaluate population productivity and trends, set harvest rates, determine spawning escapement goals, and forecast future returns (Alaska Department of Fish and Game 1974-2000 and 1975-2004). The number of spawning fish is determined by subtracting subsistence, sport-caught fish, and prespawn mortality from the total estimated escapement. The methods outlined in this protocol for tower counts can be used to provide reasonable estimates ( plus or minus 6%-10%) of reproductive salmon population size and run timing in clear rivers. 

  16. Fractional counts-the simulation of low probability events

    International Nuclear Information System (INIS)

    Coldwell, R.L.; Lasche, G.P.; Jadczyk, A.

    2001-01-01

    The code RobSim has been added to RobWin.1 It simulates spectra resulting from gamma rays striking an array of detectors made up of different components. These are frequently used to set coincidence and anti-coincidence windows that decide if individual events are part of the signal. The first problem addressed is the construction of the detector. Then owing to the statistical nature of the responses of these elements there is a random nature in the response that can be taken into account by including fractional counts in the output spectrum. This somewhat complicates the error analysis, as Poisson statistics are no longer applicable

  17. Regression models for categorical, count, and related variables an applied approach

    CERN Document Server

    Hoffmann, John P

    2016-01-01

    Social science and behavioral science students and researchers are often confronted with data that are categorical, count a phenomenon, or have been collected over time. Sociologists examining the likelihood of interracial marriage, political scientists studying voting behavior, criminologists counting the number of offenses people commit, health scientists studying the number of suicides across neighborhoods, and psychologists modeling mental health treatment success are all interested in outcomes that are not continuous. Instead, they must measure and analyze these events and phenomena in a discrete manner.   This book provides an introduction and overview of several statistical models designed for these types of outcomes--all presented with the assumption that the reader has only a good working knowledge of elementary algebra and has taken introductory statistics and linear regression analysis.   Numerous examples from the social sciences demonstrate the practical applications of these models. The chapte...

  18. Fluorescence decay data analysis correcting for detector pulse pile-up at very high count rates

    Science.gov (United States)

    Patting, Matthias; Reisch, Paja; Sackrow, Marcus; Dowler, Rhys; Koenig, Marcelle; Wahl, Michael

    2018-03-01

    Using time-correlated single photon counting for the purpose of fluorescence lifetime measurements is usually limited in speed due to pile-up. With modern instrumentation, this limitation can be lifted significantly, but some artifacts due to frequent merging of closely spaced detector pulses (detector pulse pile-up) remain an issue to be addressed. We propose a data analysis method correcting for this type of artifact and the resulting systematic errors. It physically models the photon losses due to detector pulse pile-up and incorporates the loss in the decay fit model employed to obtain fluorescence lifetimes and relative amplitudes of the decay components. Comparison of results with and without this correction shows a significant reduction of systematic errors at count rates approaching the excitation rate. This allows quantitatively accurate fluorescence lifetime imaging at very high frame rates.

  19. Study on advancement of in vivo counting using mathematical simulation

    Energy Technology Data Exchange (ETDEWEB)

    Kinase, Sakae [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    2003-05-01

    To obtain an assessment of the committed effective dose, individual monitoring for the estimation of intakes of radionuclides is required. For individual monitoring of exposure to intakes of radionuclides, direct measurement of radionuclides in the body - in vivo counting- is very useful. To advance in a precision in vivo counting which fulfills the requirements of ICRP 1990 recommendations, some problems, such as the investigation of uncertainties in estimates of body burdens by in vivo counting, and the selection of the way to improve the precision, have been studied. In the present study, a calibration technique for in vivo counting application using Monte Carlo simulation was developed. The advantage of the technique is that counting efficiency can be obtained for various shapes and sizes that are very difficult to change for phantoms. To validate the calibration technique, the response functions and counting efficiencies of a whole-body counter installed in JAERI were evaluated using the simulation and measurements. Consequently, the calculations are in good agreement with the measurements. The method for the determination of counting efficiency curves as a function of energy was developed using the present technique and a physiques correction equation was derived from the relationship between parameters of correction factor and counting efficiencies of the JAERI whole-body counter. The uncertainties in body burdens of {sup 137}Cs estimated with the JAERI whole-body counter were also investigated using the Monte Carlo simulation and measurements. It was found that the uncertainties of body burdens estimated with the whole-body counter are strongly dependent on various sources of uncertainty such as radioactivity distribution within the body and counting statistics. Furthermore, the evaluation method of the peak efficiencies of a Ge semi-conductor detector was developed by Monte Carlo simulation for optimum arrangement of Ge semi-conductor detectors for

  20. Bayesian model selection techniques as decision support for shaping a statistical analysis plan of a clinical trial: an example from a vertigo phase III study with longitudinal count data as primary endpoint.

    Science.gov (United States)

    Adrion, Christine; Mansmann, Ulrich

    2012-09-10

    A statistical analysis plan (SAP) is a critical link between how a clinical trial is conducted and the clinical study report. To secure objective study results, regulatory bodies expect that the SAP will meet requirements in pre-specifying inferential analyses and other important statistical techniques. To write a good SAP for model-based sensitivity and ancillary analyses involves non-trivial decisions on and justification of many aspects of the chosen setting. In particular, trials with longitudinal count data as primary endpoints pose challenges for model choice and model validation. In the random effects setting, frequentist strategies for model assessment and model diagnosis are complex and not easily implemented and have several limitations. Therefore, it is of interest to explore Bayesian alternatives which provide the needed decision support to finalize a SAP. We focus on generalized linear mixed models (GLMMs) for the analysis of longitudinal count data. A series of distributions with over- and under-dispersion is considered. Additionally, the structure of the variance components is modified. We perform a simulation study to investigate the discriminatory power of Bayesian tools for model criticism in different scenarios derived from the model setting. We apply the findings to the data from an open clinical trial on vertigo attacks. These data are seen as pilot data for an ongoing phase III trial. To fit GLMMs we use a novel Bayesian computational approach based on integrated nested Laplace approximations (INLAs). The INLA methodology enables the direct computation of leave-one-out predictive distributions. These distributions are crucial for Bayesian model assessment. We evaluate competing GLMMs for longitudinal count data according to the deviance information criterion (DIC) or probability integral transform (PIT), and by using proper scoring rules (e.g. the logarithmic score). The instruments under study provide excellent tools for preparing decisions

  1. Demonstration of fundamental statistics by studying timing of electronics signals in a physics-based laboratory

    Science.gov (United States)

    Beach, Shaun E.; Semkow, Thomas M.; Remling, David J.; Bradt, Clayton J.

    2017-07-01

    We have developed accessible methods to demonstrate fundamental statistics in several phenomena, in the context of teaching electronic signal processing in a physics-based college-level curriculum. A relationship between the exponential time-interval distribution and Poisson counting distribution for a Markov process with constant rate is derived in a novel way and demonstrated using nuclear counting. Negative binomial statistics is demonstrated as a model for overdispersion and justified by the effect of electronic noise in nuclear counting. The statistics of digital packets on a computer network are shown to be compatible with the fractal-point stochastic process leading to a power-law as well as generalized inverse Gaussian density distributions of time intervals between packets.

  2. Multiplicity counting from fission detector signals with time delay effects

    Science.gov (United States)

    Nagy, L.; Pázsit, I.; Pál, L.

    2018-03-01

    In recent work, we have developed the theory of using the first three auto- and joint central moments of the currents of up to three fission chambers to extract the singles, doubles and triples count rates of traditional multiplicity counting (Pázsit and Pál, 2016; Pázsit et al., 2016). The objective is to elaborate a method for determining the fissile mass, neutron multiplication, and (α, n) neutron emission rate of an unknown assembly of fissile material from the statistics of the fission chamber signals, analogous to the traditional multiplicity counting methods with detectors in the pulse mode. Such a method would be an alternative to He-3 detector systems, which would be free from the dead time problems that would be encountered in high counting rate applications, for example the assay of spent nuclear fuel. A significant restriction of our previous work was that all neutrons born in a source event (spontaneous fission) were assumed to be detected simultaneously, which is not fulfilled in reality. In the present work, this restriction is eliminated, by assuming an independent, identically distributed random time delay for all neutrons arising from one source event. Expressions are derived for the same auto- and joint central moments of the detector current(s) as in the previous case, expressed with the singles, doubles, and triples (S, D and T) count rates. It is shown that if the time-dispersion of neutron detections is of the same order of magnitude as the detector pulse width, as they typically are in measurements of fast neutrons, the multiplicity rates can still be extracted from the moments of the detector current, although with more involved calibration factors. The presented formulae, and hence also the performance of the proposed method, are tested by both analytical models of the time delay as well as with numerical simulations. Methods are suggested also for the modification of the method for large time delay effects (for thermalised neutrons).

  3. Basics of statistical physics

    CERN Document Server

    Müller-Kirsten, Harald J W

    2013-01-01

    Statistics links microscopic and macroscopic phenomena, and requires for this reason a large number of microscopic elements like atoms. The results are values of maximum probability or of averaging. This introduction to statistical physics concentrates on the basic principles, and attempts to explain these in simple terms supplemented by numerous examples. These basic principles include the difference between classical and quantum statistics, a priori probabilities as related to degeneracies, the vital aspect of indistinguishability as compared with distinguishability in classical physics, the differences between conserved and non-conserved elements, the different ways of counting arrangements in the three statistics (Maxwell-Boltzmann, Fermi-Dirac, Bose-Einstein), the difference between maximization of the number of arrangements of elements, and averaging in the Darwin-Fowler method. Significant applications to solids, radiation and electrons in metals are treated in separate chapters, as well as Bose-Eins...

  4. Kids Count: The State of the Child in Tennessee, 1994.

    Science.gov (United States)

    Tennessee State Commission on Children and Youth, Nashville.

    This Kids Count report examines trends in the well being of Tennessee's children. The statistical portrait is based on 23 indicators of child well being: (1) single-parent families; (2) family income/poverty; (3) children receiving Aid to Families with Dependent Children; (4) students participating in school nutrition programs; (5) teen…

  5. Improvement and automatization of a proportional alpha-beta counting system - FAG

    International Nuclear Information System (INIS)

    German, U.; Levinson, S.; Pelled, O.; Shemesh, Y.; Assido, H.

    1997-01-01

    An alpha and beta counting system - FAG*, for planchette samples is operated at the Health Physics department's laboratory of the NRCN. The original operation mode of the system was based on manual tasks handled by the FHT1 100 electronics. An option for a basic computer keyboard operation was available too. A computer with an appropriate I/O card was connected to the system and a new operating program was developed which enables full automatic control of the various components. The program includes activity calculations and statistical checks as well as data management. A bar-code laser system for sample number reading was integrated into the Alpha-Beta automatic counting system. The sample identification by means of an attached bar-code label enables unmistakable and reliable attribution of results to the counted sample. authors)

  6. Sex based levels of C-reactive protein and white blood cell count in subjects with metabolic syndrome: Isfahan Healthy Heart Program.

    Science.gov (United States)

    Gharipour, Mojgan; Ramezani, Mohammad Arash; Sadeghi, Masuomeh; Khosravi, Alireza; Masjedi, Mohsen; Khosravi-Boroujeni, Hossein; Rafieian-Kopaei, Mahmoud; Sarrafzadegan, Nizal

    2013-06-01

    C-reactive protein (CRP) and white blood cell (WBC) are proinflammatory markers. They are major pathophysiological for the development of metabolic syndrome (MetS). This study aimed to address the independent associations between MetS and WBC counts and serum CRP levels and evaluation of their magnitude in relation to the MetS, based on the sex in the Iranian adults. In this cross-sectional study, subjects who met the MetS criteria, based on the Adult Treatment Panel III were selected from the Isfahan Healthy Heart Program database. A questionnaire containing the demographic data, weight, height, waist, and hip circumference of the respondents was completed for each person. Blood pressure was measured and the anthropometric measurements were done, and fasting blood samples were taken for 2 h postload plasma glucose (2 hpp). Serum [total, high-density lipoprotein (HDL), and low-density lipoprotein] levels of cholesterol, triglyceride, and CRP as well as WBC counts were determined. The univariate analyses were carried out to assess the relation between the CRP levels, WBC counts with the MetS in both sexes the. In men with the abdominal obesity, the higher levels of WBC count, high serum triglyceride and blood glucose levels, a low serum HDL level, and raised systolic and diastolic blood pressure were observed. However, the higher serum CRP levels were only observed in those with the low serum HDL-cholesterol levels. The mean values of the WBC counts were statistically different between the men with and without MetS, but the mean values of the CRP levels were similar between the two groups. In women, the mean values of WBC count and CRP levels were statistically different in the subjects with and without a MetS components (except for the low serum HDL levels and high diastolic blood pressure for the WBC measures and abdominal obesity for the CRP measures) and for those with and without MetS. The age and smoking adjusted changes in the CRP levels and WBC counts

  7. Applied categorical and count data analysis

    CERN Document Server

    Tang, Wan; Tu, Xin M

    2012-01-01

    Introduction Discrete Outcomes Data Source Outline of the BookReview of Key Statistical ResultsSoftwareContingency Tables Inference for One-Way Frequency TableInference for 2 x 2 TableInference for 2 x r TablesInference for s x r TableMeasures of AssociationSets of Contingency Tables Confounding Effects Sets of 2 x 2 TablesSets of s x r TablesRegression Models for Categorical Response Logistic Regression for Binary ResponseInference about Model ParametersGoodness of FitGeneralized Linear ModelsRegression Models for Polytomous ResponseRegression Models for Count Response Poisson Regression Mode

  8. Relationship between salivary flow rates and Candida albicans counts.

    Science.gov (United States)

    Navazesh, M; Wood, G J; Brightman, V J

    1995-09-01

    Seventy-one persons (48 women, 23 men; mean age, 51.76 years) were evaluated for salivary flow rates and Candida albicans counts. Each person was seen on three different occasions. Samples of unstimulated whole, chewing-stimulated whole, acid-stimulated parotid, and candy-stimulated parotid saliva were collected under standardized conditions. An oral rinse was also obtained and evaluated for Candida albicans counts. Unstimulated and chewing-stimulated whole flow rates were negatively and significantly (p Candida counts. Unstimulated whole saliva significantly (p Candida counts of 0 versus or = 500 count. Differences in stimulated parotid flow rates were not significant among different levels of Candida counts. The results of this study reveal that whole saliva is a better predictor than parotid saliva in identification of persons with high Candida albicans counts.

  9. Hidden Markov models for zero-inflated Poisson counts with an application to substance use.

    Science.gov (United States)

    DeSantis, Stacia M; Bandyopadhyay, Dipankar

    2011-06-30

    Paradigms for substance abuse cue-reactivity research involve pharmacological or stressful stimulation designed to elicit stress and craving responses in cocaine-dependent subjects. It is unclear as to whether stress induced from participation in such studies increases drug-seeking behavior. We propose a 2-state Hidden Markov model to model the number of cocaine abuses per week before and after participation in a stress-and cue-reactivity study. The hypothesized latent state corresponds to 'high' or 'low' use. To account for a preponderance of zeros, we assume a zero-inflated Poisson model for the count data. Transition probabilities depend on the prior week's state, fixed demographic variables, and time-varying covariates. We adopt a Bayesian approach to model fitting, and use the conditional predictive ordinate statistic to demonstrate that the zero-inflated Poisson hidden Markov model outperforms other models for longitudinal count data. Copyright © 2011 John Wiley & Sons, Ltd.

  10. Two dimensional localization of electrons and positrons under high counting rate

    International Nuclear Information System (INIS)

    Barbosa, A.F.; Anjos, J.C.; Sanchez-Hernandez, A.; Pepe, I.M.; Barros, N.

    1997-12-01

    The construction of two wire chambers for the experiment E831 at Fermilab is reported. Each chamber includes three wire planes - one anode and two orthogonal cathodes - in which the wires operate as independent proportional counters. One of the chambers is rotated with respect to the other, so that four position coordinates may be encoded for a charged particle crossing both chambers. Spatial resolution is determined by the wire pitch: 1 mm for cathodes, 2 mm for anodes. 320 electronic channels are involved in the detection system readout. Global counting rates in excess to 10 7 events per second have been measured, while the average electron-positron beam intensity may be as high as 3 x 10 7 events per second. (author)

  11. Variable sampling-time technique for improving count rate performance of scintillation detectors

    International Nuclear Information System (INIS)

    Tanaka, E.; Nohara, N.; Murayama, H.

    1979-01-01

    A new technique is presented to improve the count rate capability of a scintillation spectrometer or a position sensitive detector with minimum loss of resolution. The technique is based on the combination of pulse shortening and selective integration in which the integration period is not fixed but shortened by the arrival of the following pulse. Theoretical analysis of the degradation of the statiscal component of resolution is made for the proposed system with delay line pulse shortening, and the factor of resolution loss is formulated as a function of the input pulse rate. A new method is also presented for determining the statistical component of resolution separately from the non-statistical system resolution. Preliminary experiments with a NaI(Tl) detector have been carried out, the results of which are consistent with the theoretical prediction. However, due to the non-exponential scintillation decay of the NaI(Tl) crystal, a simple delay line clipping is not satisfactory, and an RC high-pass filter has been added, which results in further degradation of the statistical resolution. (Auth.)

  12. Count Data On Cancer Death In Ohio A Bayesian Analysis

    Directory of Open Access Journals (Sweden)

    Walaa Hamdi

    2015-08-01

    Full Text Available This paper considers statistical modeling of count data on cancer death in Ohio State. We obtained count data on male and female from a website of the Centers for Disease Control and Prevention and used Bayesian analyses to find suitable models which help us to do inferences and predictions for next year. To assist us in selecting appropriate models we use criteria such as the DIC. In this paper we analyze the data to spatial longitudinal so we can capture possible correlations. Using our analyses we make predictions of the numbers of people who will die with cancer in a future year in Ohio State.

  13. Gross alpha/beta analyses in water by liquid scintillation counting

    International Nuclear Information System (INIS)

    Wong, C.T.; Lawrence Livermore National Laboratory, CA; Soliman, V.M.; Perera, S.K.

    2005-01-01

    The standard procedure for analyzing gross alpha and gross beta in water is evaporation of the sample and radioactivity determination of the resultant solids by proportional counting. This technique lacks precision, and lacks sensitivity for samples with high total dissolved solids. Additionally, the analytical results are dependent on the choice of radionuclide calibration standard and the sample matrix. Direct analysis by liquid scintillation counting has the advantages of high counting efficiencies and minimal sample preparation time. However, due to the small sample aliquants used for analysis, long count times are necessary to reach required detection limits. The procedure proposed consists of evaporating a sample aliquant to dryness, dissolving the resultant solids in a small volume of dilute acid, followed by liquid scintillation counting to determine radioactivity. This procedure can handle sample aliquants containing up to 500 mg of dissolved solids. Various acids, scintillation cocktail mixtures, instrument discriminator settings, and regions of interest (ROI) were evaluated to determine optimum counting conditions. Precision is improved and matrix effects are reduced as compared to proportional counting. Tests indicate that this is a viable alternative to proportional counting for gross alpha and gross beta analyses of water samples. (author)

  14. Counting carbohydrates

    Science.gov (United States)

    Carb counting; Carbohydrate-controlled diet; Diabetic diet; Diabetes-counting carbohydrates ... Many foods contain carbohydrates (carbs), including: Fruit and fruit juice Cereal, bread, pasta, and rice Milk and milk products, soy milk Beans, legumes, ...

  15. High cumulants of conserved charges and their statistical uncertainties

    Science.gov (United States)

    Li-Zhu, Chen; Ye-Yin, Zhao; Xue, Pan; Zhi-Ming, Li; Yuan-Fang, Wu

    2017-10-01

    We study the influence of measured high cumulants of conserved charges on their associated statistical uncertainties in relativistic heavy-ion collisions. With a given number of events, the measured cumulants randomly fluctuate with an approximately normal distribution, while the estimated statistical uncertainties are found to be correlated with corresponding values of the obtained cumulants. Generally, with a given number of events, the larger the cumulants we measure, the larger the statistical uncertainties that are estimated. The error-weighted averaged cumulants are dependent on statistics. Despite this effect, however, it is found that the three sigma rule of thumb is still applicable when the statistics are above one million. Supported by NSFC (11405088, 11521064, 11647093), Major State Basic Research Development Program of China (2014CB845402) and Ministry of Science and Technology (MoST) (2016YFE0104800)

  16. Analysis of electroperforated materials using the quadrat counts method

    Energy Technology Data Exchange (ETDEWEB)

    Miranda, E; Garzon, C; Garcia-Garcia, J [Departament d' Enginyeria Electronica, Universitat Autonoma de Barcelona, 08193 Bellaterra, Barcelona (Spain); MartInez-Cisneros, C; Alonso, J, E-mail: enrique.miranda@uab.cat [Departament de Quimica AnalItica, Universitat Autonoma de Barcelona, 08193 Bellaterra, Barcelona (Spain)

    2011-06-23

    The electroperforation distribution in thin porous materials is investigated using the quadrat counts method (QCM), a classical statistical technique aimed to evaluate the deviation from complete spatial randomness (CSR). Perforations are created by means of electrical discharges generated by needle-like tungsten electrodes. The objective of perforating a thin porous material is to enhance its air permeability, a critical issue in many industrial applications involving paper, plastics, textiles, etc. Using image analysis techniques and specialized statistical software it is shown that the perforation locations follow, beyond a certain length scale, a homogeneous 2D Poisson distribution.

  17. Repeatability of differential goat bulk milk culture and associations with somatic cell count, total bacterial count, and standard plate count

    OpenAIRE

    Koop, G.; Dik, N.; Nielen, M.; Lipman, L.J.A.

    2010-01-01

    The aims of this study were to assess how different bacterial groups in bulk milk are related to bulk milk somatic cell count (SCC), bulk milk total bacterial count (TBC), and bulk milk standard plate count (SPC) and to measure the repeatability of bulk milk culturing. On 53 Dutch dairy goat farms, 3 bulk milk samples were collected at intervals of 2 wk. The samples were cultured for SPC, coliform count, and staphylococcal count and for the presence of Staphylococcus aureus. Furthermore, SCC ...

  18. Standardization of I-125 solution by extrapolation of an efficiency wave obtained by coincidence X-(X-γ) counting method

    International Nuclear Information System (INIS)

    Iwahara, A.

    1989-01-01

    The activity concentration of 125 I was determined by X-(X-α) coincidence counting method and efficiency extrapolation curve. The measurement system consists of 2 thin NaI(T1) scintillation detectors which are horizontally movable on a track. The efficiency curve is obtained by symmetricaly changing the distance between the source and the detectors and the activity is determined by applying a linear efficiency extrapolation curve. All sum-coincidence events are included between 10 and 100 KeV window counting and the main source of uncertainty is coming from poor counting statistic around zero efficiency. The consistence of results with other methods shows that this technique can be applied to photon cascade emitters and are not discriminating by the detectors. It has been also determined the 35,5 KeV gamma-ray emission probability of 125 I by using a Gamma-X type high purity germanium detector. (author) [pt

  19. CERNDxCTA counting mode chip

    International Nuclear Information System (INIS)

    Moraes, D.; Kaplon, J.; Nygard, E.

    2008-01-01

    This ASIC is a counting mode front-end electronic optimized for the readout of CdZnTe/CdTe and silicon sensors, for possible use in applications where the flux of ionizing radiation is high. The chip is implemented in 0.25 μm CMOS technology. The circuit comprises 128 channels equipped with a transimpedance amplifier followed by a gain shaper stage with 21 ns peaking time, two discriminators and two 18-bit counters. The channel architecture is optimized for the detector characteristics in order to achieve the best energy resolution at counting rates of up to 5 M counts/second. The amplifier shows a linear sensitivity of 118 mV/fC and an equivalent noise charge of about 711 e - , for a detector capacitance of 5 pF. Complete evaluation of the circuit is presented using electronic pulses and pixel detectors

  20. Nonextensive statistical mechanics and high energy physics

    Directory of Open Access Journals (Sweden)

    Tsallis Constantino

    2014-04-01

    Full Text Available The use of the celebrated Boltzmann-Gibbs entropy and statistical mechanics is justified for ergodic-like systems. In contrast, complex systems typically require more powerful theories. We will provide a brief introduction to nonadditive entropies (characterized by indices like q, which, in the q → 1 limit, recovers the standard Boltzmann-Gibbs entropy and associated nonextensive statistical mechanics. We then present somerecent applications to systems such as high-energy collisions, black holes and others. In addition to that, we clarify and illustrate the neat distinction that exists between Lévy distributions and q-exponential ones, a point which occasionally causes some confusion in the literature, very particularly in the LHC literature

  1. Statistical estimation Monte Carlo for unreliability evaluation of highly reliable system

    International Nuclear Information System (INIS)

    Xiao Gang; Su Guanghui; Jia Dounan; Li Tianduo

    2000-01-01

    Based on analog Monte Carlo simulation, statistical Monte Carlo methods for unreliable evaluation of highly reliable system are constructed, including direct statistical estimation Monte Carlo method and weighted statistical estimation Monte Carlo method. The basal element is given, and the statistical estimation Monte Carlo estimators are derived. Direct Monte Carlo simulation method, bounding-sampling method, forced transitions Monte Carlo method, direct statistical estimation Monte Carlo and weighted statistical estimation Monte Carlo are used to evaluate unreliability of a same system. By comparing, weighted statistical estimation Monte Carlo estimator has smallest variance, and has highest calculating efficiency

  2. Bias expansion of spatial statistics and approximation of differenced ...

    Indian Academy of Sciences (India)

    Investigations of spatial statistics, computed from lattice data in the plane, can lead to a special lattice point counting problem. The statistical goal is to expand the asymptotic expectation or large-sample bias of certain spatial covariance estimators, where this bias typically depends on the shape of a spatial sampling region.

  3. A Review of Multivariate Distributions for Count Data Derived from the Poisson Distribution.

    Science.gov (United States)

    Inouye, David; Yang, Eunho; Allen, Genevera; Ravikumar, Pradeep

    2017-01-01

    The Poisson distribution has been widely studied and used for modeling univariate count-valued data. Multivariate generalizations of the Poisson distribution that permit dependencies, however, have been far less popular. Yet, real-world high-dimensional count-valued data found in word counts, genomics, and crime statistics, for example, exhibit rich dependencies, and motivate the need for multivariate distributions that can appropriately model this data. We review multivariate distributions derived from the univariate Poisson, categorizing these models into three main classes: 1) where the marginal distributions are Poisson, 2) where the joint distribution is a mixture of independent multivariate Poisson distributions, and 3) where the node-conditional distributions are derived from the Poisson. We discuss the development of multiple instances of these classes and compare the models in terms of interpretability and theory. Then, we empirically compare multiple models from each class on three real-world datasets that have varying data characteristics from different domains, namely traffic accident data, biological next generation sequencing data, and text data. These empirical experiments develop intuition about the comparative advantages and disadvantages of each class of multivariate distribution that was derived from the Poisson. Finally, we suggest new research directions as explored in the subsequent discussion section.

  4. Evaluation of clustering statistics with N-body simulations

    International Nuclear Information System (INIS)

    Quinn, T.R.

    1986-01-01

    Two series of N-body simulations are used to determine the effectiveness of various clustering statistics in revealing initial conditions from evolved models. All the simulations contained 16384 particles and were integrated with the PPPM code. One series is a family of models with power at only one wavelength. The family contains five models with the wavelength of the power separated by factors of √2. The second series is a family of all equal power combinations of two wavelengths taken from the first series. The clustering statistics examined are the two point correlation function, the multiplicity function, the nearest neighbor distribution, the void probability distribution, the distribution of counts in cells, and the peculiar velocity distribution. It is found that the covariance function, the nearest neighbor distribution, and the void probability distribution are relatively insensitive to the initial conditions. The distribution of counts in cells show a little more sensitivity, but the multiplicity function is the best of the statistics considered for revealing the initial conditions

  5. The negative binomial distribution as a model for external corrosion defect counts in buried pipelines

    International Nuclear Information System (INIS)

    Valor, Alma; Alfonso, Lester; Caleyo, Francisco; Vidal, Julio; Perez-Baruch, Eloy; Hallen, José M.

    2015-01-01

    Highlights: • Observed external-corrosion defects in underground pipelines revealed a tendency to cluster. • The Poisson distribution is unable to fit extensive count data for these type of defects. • In contrast, the negative binomial distribution provides a suitable count model for them. • Two spatial stochastic processes lead to the negative binomial distribution for defect counts. • They are the Gamma-Poisson mixed process and the compound Poisson process. • A Rogeŕs process also arises as a plausible temporal stochastic process leading to corrosion defect clustering and to negative binomially distributed defect counts. - Abstract: The spatial distribution of external corrosion defects in buried pipelines is usually described as a Poisson process, which leads to corrosion defects being randomly distributed along the pipeline. However, in real operating conditions, the spatial distribution of defects considerably departs from Poisson statistics due to the aggregation of defects in groups or clusters. In this work, the statistical analysis of real corrosion data from underground pipelines operating in southern Mexico leads to conclude that the negative binomial distribution provides a better description for defect counts. The origin of this distribution from several processes is discussed. The analysed processes are: mixed Gamma-Poisson, compound Poisson and Roger’s processes. The physical reasons behind them are discussed for the specific case of soil corrosion.

  6. Correction of the counting up number by dead time in detector systems for radiograph images

    International Nuclear Information System (INIS)

    Cerdeira E, A.; Cicuttin, A.; Cerdeira, A.; Estrada, M.; Luca, A. de

    2002-01-01

    The effect of the dead time in a detection system by counting up of particles and the contribution of this error in the final image resolution is analysed. It is given a statistical criteria for the optimization of electronic parameters such as dead time and counting up memory which help in the implementation of these systems with the minimum necessary characteristics which satisfy the resolution requirements. (Author)

  7. Counting and Surveying Homeless Youth: Recommendations from YouthCount 2.0!, a Community-Academic Partnership.

    Science.gov (United States)

    Narendorf, Sarah C; Santa Maria, Diane M; Ha, Yoonsook; Cooper, Jenna; Schieszler, Christine

    2016-12-01

    Communities across the United States are increasing efforts to find and count homeless youth. This paper presents findings and lessons learned from a community/academic partnership to count homeless youth and conduct an in depth research survey focused on the health needs of this population. Over a 4 week recruitment period, 632 youth were counted and 420 surveyed. Methodological successes included an extended counting period, broader inclusion criteria to capture those in unstable housing, use of student volunteers in health training programs, recruiting from magnet events for high risk youth, and partnering with community agencies to disseminate findings. Strategies that did not facilitate recruitment included respondent driven sampling, street canvassing beyond known hotspots, and having community agencies lead data collection. Surveying was successful in gathering data on reasons for homelessness, history in public systems of care, mental health history and needs, sexual risk behaviors, health status, and substance use. Youth were successfully surveyed across housing types including shelters or transitional housing (n = 205), those in unstable housing such as doubled up with friends or acquaintances (n = 75), and those who were literally on the streets or living in a place not meant for human habitation (n = 140). Most youth completed the self-report survey and provided detailed information about risk behaviors. Recommendations to combine research data collection with counting are presented.

  8. Improving the counting efficiency in time-correlated single photon counting experiments by dead-time optimization

    Energy Technology Data Exchange (ETDEWEB)

    Peronio, P.; Acconcia, G.; Rech, I.; Ghioni, M. [Dipartimento di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, Piazza Leonardo da Vinci 32, 20133 Milano (Italy)

    2015-11-15

    Time-Correlated Single Photon Counting (TCSPC) has been long recognized as the most sensitive method for fluorescence lifetime measurements, but often requiring “long” data acquisition times. This drawback is related to the limited counting capability of the TCSPC technique, due to pile-up and counting loss effects. In recent years, multi-module TCSPC systems have been introduced to overcome this issue. Splitting the light into several detectors connected to independent TCSPC modules proportionally increases the counting capability. Of course, multi-module operation also increases the system cost and can cause space and power supply problems. In this paper, we propose an alternative approach based on a new detector and processing electronics designed to reduce the overall system dead time, thus enabling efficient photon collection at high excitation rate. We present a fast active quenching circuit for single-photon avalanche diodes which features a minimum dead time of 12.4 ns. We also introduce a new Time-to-Amplitude Converter (TAC) able to attain extra-short dead time thanks to the combination of a scalable array of monolithically integrated TACs and a sequential router. The fast TAC (F-TAC) makes it possible to operate the system towards the upper limit of detector count rate capability (∼80 Mcps) with reduced pile-up losses, addressing one of the historic criticisms of TCSPC. Preliminary measurements on the F-TAC are presented and discussed.

  9. Data analysis in emission tomography using emission-count posteriors

    International Nuclear Information System (INIS)

    Sitek, Arkadiusz

    2012-01-01

    A novel approach to the analysis of emission tomography data using the posterior probability of the number of emissions per voxel (emission count) conditioned on acquired tomographic data is explored. The posterior is derived from the prior and the Poisson likelihood of the emission-count data by marginalizing voxel activities. Based on emission-count posteriors, examples of Bayesian analysis including estimation and classification tasks in emission tomography are provided. The application of the method to computer simulations of 2D tomography is demonstrated. In particular, the minimum-mean-square-error point estimator of the emission count is demonstrated. The process of finding this estimator can be considered as a tomographic image reconstruction technique since the estimates of the number of emissions per voxel divided by voxel sensitivities and acquisition time are the estimates of the voxel activities. As an example of a classification task, a hypothesis stating that some region of interest (ROI) emitted at least or at most r-times the number of events in some other ROI is tested. The ROIs are specified by the user. The analysis described in this work provides new quantitative statistical measures that can be used in decision making in diagnostic imaging using emission tomography. (paper)

  10. Data analysis in emission tomography using emission-count posteriors

    Science.gov (United States)

    Sitek, Arkadiusz

    2012-11-01

    A novel approach to the analysis of emission tomography data using the posterior probability of the number of emissions per voxel (emission count) conditioned on acquired tomographic data is explored. The posterior is derived from the prior and the Poisson likelihood of the emission-count data by marginalizing voxel activities. Based on emission-count posteriors, examples of Bayesian analysis including estimation and classification tasks in emission tomography are provided. The application of the method to computer simulations of 2D tomography is demonstrated. In particular, the minimum-mean-square-error point estimator of the emission count is demonstrated. The process of finding this estimator can be considered as a tomographic image reconstruction technique since the estimates of the number of emissions per voxel divided by voxel sensitivities and acquisition time are the estimates of the voxel activities. As an example of a classification task, a hypothesis stating that some region of interest (ROI) emitted at least or at most r-times the number of events in some other ROI is tested. The ROIs are specified by the user. The analysis described in this work provides new quantitative statistical measures that can be used in decision making in diagnostic imaging using emission tomography.

  11. Automation system for optical counting of nuclear tracks

    Energy Technology Data Exchange (ETDEWEB)

    Boulyga, S.F.; Boulyga, E.G.; Lomonosova, E.M.; Zhuk, I.V

    1999-06-01

    An automation system consisting of the microscope, video camera and Pentium PC with frame recorder was created. The system provides counting of nuclear tracks on the SSNTD surface with a resolution of 752 x 582 points, determination of the surface area and main axis of the track. The pattern recognition program was developed for operation in Windows 3.1 (or higher) ensuring a convenient interface with the user. In a comparison of the results on automatic track counting with the more accurate hand mode it was shown that the program enables the tracks to be detected even on images with a rather high noise level. It ensures a high accuracy of track counting being comparable with the accuracy of manual counting for densities of tracks in the range of up to 2{center_dot}10{sup 5} tracks/cm{sup 2}. The automatic system was applied in the experimental investigation of uranium and transuranium elements.

  12. Automation system for optical counting of nuclear tracks

    International Nuclear Information System (INIS)

    Boulyga, S.F.; Boulyga, E.G.; Lomonosova, E.M.; Zhuk, I.V.

    1999-01-01

    An automation system consisting of the microscope, video camera and Pentium PC with frame recorder was created. The system provides counting of nuclear tracks on the SSNTD surface with a resolution of 752 x 582 points, determination of the surface area and main axis of the track. The pattern recognition program was developed for operation in Windows 3.1 (or higher) ensuring a convenient interface with the user. In a comparison of the results on automatic track counting with the more accurate hand mode it was shown that the program enables the tracks to be detected even on images with a rather high noise level. It ensures a high accuracy of track counting being comparable with the accuracy of manual counting for densities of tracks in the range of up to 2·10 5 tracks/cm 2 . The automatic system was applied in the experimental investigation of uranium and transuranium elements

  13. Automation system for optical counting of nuclear tracks

    CERN Document Server

    Boulyga, S F; Lomonosova, E M; Zhuk, I V

    1999-01-01

    An automation system consisting of the microscope, video camera and Pentium PC with frame recorder was created. The system provides counting of nuclear tracks on the SSNTD surface with a resolution of 752 x 582 points, determination of the surface area and main axis of the track. The pattern recognition program was developed for operation in Windows 3.1 (or higher) ensuring a convenient interface with the user. In a comparison of the results on automatic track counting with the more accurate hand mode it was shown that the program enables the tracks to be detected even on images with a rather high noise level. It ensures a high accuracy of track counting being comparable with the accuracy of manual counting for densities of tracks in the range of up to 2 centre dot 10 sup 5 tracks/cm sup 2. The automatic system was applied in the experimental investigation of uranium and transuranium elements.

  14. Comparison of plate counts, Petrifilm, dipslides, and adenosine triphosphate bioluminescence for monitoring bacteria in cooling-tower waters.

    Science.gov (United States)

    Mueller, Sherry A; Anderson, James E; Kim, Byung R; Ball, James C

    2009-04-01

    Effective bacterial control in cooling-tower systems requires accurate and timely methods to count bacteria. Plate-count methods are difficult to implement on-site, because they are time- and labor-intensive and require sterile techniques. Several field-applicable methods (dipslides, Petrifilm, and adenosine triphosphate [ATP] bioluminescence) were compared with the plate count for two sample matrices--phosphate-buffered saline solution containing a pure culture of Pseudomonas fluorescens and cooling-tower water containing an undefined mixed bacterial culture. For the pure culture, (1) counts determined on nutrient agar and plate-count agar (PCA) media and expressed as colony-forming units (CFU) per milliliter were equivalent to those on R2A medium (p = 1.0 and p = 1.0, respectively); (2) Petrifilm counts were not significantly different from R2A plate counts (p = 0.99); (3) the dipslide counts were up to 2 log units higher than R2A plate counts, but this discrepancy was not statistically significant (p = 0.06); and (4) a discernable correlation (r2 = 0.67) existed between ATP readings and plate counts. For cooling-tower water samples (n = 62), (1) bacterial counts using R2A medium were higher (but not significant; p = 0.63) than nutrient agar and significantly higher than tryptone-glucose yeast extract (TGE; p = 0.03) and PCA (p < 0.001); (2) Petrifilm counts were significantly lower than nutrient agar or R2A (p = 0.02 and p < 0.001, respectively), but not statistically different from TGE, PCA, and dipslides (p = 0.55, p = 0.69, and p = 0.91, respectively); (3) the dipslide method yielded bacteria counts 1 to 3 log units lower than nutrient agar and R2A (p < 0.001), but was not significantly different from Petrifilm (p = 0.91), PCA (p = 1.00) or TGE (p = 0.07); (4) the differences between dipslides and the other methods became greater with a 6-day incubation time; and (5) the correlation between ATP readings and plate counts varied from system to system, was poor

  15. High resolution micro-CT of low attenuating organic materials using large area photon-counting detector

    International Nuclear Information System (INIS)

    Kumpová, I.; Jandejsek, I.; Jakůbek, J.; Vopálenský, M.; Vavřík, D.; Fíla, T.; Koudelka, P.; Kytýř, D.; Zlámal, P.; Gantar, A.

    2016-01-01

    To overcome certain limitations of contemporary materials used for bone tissue engineering, such as inflammatory response after implantation, a whole new class of materials based on polysaccharide compounds is being developed. Here, nanoparticulate bioactive glass reinforced gelan-gum (GG-BAG) has recently been proposed for the production of bone scaffolds. This material offers promising biocompatibility properties, including bioactivity and biodegradability, with the possibility of producing scaffolds with directly controlled microgeometry. However, to utilize such a scaffold with application-optimized properties, large sets of complex numerical simulations using the real microgeometry of the material have to be carried out during the development process. Because the GG-BAG is a material with intrinsically very low attenuation to X-rays, its radiographical imaging, including tomographical scanning and reconstructions, with resolution required by numerical simulations might be a very challenging task. In this paper, we present a study on X-ray imaging of GG-BAG samples. High-resolution volumetric images of investigated specimens were generated on the basis of micro-CT measurements using a large area flat-panel detector and a large area photon-counting detector. The photon-counting detector was composed of a 010× 1 matrix of Timepix edgeless silicon pixelated detectors with tiling based on overlaying rows (i.e. assembled so that no gap is present between individual rows of detectors). We compare the results from both detectors with the scanning electron microscopy on selected slices in transversal plane. It has been shown that the photon counting detector can provide approx. 3× better resolution of the details in low-attenuating materials than the integrating flat panel detectors. We demonstrate that employment of a large area photon counting detector is a good choice for imaging of low attenuating materials with the resolution sufficient for numerical

  16. High resolution micro-CT of low attenuating organic materials using large area photon-counting detector

    Science.gov (United States)

    Kumpová, I.; Vavřík, D.; Fíla, T.; Koudelka, P.; Jandejsek, I.; Jakůbek, J.; Kytýř, D.; Zlámal, P.; Vopálenský, M.; Gantar, A.

    2016-02-01

    To overcome certain limitations of contemporary materials used for bone tissue engineering, such as inflammatory response after implantation, a whole new class of materials based on polysaccharide compounds is being developed. Here, nanoparticulate bioactive glass reinforced gelan-gum (GG-BAG) has recently been proposed for the production of bone scaffolds. This material offers promising biocompatibility properties, including bioactivity and biodegradability, with the possibility of producing scaffolds with directly controlled microgeometry. However, to utilize such a scaffold with application-optimized properties, large sets of complex numerical simulations using the real microgeometry of the material have to be carried out during the development process. Because the GG-BAG is a material with intrinsically very low attenuation to X-rays, its radiographical imaging, including tomographical scanning and reconstructions, with resolution required by numerical simulations might be a very challenging task. In this paper, we present a study on X-ray imaging of GG-BAG samples. High-resolution volumetric images of investigated specimens were generated on the basis of micro-CT measurements using a large area flat-panel detector and a large area photon-counting detector. The photon-counting detector was composed of a 010× 1 matrix of Timepix edgeless silicon pixelated detectors with tiling based on overlaying rows (i.e. assembled so that no gap is present between individual rows of detectors). We compare the results from both detectors with the scanning electron microscopy on selected slices in transversal plane. It has been shown that the photon counting detector can provide approx. 3× better resolution of the details in low-attenuating materials than the integrating flat panel detectors. We demonstrate that employment of a large area photon counting detector is a good choice for imaging of low attenuating materials with the resolution sufficient for numerical simulations.

  17. Analysis of dental caries using generalized linear and count regression models

    Directory of Open Access Journals (Sweden)

    Javali M. Phil

    2013-11-01

    Full Text Available Generalized linear models (GLM are generalization of linear regression models, which allow fitting regression models to response data in all the sciences especially medical and dental sciences that follow a general exponential family. These are flexible and widely used class of such models that can accommodate response variables. Count data are frequently characterized by overdispersion and excess zeros. Zero-inflated count models provide a parsimonious yet powerful way to model this type of situation. Such models assume that the data are a mixture of two separate data generation processes: one generates only zeros, and the other is either a Poisson or a negative binomial data-generating process. Zero inflated count regression models such as the zero-inflated Poisson (ZIP, zero-inflated negative binomial (ZINB regression models have been used to handle dental caries count data with many zeros. We present an evaluation framework to the suitability of applying the GLM, Poisson, NB, ZIP and ZINB to dental caries data set where the count data may exhibit evidence of many zeros and over-dispersion. Estimation of the model parameters using the method of maximum likelihood is provided. Based on the Vuong test statistic and the goodness of fit measure for dental caries data, the NB and ZINB regression models perform better than other count regression models.

  18. Material screening with HPGe counting station for PandaX experiment

    International Nuclear Information System (INIS)

    Wang, X.; Chen, X.; Fu, C.; Ji, X.; Liu, X.; Xie, P.; Zhang, T.; Mao, Y.; Wang, S.; Wang, H.

    2016-01-01

    A gamma counting station based on high-purity germanium (HPGe) detector was set up for the material screening of the PandaX dark matter experiments in the China Jinping Underground Laboratory. Low background gamma rate of 2.6 counts/min within the energy range of 20 to 2700 keV is achieved due to the well-designed passive shield. The sentivities of the HPGe detetector reach mBq/kg level for isotopes like K, U, Th, and even better for Co and Cs, resulted from the low-background rate and the high relative detection efficiency of 175%. The structure and performance of the counting station are described in this article. Detailed counting results for the radioactivity in materials used by the PandaX dark-matter experiment are presented. The upgrading plan of the counting station is also discussed.

  19. Learning word vector representations based on acoustic counts

    OpenAIRE

    Ribeiro, Sam; Watts, Oliver; Yamagishi, Junichi

    2017-01-01

    This paper presents a simple count-based approach to learning word vector representations by leveraging statistics of cooccurrences between text and speech. This type of representation requires two discrete sequences of units defined across modalities. Two possible methods for the discretization of an acoustic signal are presented, which are then applied to fundamental frequency and energy contours of a transcribed corpus of speech, yielding a sequence of textual objects (e.g. words, syllable...

  20. Simple counting technique for measuring mixtures of two pure beta-emitting radionuclides

    CSIR Research Space (South Africa)

    Van Wyngaardt, WM

    2006-08-01

    Full Text Available on counting statistics. The method is demonstrated experimentally for various mixture combinations of C-14 and Ni-63. It is shown that the accuracy of the technique can be enhanced by improving the determination of the figure-of-merit....

  1. Relationship between salivary flow rates and Candida counts in subjects with xerostomia.

    Science.gov (United States)

    Torres, Sandra R; Peixoto, Camila Bernardo; Caldas, Daniele Manhães; Silva, Eline Barboza; Akiti, Tiyomi; Nucci, Márcio; de Uzeda, Milton

    2002-02-01

    This study evaluated the relationship between salivary flow and Candida colony counts in the saliva of patients with xerostomia. Sialometry and Candida colony-forming unit (CFU) counts were taken from 112 subjects who reported xerostomia in a questionnaire. Chewing-stimulated whole saliva was collected and streaked in Candida plates and counted in 72 hours. Species identification was accomplished under standard methods. There was a significant inverse relationship between salivary flow and Candida CFU counts (P =.007) when subjects with high colony counts were analyzed (cutoff point of 400 or greater CFU/mL). In addition, the median sialometry of men was significantly greater than that of women (P =.003), even after controlling for confounding variables like underlying disease and medications. Sjögren's syndrome was associated with low salivary flow rate (P =.007). There was no relationship between the median Candida CFU counts and gender or age. There was a high frequency (28%) of mixed colonization. Candida albicans was the most frequent species, followed by C parapsilosis, C tropicalis, and C krusei. In subjects with high Candida CFU counts there was an inverse relationship between salivary flow and Candida CFU counts.

  2. KANSAS KIDS COUNT Data Book, 2000: Kansas Children at Risk.

    Science.gov (United States)

    Kansas Action for Children, Inc., Topeka.

    This KIDS COUNT Data Book provides state and county data on the well-being of Kansas' children. The statistical portrait is based on 22 indicators of well-being: (1) births to single teens; (2) children living in poverty; (3) children receiving free school meals; (4) children in families receiving economic assistance; (5) childhood deaths; (6)…

  3. Auto-counting of high density overlapping tracks and neutron spectrum measurement using CR-39 detectors and in-house image analysis program

    International Nuclear Information System (INIS)

    Paul, Sabyasachi; Tripathy, S.P.; Sahoo, G.S.; Joshi, D.S.; Bandyopadhyay, T.

    2014-01-01

    An effort is made in this work to overcome the difficulty of counting highly dense and overlapping tracks in solid polymeric track detectors (SPTD) such as CR-39. A program is developed to automatically count the track density which is found to be faster and more precise compared to other commonly used image analysing software. The results obtained by the present methodology are compared with those obtained using other software. (author)

  4. CD4 cell count recovery in HIV/TB co-infected patients versus TB uninfected HIV patients

    Directory of Open Access Journals (Sweden)

    Wanchu A

    2010-10-01

    Full Text Available Background: There is lack of data comparing the improvement in CD4 count following antitubercular (ATT and antiretroviral therapy (ART in patients presenting with Human Immunodeficiency Virus/Tuberculosis (HIV/TB dual infection compared with CD4 matched cohort of TB uninfected HIV patients initiated on ART. We sought to test the hypothesis; TB additionally contributes to reduction in CD4 count in HIV/TB co-infected patients and this would result in greater improvement in count following treatment compared with CD4 matched TB uninfected individuals. Materials and Methods: In a retrospective cohort study design we studied the change in CD4 cell counts in two groups of patients - those with CD4 cell count >100 cells / mm 3 (Group 1 and <100/mm 3 (Group 2 at presentation. In each group the change in CD4 cell count in dually infected patients following six-month ATT and ART was compared to cohorts of CD4 matched TB uninfected patients initiated on ART. Results: In Group 1 (52 patients dually infected subjects′ CD4 count improved from 150 cells/ mm 3 to 345 cells/mm 3 (P=0.001. In the control TB uninfected patients, the change was from 159 cells/mm 3 to 317 cells/mm 3 (P=0.001. Additional improvement in dually infected patients compared to the control group was not statistically significant (P=0.24. In Group 2 (65 patients dually infected subjects count improved from 49 cells/mm3 to 249 cells/mm 3 (P=0.001 where as in control TB uninfected patients improvement was from 50 cells/ mm 3 to 205 cells/mm 3 (P=0.001, there being statistically significant additional improvement in dually infected subjects (P=0.01. Conclusion: Greater increment in CD4 counts with ATT and ART in dually infected patients suggests that TB additionally influences the reduction of CD4 counts in HIV patients.

  5. Standardization of 241Am by digital coincidence counting, liquid scintillation counting and defined solid angle counting

    International Nuclear Information System (INIS)

    Balpardo, C.; Capoulat, M.E.; Rodrigues, D.; Arenillas, P.

    2010-01-01

    The nuclide 241 Am decays by alpha emission to 237 Np. Most of the decays (84.6%) populate the excited level of 237 Np with energy of 59.54 keV. Digital coincidence counting was applied to standardize a solution of 241 Am by alpha-gamma coincidence counting with efficiency extrapolation. Electronic discrimination was implemented with a pressurized proportional counter and the results were compared with two other independent techniques: Liquid scintillation counting using the logical sum of double coincidences in a TDCR array and defined solid angle counting taking into account activity inhomogeneity in the active deposit. The results show consistency between the three methods within a limit of a 0.3%. An ampoule of this solution will be sent to the International Reference System (SIR) during 2009. Uncertainties were analysed and compared in detail for the three applied methods.

  6. Use of the Beta-Binomial Model for Central Statistical Monitoring of Multicenter Clinical Trials

    OpenAIRE

    Desmet, Lieven; Venet, David; Doffagne, Erik; Timmermans, Catherine; Legrand, Catherine; Burzykowski, Tomasz; Buyse, Marc

    2017-01-01

    As part of central statistical monitoring of multicenter clinical trial data, we propose a procedure based on the beta-binomial distribution for the detection of centers with atypical values for the probability of some event. The procedure makes no assumptions about the typical event proportion and uses the event counts from all centers to derive a reference model. The procedure is shown through simulations to have high sensitivity and high specificity if the contamination rate is small and t...

  7. Positron imaging system with improved count rate and tomographic capability

    International Nuclear Information System (INIS)

    Muehllehner, G.; Buchin, M.P.

    1980-01-01

    Improvements to a positron camera imaging system are described. A pair of Angear-type scintillation cameras serve as the detectors, each camera being positioned on opposite sides of the organ of interest. Pulse shaping circuits reduce the pulse duration below 900 nanoseconds and the integration time below 500 noneseconds, improving the count rate capability and the counting statistics of the system and thus the image quality and processing speed. The invention also provides means for rotating the opposed camera heads about an axis which passes through the organ of interest. The cameras do not use collimators, and are capable of accepting radiation travelling in planes not perpendicular to the scintillation crystals. (LL)

  8. Bayesian model selection techniques as decision support for shaping a statistical analysis plan of a clinical trial: An example from a vertigo phase III study with longitudinal count data as primary endpoint

    Directory of Open Access Journals (Sweden)

    Adrion Christine

    2012-09-01

    Full Text Available Abstract Background A statistical analysis plan (SAP is a critical link between how a clinical trial is conducted and the clinical study report. To secure objective study results, regulatory bodies expect that the SAP will meet requirements in pre-specifying inferential analyses and other important statistical techniques. To write a good SAP for model-based sensitivity and ancillary analyses involves non-trivial decisions on and justification of many aspects of the chosen setting. In particular, trials with longitudinal count data as primary endpoints pose challenges for model choice and model validation. In the random effects setting, frequentist strategies for model assessment and model diagnosis are complex and not easily implemented and have several limitations. Therefore, it is of interest to explore Bayesian alternatives which provide the needed decision support to finalize a SAP. Methods We focus on generalized linear mixed models (GLMMs for the analysis of longitudinal count data. A series of distributions with over- and under-dispersion is considered. Additionally, the structure of the variance components is modified. We perform a simulation study to investigate the discriminatory power of Bayesian tools for model criticism in different scenarios derived from the model setting. We apply the findings to the data from an open clinical trial on vertigo attacks. These data are seen as pilot data for an ongoing phase III trial. To fit GLMMs we use a novel Bayesian computational approach based on integrated nested Laplace approximations (INLAs. The INLA methodology enables the direct computation of leave-one-out predictive distributions. These distributions are crucial for Bayesian model assessment. We evaluate competing GLMMs for longitudinal count data according to the deviance information criterion (DIC or probability integral transform (PIT, and by using proper scoring rules (e.g. the logarithmic score. Results The instruments under study

  9. Benjamin Thompson, Count Rumford Count Rumford on the nature of heat

    CERN Document Server

    Brown, Sanborn C

    1967-01-01

    Men of Physics: Benjamin Thompson - Count Rumford: Count Rumford on the Nature of Heat covers the significant contributions of Count Rumford in the fields of physics. Count Rumford was born with the name Benjamin Thompson on March 23, 1753, in Woburn, Massachusetts. This book is composed of two parts encompassing 11 chapters, and begins with a presentation of Benjamin Thompson's biography and his interest in physics, particularly as an advocate of an """"anti-caloric"""" theory of heat. The subsequent chapters are devoted to his many discoveries that profoundly affected the physical thought

  10. Statistical and heuristic image noise extraction (SHINE): a new method for processing Poisson noise in scintigraphic images

    International Nuclear Information System (INIS)

    Hannequin, Pascal; Mas, Jacky

    2002-01-01

    Poisson noise is one of the factors degrading scintigraphic images, especially at low count level, due to the statistical nature of photon detection. We have developed an original procedure, named statistical and heuristic image noise extraction (SHINE), to reduce the Poisson noise contained in the scintigraphic images, preserving the resolution, the contrast and the texture. The SHINE procedure consists in dividing the image into 4 x 4 blocks and performing a correspondence analysis on these blocks. Each block is then reconstructed using its own significant factors which are selected using an original statistical variance test. The SHINE procedure has been validated using a line numerical phantom and a hot spots and cold spots real phantom. The reference images are the noise-free simulated images for the numerical phantom and an extremely high counts image for the real phantom. The SHINE procedure has then been applied to the Jaszczak phantom and clinical data including planar bone scintigraphy, planar Sestamibi scintigraphy and Tl-201 myocardial SPECT. The SHINE procedure reduces the mean normalized error between the noisy images and the corresponding reference images. This reduction is constant and does not change with the count level. The SNR in a SHINE processed image is close to that of the corresponding raw image with twice the number of counts. The visual results with the Jaszczak phantom SPECT have shown that SHINE preserves the contrast and the resolution of the slices well. Clinical examples have shown no visual difference between the SHINE images and the corresponding raw images obtained with twice the acquisition duration. SHINE is an entirely automatic procedure which enables halving the acquisition time or the injected dose in scintigraphic acquisitions. It can be applied to all scintigraphic images, including PET data, and to all low-count photon images

  11. High count-rate study of two TES x-ray microcalorimeters with different transition temperatures

    Science.gov (United States)

    Lee, Sang-Jun; Adams, Joseph S.; Bandler, Simon R.; Betancourt-Martinez, Gabriele L.; Chervenak, James A.; Eckart, Megan E.; Finkbeiner, Fred M.; Kelley, Richard L.; Kilbourne, Caroline A.; Porter, Frederick S.; Sadleir, John E.; Smith, Stephen J.; Wassell, Edward J.

    2017-10-01

    We have developed transition-edge sensor (TES) microcalorimeter arrays with high count-rate capability and high energy resolution to carry out x-ray imaging spectroscopy observations of various astronomical sources and the Sun. We have studied the dependence of the energy resolution and throughput (fraction of processed pulses) on the count rate for such microcalorimeters with two different transition temperatures (T c). Devices with both transition temperatures were fabricated within a single microcalorimeter array directly on top of a solid substrate where the thermal conductance of the microcalorimeter is dependent upon the thermal boundary resistance between the TES sensor and the dielectric substrate beneath. Because the thermal boundary resistance is highly temperature dependent, the two types of device with different T cs had very different thermal decay times, approximately one order of magnitude different. In our earlier report, we achieved energy resolutions of 1.6 and 2.3 eV at 6 keV from lower and higher T c devices, respectively, using a standard analysis method based on optimal filtering in the low flux limit. We have now measured the same devices at elevated x-ray fluxes ranging from 50 Hz to 1000 Hz per pixel. In the high flux limit, however, the standard optimal filtering scheme nearly breaks down because of x-ray pile-up. To achieve the highest possible energy resolution for a fixed throughput, we have developed an analysis scheme based on the so-called event grade method. Using the new analysis scheme, we achieved 5.0 eV FWHM with 96% throughput for 6 keV x-rays of 1025 Hz per pixel with the higher T c (faster) device, and 5.8 eV FWHM with 97% throughput with the lower T c (slower) device at 722 Hz.

  12. Counting cormorants

    DEFF Research Database (Denmark)

    Bregnballe, Thomas; Carss, David N; Lorentsen, Svein-Håkon

    2013-01-01

    This chapter focuses on Cormorant population counts for both summer (i.e. breeding) and winter (i.e. migration, winter roosts) seasons. It also explains differences in the data collected from undertaking ‘day’ versus ‘roost’ counts, gives some definitions of the term ‘numbers’, and presents two...

  13. Cellular and soluble components decrease the viable pathogen counts in milk from dairy cows with subclinical mastitis.

    Science.gov (United States)

    Koshiishi, Tomoko; Watanabe, Masako; Miyake, Hajime; Hisaeda, Keiichi; Isobe, Naoki

    2017-08-10

    The present study was undertaken to clarify the factors that reduce the viable pathogen count in milk collected from the udders of subclinical mastitic cows during preservation. Milk was centrifuged to divide somatic cells (cellular components, precipitates) and antimicrobial peptides (soluble components, supernatants without fat layer); each fraction was cultured with bacteria, and the number of viable bacteria was assessed prior to and after culture. In 28.8% of milk samples, we noted no viable bacteria immediately after collection; this value increased significantly after a 5-hr incubation of milk with cellular components but not with soluble components (48.1 and 28.8%, respectively). After culture with cellular components, the numbers of bacteria (excluding Staphylococcus aureus and Streptococcus uberis) and yeast decreased dramatically, although the differences were not statistically significant. After cultivation with soluble components, only yeasts showed a tendency toward decreased mean viability, whereas the mean bacterial counts of S. uberis and T. pyogenes tended to increase after 5-hr preservation with soluble components. These results suggest that most pathogens in high somatic cell count (SCC) milk decreased during preservation at 15 to 25°C, due to both the cellular components and antimicrobial components in the milk. Particularly, the cellular components more potently reduced bacterial counts during preservation.

  14. Statistical Analysis for High-Dimensional Data : The Abel Symposium 2014

    CERN Document Server

    Bühlmann, Peter; Glad, Ingrid; Langaas, Mette; Richardson, Sylvia; Vannucci, Marina

    2016-01-01

    This book features research contributions from The Abel Symposium on Statistical Analysis for High Dimensional Data, held in Nyvågar, Lofoten, Norway, in May 2014. The focus of the symposium was on statistical and machine learning methodologies specifically developed for inference in “big data” situations, with particular reference to genomic applications. The contributors, who are among the most prominent researchers on the theory of statistics for high dimensional inference, present new theories and methods, as well as challenging applications and computational solutions. Specific themes include, among others, variable selection and screening, penalised regression, sparsity, thresholding, low dimensional structures, computational challenges, non-convex situations, learning graphical models, sparse covariance and precision matrices, semi- and non-parametric formulations, multiple testing, classification, factor models, clustering, and preselection. Highlighting cutting-edge research and casting light on...

  15. Analysis of statistical misconception in terms of statistical reasoning

    Science.gov (United States)

    Maryati, I.; Priatna, N.

    2018-05-01

    Reasoning skill is needed for everyone to face globalization era, because every person have to be able to manage and use information from all over the world which can be obtained easily. Statistical reasoning skill is the ability to collect, group, process, interpret, and draw conclusion of information. Developing this skill can be done through various levels of education. However, the skill is low because many people assume that statistics is just the ability to count and using formulas and so do students. Students still have negative attitude toward course which is related to research. The purpose of this research is analyzing students’ misconception in descriptive statistic course toward the statistical reasoning skill. The observation was done by analyzing the misconception test result and statistical reasoning skill test; observing the students’ misconception effect toward statistical reasoning skill. The sample of this research was 32 students of math education department who had taken descriptive statistic course. The mean value of misconception test was 49,7 and standard deviation was 10,6 whereas the mean value of statistical reasoning skill test was 51,8 and standard deviation was 8,5. If the minimal value is 65 to state the standard achievement of a course competence, students’ mean value is lower than the standard competence. The result of students’ misconception study emphasized on which sub discussion that should be considered. Based on the assessment result, it was found that students’ misconception happen on this: 1) writing mathematical sentence and symbol well, 2) understanding basic definitions, 3) determining concept that will be used in solving problem. In statistical reasoning skill, the assessment was done to measure reasoning from: 1) data, 2) representation, 3) statistic format, 4) probability, 5) sample, and 6) association.

  16. Quantitative annular dark field scanning transmission electron microscopy for nanoparticle atom-counting: What are the limits?

    International Nuclear Information System (INIS)

    De Backer, A; De Wael, A; Gonnissen, J; Martinez, G T; Béché, A; Van Aert, S; MacArthur, K E; Jones, L; Nellist, P D

    2015-01-01

    Quantitative atomic resolution annular dark field scanning transmission electron microscopy (ADF STEM) has become a powerful technique for nanoparticle atom-counting. However, a lot of nanoparticles provide a severe characterisation challenge because of their limited size and beam sensitivity. Therefore, quantitative ADF STEM may greatly benefit from statistical detection theory in order to optimise the instrumental microscope settings such that the incoming electron dose can be kept as low as possible whilst still retaining single-atom precision. The principles of detection theory are used to quantify the probability of error for atom-counting. This enables us to decide between different image performance measures and to optimise the experimental detector settings for atom-counting in ADF STEM in an objective manner. To demonstrate this, ADF STEM imaging of an industrial catalyst has been conducted using the near-optimal detector settings. For this experiment, we discussed the limits for atomcounting diagnosed by combining a thorough statistical method and detailed image simulations. (paper)

  17. Seed counting system evaluation using arduino microcontroller

    Directory of Open Access Journals (Sweden)

    Paulo Fernando Escobar Paim

    2018-01-01

    Full Text Available The development of automated systems has been highlighted in the most diverse productive sectors, among them, the agricultural sector. These systems aim to optimize activities by increasing operational efficiency and quality of work. In this sense, the present work has the objective of evaluating a prototype developed for seed count in laboratory, using Arduino microcontroller. The prototype of the system for seed counting was built using a dosing mechanism commonly used in seeders, electric motor, Arduino Uno, light dependent resistor and light emitting diode. To test the prototype, a completely randomized design (CRD was used in a two-factorial scheme composed of three groups defined according to the number of seeds (500, 1000 and 1500 seeds tested, three speeds of the dosing disc that allowed the distribution in 17, 21 and 32 seeds per second, with 40 repetitions evaluating the seed counting prototype performance in different speeds. The prototype of the bench counter showed a moderate variability of seed number of counted within the nine tests and a high precision in the seed count on the distribution speeds of 17 and 21 seeds per second (s-1 up to 1500 seeds tested. Therefore, based on the observed results, the developed prototype presents itself as an excellent tool for counting seeds in laboratory.

  18. Relationship between γ detection dead-time and count correction factor

    International Nuclear Information System (INIS)

    Wu Huailong; Zhang Jianhua; Chu Chengsheng; Hu Guangchun; Zhang Changfan; Hu Gen; Gong Jian; Tian Dongfeng

    2015-01-01

    The relationship between dead-time and count correction factor was investigated by using interference source for purpose of high γ activity measurement. The count rates maintain several 10 s"-"l with γ energy of 0.3-1.3 MeV for 10"4-10"5 Bq radioactive source. It is proved that the relationship between count loss and dead-time is unconcerned at various energy and various count intensities. The same correction formula can be used for any nuclide measurement. (authors)

  19. Counting the peaks in the excitation function for precompound processes

    International Nuclear Information System (INIS)

    Bonetti, R.; Hussein, M.S.; Mello, P.A.

    1983-01-01

    The ''counting of maxima'' method of Brink and Stephen, conventionally used for the extraction of the correlation width of statistical (compound nucleus) reactions, is generalized to include precompound processes as well. It is found that this method supplies an important independent check of the results obtained from autocorrelation studies. An application is made to the reaction 25 Mg( 3 He, p)

  20. Characteristics of Febrile Patients with Normal White Blood Cell Counts and High C-Reactive Protein Levels in an Emergency Department

    Directory of Open Access Journals (Sweden)

    Kuan-Ting Liu

    2008-05-01

    Full Text Available Fever is one of the more common chief complaints of patients who visit emergency departments (ED. Many febrile patients have markedly elevated C-reactive protein (CRP levels and normal white blood cell (WBC counts. Most of these patients have bacterial infection and no previous underlying disease of impaired WBC functioning. We reviewed patients who visited our ED between November 2003 and July 2004. The WBC count and CRP level of patients over 18 years of age who visited the ED because of or with fever were recorded. Patients who had normal WBC count (4,000–10,000/mL and high CRP level (> 100 mg/L were included. The data, including gender, age and length of hospital stay, were reviewed. Underlying diseases, diagnosis of the febrile disease and final condition were recorded according to the chart. Within the study period, 54,078 patients visited our ED. Of 5,628 febrile adults, 214 (3.8% had elevated CRP level and normal WBC count. The major cause of febrility was infection (82.24%. Most of these patients were admitted (92.99%. There were 32 patients with malignant neoplasm, nine with liver cirrhosis, 66 with diabetes mellitus and 11 with uremia. There were no significant differences in age and gender between patients with and those without neoplasm. However, a higher inhospital mortality rate and other causes of febrility were noted in patients with neoplasm. It was not rare in febrile patients who visited the ED to have a high CRP level but normal WBC count. These patients did not necessarily have an underlying malignant neoplasm or hematologic illness. Factors other than malignant neoplasm or hematologic illness may be associated with the WBC response, and CRP may be a better indicator of infection under such conditions.

  1. Alberta oil and gas industry: Annual statistics for 1997. Statistical series number 98-17

    International Nuclear Information System (INIS)

    1998-01-01

    This document presents annual statistics for the Alberta oil and gas industry for the supply and disposition of crude oil and equivalent; gas; ethane; propane; butanes; NGL mixes; and sulfur. Figures are given for deliveries and prices for the current year and also historically (1987--1996). Figures are also provided for the number of wells drilled during the year, meters drilled, and the annual well count

  2. A matrix-inversion method for gamma-source mapping from gamma-count data - 59082

    International Nuclear Information System (INIS)

    Bull, Richard K.; Adsley, Ian; Burgess, Claire

    2012-01-01

    Gamma ray counting is often used to survey the distribution of active waste material in various locations. Ideally the output from such surveys would be a map of the activity of the waste. In this paper a simple matrix-inversion method is presented. This allows an array of gamma-count data to be converted to an array of source activities. For each survey area the response matrix is computed using the gamma-shielding code Microshield [1]. This matrix links the activity array to the count array. The activity array is then obtained via matrix inversion. The method was tested on artificially-created arrays of count-data onto which statistical noise had been added. The method was able to reproduce, quite faithfully, the original activity distribution used to generate the dataset. The method has been applied to a number of practical cases, including the distribution of activated objects in a hot cell and to activated Nimonic springs amongst fuel-element debris in vaults at a nuclear plant. (authors)

  3. Development of Fast High-Resolution Muon Drift-Tube Detectors for High Counting Rates

    CERN Document Server

    INSPIRE-00287945; Dubbert, J.; Horvat, S.; Kortner, O.; Kroha, H.; Legger, F.; Richter, R.; Adomeit, S.; Biebel, O.; Engl, A.; Hertenberger, R.; Rauscher, F.; Zibell, A.

    2011-01-01

    Pressurized drift-tube chambers are e?cient detectors for high-precision tracking over large areas. The Monitored Drift-Tube (MDT) chambers of the muon spectrometer of the ATLAS detector at the Large Hadron Collider (LHC) reach a spatial resolution of 35 micons and almost 100% tracking e?ciency with 6 layers of 30 mm diameter drift tubes operated with Ar:CO2 (93:7) gas mixture at 3 bar and a gas gain of 20000. The ATLAS MDT chambers are designed to cope with background counting rates due to neutrons and gamma-rays of up to about 300 kHz per tube which will be exceeded for LHC luminosities larger than the design value of 10-34 per square cm and second. Decreasing the drift-tube diameter to 15 mm while keeping the other parameters, including the gas gain, unchanged reduces the maximum drift time from about 700 ns to 200 ns and the drift-tube occupancy by a factor of 7. New drift-tube chambers for the endcap regions of the ATLAS muon spectrometer have been designed. A prototype chamber consisting of 12 times 8 l...

  4. Repeatability of differential goat bulk milk culture and associations with somatic cell count, total bacterial count, and standard plate count

    NARCIS (Netherlands)

    Koop, G.; Dik, N.; Nielen, M.; Lipman, L.J.A.

    2010-01-01

    The aims of this study were to assess how different bacterial groups in bulk milk are related to bulk milk somatic cell count (SCC), bulk milk total bacterial count (TBC), and bulk milk standard plate count (SPC) and to measure the repeatability of bulk milk culturing. On 53 Dutch dairy goat farms,

  5. Statistics for experimentalists

    CERN Document Server

    Cooper, B E

    2014-01-01

    Statistics for Experimentalists aims to provide experimental scientists with a working knowledge of statistical methods and search approaches to the analysis of data. The book first elaborates on probability and continuous probability distributions. Discussions focus on properties of continuous random variables and normal variables, independence of two random variables, central moments of a continuous distribution, prediction from a normal distribution, binomial probabilities, and multiplication of probabilities and independence. The text then examines estimation and tests of significance. Topics include estimators and estimates, expected values, minimum variance linear unbiased estimators, sufficient estimators, methods of maximum likelihood and least squares, and the test of significance method. The manuscript ponders on distribution-free tests, Poisson process and counting problems, correlation and function fitting, balanced incomplete randomized block designs and the analysis of covariance, and experiment...

  6. Signal shaping and tail cancellation for gas proportional detectors at high counting rates

    International Nuclear Information System (INIS)

    Boie, R.A.; Hrisoho, A.T.; Rehak, P.

    1982-01-01

    A low noise, wide bandwidth preamplifier and signal processing filter were developed for high counting rate proportional counters. The filter consists of a seven pole Gaussian integrator with symmetrical weighting function and continuously variable shaping time, tausub(s), of 8-50 ns (fwhm) preceded by a second order pole/zero circuit which cancels the long (1/t) tails of the chamber signals. The preamplifier is an optimized common base input design with 2 ns rise time and an equivalent noise input charge < 2000 r.m.s. electrons, when connected to a chamber with 10 pF capacitance and at a filtering time, tausub(s), of 10 ns. (orig.)

  7. Big drop. [Decline in sperm counts from environmental pollution

    Energy Technology Data Exchange (ETDEWEB)

    Castleman, M.

    Statisticians at the University of Copenhagen have published a review of sperm-count studies carried out from 1938-1990. After controlling for such things as counting techniques and sexual activity, they found that average sperm counts have decreased 42% over the last 50 years. After eliminating possible causes such as heat and differences in counting techniques, the only plausible cause left was environmental pollution. This finding correlates with a study performed in 1979 which found abnormally high concentrations of such toxic pollutants as DDT and PCBs in semen. The Copenhagen study suggests that even at low levels of exposure, environmental pollution is having biological consequences.

  8. Do we need statistics when we have linguistics?

    Directory of Open Access Journals (Sweden)

    Cantos Gómez Pascual

    2002-01-01

    Full Text Available Statistics is known to be a quantitative approach to research. However, most of the research done in the fields of language and linguistics is of a different kind, namely qualitative. Succinctly, qualitative analysis differs from quantitative analysis is that in the former no attempt is made to assign frequencies, percentages and the like, to the linguistic features found or identified in the data. In quantitative research, linguistic features are classified and counted, and even more complex statistical models are constructed in order to explain these observed facts. In qualitative research, however, we use the data only for identifying and describing features of language usage and for providing real occurrences/examples of particular phenomena. In this paper, we shall try to show how quantitative methods and statistical techniques can supplement qualitative analyses of language. We shall attempt to present some mathematical and statistical properties of natural languages, and introduce some of the quantitative methods which are of the most value in working empirically with texts and corpora, illustrating the various issues with numerous examples and moving from the most basic descriptive techniques (frequency counts and percentages to decision-taking techniques (chi-square and z-score and to more sophisticated statistical language models (Type-Token/Lemma-Token/Lemma-Type formulae, cluster analysis and discriminant function analysis.

  9. Multiple-Event, Single-Photon Counting Imaging Sensor

    Science.gov (United States)

    Zheng, Xinyu; Cunningham, Thomas J.; Sun, Chao; Wang, Kang L.

    2011-01-01

    The single-photon counting imaging sensor is typically an array of silicon Geiger-mode avalanche photodiodes that are monolithically integrated with CMOS (complementary metal oxide semiconductor) readout, signal processing, and addressing circuits located in each pixel and the peripheral area of the chip. The major problem is its single-event method for photon count number registration. A single-event single-photon counting imaging array only allows registration of up to one photon count in each of its pixels during a frame time, i.e., the interval between two successive pixel reset operations. Since the frame time can t be too short, this will lead to very low dynamic range and make the sensor merely useful for very low flux environments. The second problem of the prior technique is a limited fill factor resulting from consumption of chip area by the monolithically integrated CMOS readout in pixels. The resulting low photon collection efficiency will substantially ruin any benefit gained from the very sensitive single-photon counting detection. The single-photon counting imaging sensor developed in this work has a novel multiple-event architecture, which allows each of its pixels to register as more than one million (or more) photon-counting events during a frame time. Because of a consequently boosted dynamic range, the imaging array of the invention is capable of performing single-photon counting under ultra-low light through high-flux environments. On the other hand, since the multiple-event architecture is implemented in a hybrid structure, back-illumination and close-to-unity fill factor can be realized, and maximized quantum efficiency can also be achieved in the detector array.

  10. CERN_DxCTA counting mode chip

    CERN Document Server

    Moraes, D; Nygård, E

    2008-01-01

    This ASIC is a counting mode front-end electronic optimized for the readout of CdZnTe/CdTe and silicon sensors, for possible use in applications where the flux of ionizing radiation is high. The chip is implemented in 0.25 μm CMOS technology. The circuit comprises 128 channels equipped with a transimpedance amplifier followed by a gain shaper stage with 21 ns peaking time, two discriminators and two 18-bit counters. The channel architecture is optimized for the detector characteristics in order to achieve the best energy resolution at counting rates of up to 5 M counts/second. The amplifier shows a linear sensitivity of 118 mV/fC and an equivalent noise charge of about 711 e−, for a detector capacitance of 5 pF. Complete evaluation of the circuit is presented using electronic pulses and pixel detectors.

  11. Standardization of {sup 241}Am by digital coincidence counting, liquid scintillation counting and defined solid angle counting

    Energy Technology Data Exchange (ETDEWEB)

    Balpardo, C., E-mail: balpardo@cae.cnea.gov.a [Laboratorio de Metrologia de Radioisotopos, CNEA, Buenos Aires (Argentina); Capoulat, M.E.; Rodrigues, D.; Arenillas, P. [Laboratorio de Metrologia de Radioisotopos, CNEA, Buenos Aires (Argentina)

    2010-07-15

    The nuclide {sup 241}Am decays by alpha emission to {sup 237}Np. Most of the decays (84.6%) populate the excited level of {sup 237}Np with energy of 59.54 keV. Digital coincidence counting was applied to standardize a solution of {sup 241}Am by alpha-gamma coincidence counting with efficiency extrapolation. Electronic discrimination was implemented with a pressurized proportional counter and the results were compared with two other independent techniques: Liquid scintillation counting using the logical sum of double coincidences in a TDCR array and defined solid angle counting taking into account activity inhomogeneity in the active deposit. The results show consistency between the three methods within a limit of a 0.3%. An ampoule of this solution will be sent to the International Reference System (SIR) during 2009. Uncertainties were analysed and compared in detail for the three applied methods.

  12. Do your syringes count?

    International Nuclear Information System (INIS)

    Brewster, K.

    2002-01-01

    Full text: This study was designed to investigate anecdotal evidence that residual Sestamibi (MIBI) activity vaned in certain situations. For rest studies different brands of syringes were tested to see if the residuals varied. The period of time MIBI doses remained in the syringe between dispensing and injection was also considered as a possible source of increased residual counts. Stress Mibi syringe residual activities were measured to assess if the method of stress test affected residual activity. MIBI was reconstituted using 13 Gbq of Technetium in 3mls of normal saline then boiled for 10 minutes. Doses were dispensed according to department protocol and injected via cannula. Residual syringes were collected for three syringe types. In each case the barrel and plunger were measured separately. As the syringe is flushed during the exercise stress test and not the pharmacological stress test the chosen method was recorded. No relationship was demonstrated between the time MIBI remained in a syringe prior to injection and residual activity. Residual activity was not affected by method of stress test used. Actual injected activity can be calculated if the amount of activity remaining in the syringe post injection is known. Imaging time can be adjusted for residual activity to optimise count statistics. Preliminary results in this study indicate there is no difference in residual activity between syringe brands.Copyright (2002) The Australian and New Zealand Society of Nuclear Medicine Inc

  13. Relativistic beaming and quasar statistics

    International Nuclear Information System (INIS)

    Orr, M.J.L.; Browne, I.W.A.

    1982-01-01

    The statistical predictions of a unified scheme for the radio emission from quasars are explored. This scheme attributes the observed differences between flat- and steep-spectrum quasars to projection and the effects of relativistic beaming of the emission from the nuclear components. We use a simple quasar model consisting of a compact relativistically beamed core with spectral index zero and unbeamed lobes, spectral index - 1, to predict the proportion of flat-spectrum sources in flux-limited samples selected at different frequencies. In our model this fraction depends on the core Lorentz factor, γ and we find that a value of approximately 5 gives satisfactory agreement with observation. In a similar way the model is used to construct the expected number/flux density counts for flat-spectrum quasars from the observed steep-spectrum counts. Again, good agreement with the observations is obtained if the average core Lorentz factor is about 5. Independent estimates of γ from observations of superluminal motion in quasars are of the same order of magnitude. We conclude that the statistical properties of quasars are entirely consistent with the predictions of simple relativistic-beam models. (author)

  14. Quantifying the sources of variability in equine faecal egg counts: implications for improving the utility of the method.

    Science.gov (United States)

    Denwood, M J; Love, S; Innocent, G T; Matthews, L; McKendrick, I J; Hillary, N; Smith, A; Reid, S W J

    2012-08-13

    The faecal egg count (FEC) is the most widely used means of quantifying the nematode burden of horses, and is frequently used in clinical practice to inform treatment and prevention. The statistical process underlying the FEC is complex, comprising a Poisson counting error process for each sample, compounded with an underlying continuous distribution of means between samples. Being able to quantify the sources of variability contributing to this distribution of means is a necessary step towards providing estimates of statistical power for future FEC and FECRT studies, and may help to improve the usefulness of the FEC technique by identifying and minimising unwanted sources of variability. Obtaining such estimates require a hierarchical statistical model coupled with repeated FEC observations from a single animal over a short period of time. Here, we use this approach to provide the first comparative estimate of multiple sources of within-horse FEC variability. The results demonstrate that a substantial proportion of the observed variation in FEC between horses occurs as a result of variation in FEC within an animal, with the major sources being aggregation of eggs within faeces and variation in egg concentration between faecal piles. The McMaster procedure itself is associated with a comparatively small coefficient of variation, and is therefore highly repeatable when a sufficiently large number of eggs are observed to reduce the error associated with the counting process. We conclude that the variation between samples taken from the same animal is substantial, but can be reduced through the use of larger homogenised faecal samples. Estimates are provided for the coefficient of variation (cv) associated with each within animal source of variability in observed FEC, allowing the usefulness of individual FEC to be quantified, and providing a basis for future FEC and FECRT studies. Copyright © 2012 Elsevier B.V. All rights reserved.

  15. Analysis of thrips distribution: application of spatial statistics and Kriging

    Science.gov (United States)

    John Aleong; Bruce L. Parker; Margaret Skinner; Diantha Howard

    1991-01-01

    Kriging is a statistical technique that provides predictions for spatially and temporally correlated data. Observations of thrips distribution and density in Vermont soils are made in both space and time. Traditional statistical analysis of such data assumes that the counts taken over space and time are independent, which is not necessarily true. Therefore, to analyze...

  16. Ballistic deficit correction methods for large Ge detectors-high counting rate study

    International Nuclear Information System (INIS)

    Duchene, G.; Moszynski, M.

    1995-01-01

    This study presents different ballistic correction methods versus input count rate (from 3 to 50 kcounts/s) using four large Ge detectors of about 70 % relative efficiency. It turns out that the Tennelec TC245 linear amplifier in the BDC mode (Hinshaw method) is the best compromise for energy resolution throughout. All correction methods lead to narrow sum-peaks indistinguishable from single Γ lines. The full energy peak throughput is found representative of the pile-up inspection dead time of the corrector circuits. This work also presents a new and simple representation, plotting simultaneously energy resolution and throughput versus input count rate. (TEC). 12 refs., 11 figs

  17. Correlation of Leukocyte Count and Percentage of Segmented Neutrophils with Pathohistological Findings of Appendix in Children

    Directory of Open Access Journals (Sweden)

    Marko Baskovic

    2018-01-01

    Full Text Available BackgroundAppendicitis is the most common indication for an emergency operation in children's age. Although none of the laboratory values has not high sensitivity and specificity for the diagnosis of appendicitis, leukocyte count and the percentage of segmented neutrophils are most commonly used. The aim of this study was to determine whether there is a statistically significant correlation between leukocyte count and the percentage of segmented neutrophils compared to the pathohistological finding of appendix in children. Materials and MethodsWe retrospectively analyzed the data in the period from 1 January 2016 to 31 December 2016. The analysis was made on 211 patients. Spearman's correlation coefficient (rs was calculated. We determined the specificity and sensitivity of leukocyte count and the percentage of segmented neutrophils used in the calculation of Alvorado and Pediatric Appendicitis score.ResultsThe results of the research have shown that the correlation between leukocyte count and the pathohistological findings is weak (rs = 0.29, p = 3.61*10-8, while there is no correlation between the percentage of segmented neutrophils and pathohistological findings (rs = 0.18, p = 7.08 *10-5. The sensitivity of leukocyte count is 93% and the specificity is 30%, while the sensitivity to the percentage of segmented neutrophils is 71% and the specificity is 50%. ROC analysis for leukocytes shows area under the curve of 0.648, while for segmented neutrophils of 0.574.ConclusionGiven the correlation results obtained, the clinical experience of physicians will still have one of the leading roles in diagnosing acute appendicitis in children.

  18. Counting the peaks in the excitation function for precompound processes

    International Nuclear Information System (INIS)

    Bonetti, R.; Hussein, M.S.; Mello, P.A.

    1982-07-01

    The 'counting of maxima' method of Brink and Stephen, conventionally used for the extraction of the correlation width of statistical (compound nucleus) reactions, is generalized to include precompound processes as well. It is found that this method supplies an important independent check of the results obtained from auto-correlation studies. An application is made to the reaction 25 Mg( 3 He,p). (Author) [pt

  19. Focus in High School Mathematics: Statistics and Probability

    Science.gov (United States)

    National Council of Teachers of Mathematics, 2009

    2009-01-01

    Reasoning about and making sense of statistics and probability are essential to students' future success. This volume belongs to a series that supports National Council of Teachers of Mathematics' (NCTM's) "Focus in High School Mathematics: Reasoning and Sense Making" by providing additional guidance for making reasoning and sense making part of…

  20. Counting paths with Schur transitions

    Energy Technology Data Exchange (ETDEWEB)

    Díaz, Pablo [Department of Physics and Astronomy, University of Lethbridge, Lethbridge, Alberta, T1K 3M4 (Canada); Kemp, Garreth [Department of Physics, University of Johannesburg, P.O. Box 524, Auckland Park 2006 (South Africa); Véliz-Osorio, Alvaro, E-mail: aveliz@gmail.com [Mandelstam Institute for Theoretical Physics, University of the Witwatersrand, WITS 2050, Johannesburg (South Africa); School of Physics and Astronomy, Queen Mary, University of London, Mile End Road, London E1 4NS (United Kingdom)

    2016-10-15

    In this work we explore the structure of the branching graph of the unitary group using Schur transitions. We find that these transitions suggest a new combinatorial expression for counting paths in the branching graph. This formula, which is valid for any rank of the unitary group, reproduces known asymptotic results. We proceed to establish the general validity of this expression by a formal proof. The form of this equation strongly hints towards a quantum generalization. Thus, we introduce a notion of quantum relative dimension and subject it to the appropriate consistency tests. This new quantity finds its natural environment in the context of RCFTs and fractional statistics; where the already established notion of quantum dimension has proven to be of great physical importance.

  1. Clustering method for counting passengers getting in a bus with single camera

    Science.gov (United States)

    Yang, Tao; Zhang, Yanning; Shao, Dapei; Li, Ying

    2010-03-01

    Automatic counting of passengers is very important for both business and security applications. We present a single-camera-based vision system that is able to count passengers in a highly crowded situation at the entrance of a traffic bus. The unique characteristics of the proposed system include, First, a novel feature-point-tracking- and online clustering-based passenger counting framework, which performs much better than those of background-modeling-and foreground-blob-tracking-based methods. Second, a simple and highly accurate clustering algorithm is developed that projects the high-dimensional feature point trajectories into a 2-D feature space by their appearance and disappearance times and counts the number of people through online clustering. Finally, all test video sequences in the experiment are captured from a real traffic bus in Shanghai, China. The results show that the system can process two 320×240 video sequences at a frame rate of 25 fps simultaneously, and can count passengers reliably in various difficult scenarios with complex interaction and occlusion among people. The method achieves high accuracy rates up to 96.5%.

  2. On statistical analysis of compound point process

    Czech Academy of Sciences Publication Activity Database

    Volf, Petr

    2006-01-01

    Roč. 35, 2-3 (2006), s. 389-396 ISSN 1026-597X R&D Projects: GA ČR(CZ) GA402/04/1294 Institutional research plan: CEZ:AV0Z10750506 Keywords : counting process * compound process * hazard function * Cox -model Subject RIV: BB - Applied Statistics, Operational Research

  3. Statistics of multi-tube detecting systems

    International Nuclear Information System (INIS)

    Grau Carles, P.; Grau Malonda, A.

    1994-01-01

    In this paper three new statistical theorems are demonstrated and applied. These theorems simplify very much the obtention of the formulae to compute the counting efficiency when the detection system is formed by several photomultipliers associated in coincidence and sume. These theorems are applied to several photomultiplier arrangements in order to show their potential and the application. way

  4. Pedestrian count estimation using texture feature with spatial distribution

    Directory of Open Access Journals (Sweden)

    Hongyu Hu

    2016-12-01

    Full Text Available We present a novel pedestrian count estimation approach based on global image descriptors formed from multi-scale texture features that considers spatial distribution. For regions of interest, local texture features are represented based on histograms of multi-scale block local binary pattern, which jointly constitute the feature vector of the whole image. Therefore, to achieve an effective estimation of pedestrian count, principal component analysis is used to reduce the dimension of the global representation features, and a fitting model between image global features and pedestrian count is constructed via support vector regression. The experimental result shows that the proposed method exhibits high accuracy on pedestrian count estimation and can be applied well in the real world.

  5. Detectors for proton counting. Si-APD and scintillation detectors

    International Nuclear Information System (INIS)

    Kishimoto, Shunji

    2008-01-01

    Increased intensity of synchrotron radiation requests users to prepare photon pulse detectors having higher counting rates. As detectors for photon counting, silicon-avalanche photodiode (Si-APD) and scintillation detectors were chosen for the fifth series of detectors. Principle of photon detection by pulse and need of amplification function of the detector were described. Structure and working principle, high counting rate measurement system, bunch of electrons vs. counting rate, application example of NMR time spectroscopy measurement and comments for users were described for the Si-APD detector. Structure of scintillator and photomultiplier tube, characteristics of scintillator and performance of detector were shown for the NaI detector. Future development of photon pulse detectors was discussed. (T. Tanaka)

  6. Fluorescent Biosensors Based on Single-Molecule Counting.

    Science.gov (United States)

    Ma, Fei; Li, Ying; Tang, Bo; Zhang, Chun-Yang

    2016-09-20

    Biosensors for highly sensitive, selective, and rapid quantification of specific biomolecules make great contributions to biomedical research, especially molecular diagnostics. However, conventional methods for biomolecular assays often suffer from insufficient sensitivity and poor specificity. In some case (e.g., early disease diagnostics), the concentration of target biomolecules is too low to be detected by these routine approaches, and cumbersome procedures are needed to improve the detection sensitivity. Therefore, there is an urgent need for rapid and ultrasensitive analytical tools. In this respect, single-molecule fluorescence approaches may well satisfy the requirement and hold promising potential for the development of ultrasensitive biosensors. Encouragingly, owing to the advances in single-molecule microscopy and spectroscopy over past decades, the detection of single fluorescent molecule comes true, greatly boosting the development of highly sensitive biosensors. By in vitro/in vivo labeling of target biomolecules with proper fluorescent tags, the quantification of certain biomolecule at the single-molecule level is achieved. In comparison with conventional ensemble measurements, single-molecule detection-based analytical methods possess the advantages of ultrahigh sensitivity, good selectivity, rapid analysis time, and low sample consumption. Consequently, single-molecule detection may be potentially employed as an ideal analytical approach to quantify low-abundant biomolecules with rapidity and simplicity. In this Account, we will summarize our efforts for developing a series of ultrasensitive biosensors based on single-molecule counting. Single-molecule counting is a member of single-molecule detection technologies and may be used as a very simple and ultrasensitive method to quantify target molecules by simply counting the individual fluorescent bursts. In the fluorescent sensors, the signals of target biomolecules may be translated to the

  7. A constant velocity Moessbauer spectrometer free of long-term instrumental drifts in the count rate

    International Nuclear Information System (INIS)

    Sarma, P.R.; Sharma, A.K.; Tripathi, K.C.

    1979-01-01

    Two new control circuits to be used with a constant velocity Moessbauer spectrometer with a loud-speaker drive have been described. The wave-forms generated in the circuits are of the stair-case type instead of the usual square wave-form, so that in each oscillation of the source it remains stationary for a fraction of the time-period. The gamma-rays counted during this period are monitored along with the positive and negative velocity counts and are used to correct any fluctuation in the count rate by feeding these pulses into the timer. The associated logic circuits have been described and the statistical errors involved in the circuits have been computed. (auth.)

  8. Clean Hands Count

    Medline Plus

    Full Text Available ... Like this video? Sign in to make your opinion count. Sign in 131 2 Don't like this video? Sign in to make your opinion count. Sign in 3 Loading... Loading... Transcript The ...

  9. Statistical decision making with a dual-detector probe

    International Nuclear Information System (INIS)

    Hickernell, T.S.

    1988-01-01

    Conventional imaging techniques for cancer detection have difficulty finding small, deep tumors. Single-detector radiation probes have been developed to search for deep lesions in a patient who has been given a tumor-seeking radiopharmaceutical. These probes perform poorly, however, when the background activity in the patient varies greatly from site to site. We have developed a surgical dual-detector probe that solves the problem of background activity variation, by simultaneously monitoring counts from a region of interest and counts from adjacent normal tissue. A comparison of counts from the detectors can reveal the class of tissue, tumor or normal, in the region of interest. In this study, we apply methods from statistical decision theory and derive a suitable comparison of counts to help us decide whether a tumor is present in the region of interest. We use the Hotelling trace criterion with a few assumptions to find a linear discriminant function, which can be reduced to a normalized subtraction of the counts for large background count-rate variations. Using a spatial response map of the dual probe, a computer torso phantom, and estimates of activity distribution, we simulate a surgical staging procedure to test the dual probe and the discriminant functions

  10. Bayesian analysis of energy and count rate data for detection of low count rate radioactive sources

    Energy Technology Data Exchange (ETDEWEB)

    Klumpp, John [Colorado State University, Department of Environmental and Radiological Health Sciences, Molecular and Radiological Biosciences Building, Colorado State University, Fort Collins, Colorado, 80523 (United States)

    2013-07-01

    We propose a radiation detection system which generates its own discrete sampling distribution based on past measurements of background. The advantage to this approach is that it can take into account variations in background with respect to time, location, energy spectra, detector-specific characteristics (i.e. different efficiencies at different count rates and energies), etc. This would therefore be a 'machine learning' approach, in which the algorithm updates and improves its characterization of background over time. The system would have a 'learning mode,' in which it measures and analyzes background count rates, and a 'detection mode,' in which it compares measurements from an unknown source against its unique background distribution. By characterizing and accounting for variations in the background, general purpose radiation detectors can be improved with little or no increase in cost. The statistical and computational techniques to perform this kind of analysis have already been developed. The necessary signal analysis can be accomplished using existing Bayesian algorithms which account for multiple channels, multiple detectors, and multiple time intervals. Furthermore, Bayesian machine-learning techniques have already been developed which, with trivial modifications, can generate appropriate decision thresholds based on the comparison of new measurements against a nonparametric sampling distribution. (authors)

  11. Investigation on Carbohydrate Counting Method in Type 1 Diabetic Patients

    Directory of Open Access Journals (Sweden)

    Osman Son

    2014-01-01

    Full Text Available Objective. The results from Diabetes Control and Complications Trial (DCCT have propounded the importance of the approach of treatment by medical nutrition when treating diabetes mellitus (DM. During this study, we tried to inquire carbohydrate (Kh count method’s positive effects on the type 1 DM treatment’s success as well as on the life quality of the patients. Methods. 22 of 37 type 1 DM patients who applied to Eskişehir Osmangazi University, Faculty of Medicine Hospital, Department of Endocrinology and Metabolism, had been treated by Kh count method and 15 of them are treated by multiple dosage intensive insulin treatment with applying standard diabetic diet as a control group and both of groups were under close follow-up for 6 months. Required approval was taken from the Ethical Committee of Eskişehir Osmangazi University, Medical Faculty, as well as informed consent from the patients. The body weight of patients who are treated by carbohydrate count method and multiple dosage intensive insulin treatment during the study beginning and after 6-month term, body mass index, and body compositions are analyzed. A short life quality and medical research survey applied. At statistical analysis, t-test, chi-squared test, and Mann-Whitney U test were used. Results. There had been no significant change determined at glycemic control indicators between the Kh counting group and the standard diabetic diet and multiple dosage insulin treatment group in our study. Conclusion. As a result, Kh counting method which offers a flexible nutrition plan to diabetic individuals is a functional method.

  12. Drone-based Object Counting by Spatially Regularized Regional Proposal Network

    OpenAIRE

    Hsieh, Meng-Ru; Lin, Yen-Liang; Hsu, Winston H.

    2017-01-01

    Existing counting methods often adopt regression-based approaches and cannot precisely localize the target objects, which hinders the further analysis (e.g., high-level understanding and fine-grained classification). In addition, most of prior work mainly focus on counting objects in static environments with fixed cameras. Motivated by the advent of unmanned flying vehicles (i.e., drones), we are interested in detecting and counting objects in such dynamic environments. We propose Layout Prop...

  13. Noun Countability; Count Nouns and Non-count Nouns, What are the Syntactic Differences Between them?

    Directory of Open Access Journals (Sweden)

    Azhar A. Alkazwini

    2016-11-01

    Full Text Available Words that function as the subjects of verbs, objects of verbs or prepositions and which can have a plural form and possessive ending are known as nouns. They are described as referring to persons, places, things, states, or qualities and might also be used as an attributive modifier. In this paper, classes and subclasses of nouns shall be presented, then, noun countability branching into count and non-count nous shall be discussed. A number of present examples illustrating differences between count and non-count nouns and this includes determiner-head-co-occurrence restrictions of number, subject-verb agreement, in addition to some exceptions to this agreement rule shall be discussed. Also, the lexically inherent number in nouns and how inherently plural nouns are classified in terms of (+/- count are illustrated. This research will discuss partitive construction of count and non-count nouns, nouns as attributive modifier and, finally, conclude with the fact that there are syntactic difference between count and non-count in the English Language.

  14. Evidence for significant influence of host immunity on changes in differential blood count during malaria.

    Science.gov (United States)

    Berens-Riha, Nicole; Kroidl, Inge; Schunk, Mirjam; Alberer, Martin; Beissner, Marcus; Pritsch, Michael; Kroidl, Arne; Fröschl, Günter; Hanus, Ingrid; Bretzel, Gisela; von Sonnenburg, Frank; Nothdurft, Hans Dieter; Löscher, Thomas; Herbinger, Karl-Heinz

    2014-04-23

    Malaria has been shown to change blood counts. Recently, a few studies have investigated the alteration of the peripheral blood monocyte-to-lymphocyte count ratio (MLCR) and the neutrophil-to-lymphocyte count ratio (NLCR) during infection with Plasmodium falciparum. Based on these findings this study investigates the predictive values of blood count alterations during malaria across different sub-populations. Cases and controls admitted to the Department of Infectious Diseases and Tropical Medicine from January 2000 through December 2010 were included in this comparative analysis. Blood count values and other variables at admission controlled for age, gender and immune status were statistically investigated. The study population comprised 210 malaria patients, infected with P. falciparum (68%), Plasmodium vivax (21%), Plasmodium ovale (7%) and Plasmodium malariae (4%), and 210 controls. A positive correlation of parasite density with NLCR and neutrophil counts, and a negative correlation of parasite density with thrombocyte, leucocyte and lymphocyte counts were found. An interaction with semi-immunity was observed; ratios were significantly different in semi-immune compared to non-immune patients (P value of the ratios was fair but limited. However, these changes were less pronounced in patients with semi-immunity. The ratios might constitute easily applicable surrogate biomarkers for immunity.

  15. Clean Hands Count

    Medline Plus

    Full Text Available ... starting stop Loading... Watch Queue Queue __count__/__total__ It’s YouTube. Uninterrupted. Loading... Want music and videos with ... ads? Get YouTube Red. Working... Not now Try it free Find out why Close Clean Hands Count ...

  16. Statistical properties of several models of fractional random point processes

    Science.gov (United States)

    Bendjaballah, C.

    2011-08-01

    Statistical properties of several models of fractional random point processes have been analyzed from the counting and time interval statistics points of view. Based on the criterion of the reduced variance, it is seen that such processes exhibit nonclassical properties. The conditions for these processes to be treated as conditional Poisson processes are examined. Numerical simulations illustrate part of the theoretical calculations.

  17.  Risk of discontinuation of nevirapine due to toxicities in antiretroviral naive and experienced HIV-infected patients with high and low CD4 counts

    DEFF Research Database (Denmark)

    Mocroft, A; Staszewski, S; Weber, R

    2007-01-01

    AND METHODS: 1,571 EuroSIDA patients started NVPc after 1/1/1999, with CD4+ T-cell counts and viral load measured in the 6 months before starting treatment, and were stratified into four groups based on CD4+ T-cell counts at initiation of NVPc (high [H], > 400/mm3 or > 250/mm3 for male or female, respectively...

  18. Maximum entropy deconvolution of low count nuclear medicine images

    International Nuclear Information System (INIS)

    McGrath, D.M.

    1998-12-01

    Maximum entropy is applied to the problem of deconvolving nuclear medicine images, with special consideration for very low count data. The physics of the formation of scintigraphic images is described, illustrating the phenomena which degrade planar estimates of the tracer distribution. Various techniques which are used to restore these images are reviewed, outlining the relative merits of each. The development and theoretical justification of maximum entropy as an image processing technique is discussed. Maximum entropy is then applied to the problem of planar deconvolution, highlighting the question of the choice of error parameters for low count data. A novel iterative version of the algorithm is suggested which allows the errors to be estimated from the predicted Poisson mean values. This method is shown to produce the exact results predicted by combining Poisson statistics and a Bayesian interpretation of the maximum entropy approach. A facility for total count preservation has also been incorporated, leading to improved quantification. In order to evaluate this iterative maximum entropy technique, two comparable methods, Wiener filtering and a novel Bayesian maximum likelihood expectation maximisation technique, were implemented. The comparison of results obtained indicated that this maximum entropy approach may produce equivalent or better measures of image quality than the compared methods, depending upon the accuracy of the system model used. The novel Bayesian maximum likelihood expectation maximisation technique was shown to be preferable over many existing maximum a posteriori methods due to its simplicity of implementation. A single parameter is required to define the Bayesian prior, which suppresses noise in the solution and may reduce the processing time substantially. Finally, maximum entropy deconvolution was applied as a pre-processing step in single photon emission computed tomography reconstruction of low count data. Higher contrast results were

  19. Statistical properties of superimposed stationary spike trains.

    Science.gov (United States)

    Deger, Moritz; Helias, Moritz; Boucsein, Clemens; Rotter, Stefan

    2012-06-01

    The Poisson process is an often employed model for the activity of neuronal populations. It is known, though, that superpositions of realistic, non- Poisson spike trains are not in general Poisson processes, not even for large numbers of superimposed processes. Here we construct superimposed spike trains from intracellular in vivo recordings from rat neocortex neurons and compare their statistics to specific point process models. The constructed superimposed spike trains reveal strong deviations from the Poisson model. We find that superpositions of model spike trains that take the effective refractoriness of the neurons into account yield a much better description. A minimal model of this kind is the Poisson process with dead-time (PPD). For this process, and for superpositions thereof, we obtain analytical expressions for some second-order statistical quantities-like the count variability, inter-spike interval (ISI) variability and ISI correlations-and demonstrate the match with the in vivo data. We conclude that effective refractoriness is the key property that shapes the statistical properties of the superposition spike trains. We present new, efficient algorithms to generate superpositions of PPDs and of gamma processes that can be used to provide more realistic background input in simulations of networks of spiking neurons. Using these generators, we show in simulations that neurons which receive superimposed spike trains as input are highly sensitive for the statistical effects induced by neuronal refractoriness.

  20. Subcritical Multiplicative Chaos for Regularized Counting Statistics from Random Matrix Theory

    Science.gov (United States)

    Lambert, Gaultier; Ostrovsky, Dmitry; Simm, Nick

    2018-05-01

    For an {N × N} Haar distributed random unitary matrix U N , we consider the random field defined by counting the number of eigenvalues of U N in a mesoscopic arc centered at the point u on the unit circle. We prove that after regularizing at a small scale {ɛN > 0}, the renormalized exponential of this field converges as N \\to ∞ to a Gaussian multiplicative chaos measure in the whole subcritical phase. We discuss implications of this result for obtaining a lower bound on the maximum of the field. We also show that the moments of the total mass converge to a Selberg-like integral and by taking a further limit as the size of the arc diverges, we establish part of the conjectures in Ostrovsky (Nonlinearity 29(2):426-464, 2016). By an analogous construction, we prove that the multiplicative chaos measure coming from the sine process has the same distribution, which strongly suggests that this limiting object should be universal. Our approach to the L 1-phase is based on a generalization of the construction in Berestycki (Electron Commun Probab 22(27):12, 2017) to random fields which are only asymptotically Gaussian. In particular, our method could have applications to other random fields coming from either random matrix theory or a different context.

  1. Statistics of multi-tube detecting systems

    International Nuclear Information System (INIS)

    Grau Carles, P.; Grau Malonda, A.

    1994-01-01

    In this paper three new statistical theorems are demonstrated and applied. These theorems simplify very much the obtention of the formulae to compute the counting efficiency when the detection system is formed by several photomultipliers associated in coincidence and sum. These theorems are applied to several photomultiplier arrangements in order to show their potential and the application way. (Author) 6 refs

  2. Count-to-count time interval distribution analysis in a fast reactor

    International Nuclear Information System (INIS)

    Perez-Navarro Gomez, A.

    1973-01-01

    The most important kinetic parameters have been measured at the zero power fast reactor CORAL-I by means of the reactor noise analysis in the time domain, using measurements of the count-to-count time intervals. (Author) 69 refs

  3. Statistics and error considerations at the application of SSND T-technique in radon measurement

    International Nuclear Information System (INIS)

    Jonsson, G.

    1993-01-01

    Plastic films are used for the detection of alpha particles from disintegrating radon and radon daughter nuclei. After etching there are tracks (cones) or holes in the film as a result of the exposure. The step from a counted number of tracks/holes per surface unit of the film to a reliable value of the radon and radon daughter level is surrounded by statistical considerations of different nature. Some of them are the number of counted tracks, the length of the time of exposure, the season of the time of exposure, the etching technique and the method of counting the tracks or holes. The number of background tracks of an unexposed film increases the error of the measured radon level. Some of the mentioned effects of statistical nature will be discussed in the report. (Author)

  4. Development of an automated asbestos counting software based on fluorescence microscopy.

    Science.gov (United States)

    Alexandrov, Maxym; Ichida, Etsuko; Nishimura, Tomoki; Aoki, Kousuke; Ishida, Takenori; Hirota, Ryuichi; Ikeda, Takeshi; Kawasaki, Tetsuo; Kuroda, Akio

    2015-01-01

    An emerging alternative to the commonly used analytical methods for asbestos analysis is fluorescence microscopy (FM), which relies on highly specific asbestos-binding probes to distinguish asbestos from interfering non-asbestos fibers. However, all types of microscopic asbestos analysis require laborious examination of large number of fields of view and are prone to subjective errors and large variability between asbestos counts by different analysts and laboratories. A possible solution to these problems is automated counting of asbestos fibers by image analysis software, which would lower the cost and increase the reliability of asbestos testing. This study seeks to develop a fiber recognition and counting software for FM-based asbestos analysis. We discuss the main features of the developed software and the results of its testing. Software testing showed good correlation between automated and manual counts for the samples with medium and high fiber concentrations. At low fiber concentrations, the automated counts were less accurate, leading us to implement correction mode for automated counts. While the full automation of asbestos analysis would require further improvements in accuracy of fiber identification, the developed software could already assist professional asbestos analysts and record detailed fiber dimensions for the use in epidemiological research.

  5. Quantum Biometrics with Retinal Photon Counting

    Science.gov (United States)

    Loulakis, M.; Blatsios, G.; Vrettou, C. S.; Kominis, I. K.

    2017-10-01

    It is known that the eye's scotopic photodetectors, rhodopsin molecules, and their associated phototransduction mechanism leading to light perception, are efficient single-photon counters. We here use the photon-counting principles of human rod vision to propose a secure quantum biometric identification based on the quantum-statistical properties of retinal photon detection. The photon path along the human eye until its detection by rod cells is modeled as a filter having a specific transmission coefficient. Precisely determining its value from the photodetection statistics registered by the conscious observer is a quantum parameter estimation problem that leads to a quantum secure identification method. The probabilities for false-positive and false-negative identification of this biometric technique can readily approach 10-10 and 10-4, respectively. The security of the biometric method can be further quantified by the physics of quantum measurements. An impostor must be able to perform quantum thermometry and quantum magnetometry with energy resolution better than 10-9ℏ , in order to foil the device by noninvasively monitoring the biometric activity of a user.

  6. An Economical Fast Discriminator for Nuclear Pulse Counting

    International Nuclear Information System (INIS)

    Issarachai, Opas; Punnachaiya, Suvit

    2009-07-01

    Full text: This research work was aimed to develop a fast discriminator at low cost but high capability for discrimination a nanosecond nuclear pulse. The fast discriminator can be used in association with fast photon counting system. The designed structure consisted of the ultra-fast voltage comparator using ADCMP601 integrated circuit, the monostable multivibrator with controllable pulse width output by propagation delay of logic gate, and the fast response buffer amplifier. The tested results of pulse height discrimination of 0-5 V nuclear pulse with 20 ns (FWHM) pulse width showed the correlation coefficient (R 2 ) between discrimination level and pulse height was 0.998, while the pulse rate more than 10 MHz could be counted. The 30 ns logic pulse width output revealed high stable and could be smoothly driven to low impedance load at 50 Ω. For pulse signal transmission to the counter, it was also found that the termination of reflected signal must be considered because it may cause pulse counting error

  7. Statistical modelling of citation exchange between statistics journals.

    Science.gov (United States)

    Varin, Cristiano; Cattelan, Manuela; Firth, David

    2016-01-01

    Rankings of scholarly journals based on citation data are often met with scepticism by the scientific community. Part of the scepticism is due to disparity between the common perception of journals' prestige and their ranking based on citation counts. A more serious concern is the inappropriate use of journal rankings to evaluate the scientific influence of researchers. The paper focuses on analysis of the table of cross-citations among a selection of statistics journals. Data are collected from the Web of Science database published by Thomson Reuters. Our results suggest that modelling the exchange of citations between journals is useful to highlight the most prestigious journals, but also that journal citation data are characterized by considerable heterogeneity, which needs to be properly summarized. Inferential conclusions require care to avoid potential overinterpretation of insignificant differences between journal ratings. Comparison with published ratings of institutions from the UK's research assessment exercise shows strong correlation at aggregate level between assessed research quality and journal citation 'export scores' within the discipline of statistics.

  8. A Multispectral Photon-Counting Double Random Phase Encoding Scheme for Image Authentication

    Directory of Open Access Journals (Sweden)

    Faliu Yi

    2014-05-01

    Full Text Available In this paper, we propose a new method for color image-based authentication that combines multispectral photon-counting imaging (MPCI and double random phase encoding (DRPE schemes. The sparsely distributed information from MPCI and the stationary white noise signal from DRPE make intruder attacks difficult. In this authentication method, the original multispectral RGB color image is down-sampled into a Bayer image. The three types of color samples (red, green and blue color in the Bayer image are encrypted with DRPE and the amplitude part of the resulting image is photon counted. The corresponding phase information that has nonzero amplitude after photon counting is then kept for decryption. Experimental results show that the retrieved images from the proposed method do not visually resemble their original counterparts. Nevertheless, the original color image can be efficiently verified with statistical nonlinear correlations. Our experimental results also show that different interpolation algorithms applied to Bayer images result in different verification effects for multispectral RGB color images.

  9. A multispectral photon-counting double random phase encoding scheme for image authentication.

    Science.gov (United States)

    Yi, Faliu; Moon, Inkyu; Lee, Yeon H

    2014-05-20

    In this paper, we propose a new method for color image-based authentication that combines multispectral photon-counting imaging (MPCI) and double random phase encoding (DRPE) schemes. The sparsely distributed information from MPCI and the stationary white noise signal from DRPE make intruder attacks difficult. In this authentication method, the original multispectral RGB color image is down-sampled into a Bayer image. The three types of color samples (red, green and blue color) in the Bayer image are encrypted with DRPE and the amplitude part of the resulting image is photon counted. The corresponding phase information that has nonzero amplitude after photon counting is then kept for decryption. Experimental results show that the retrieved images from the proposed method do not visually resemble their original counterparts. Nevertheless, the original color image can be efficiently verified with statistical nonlinear correlations. Our experimental results also show that different interpolation algorithms applied to Bayer images result in different verification effects for multispectral RGB color images.

  10. Effect of fractionated regional external beam radiotherapy on peripheral blood cell count

    International Nuclear Information System (INIS)

    Zachariah, B.; Jacob, S.S.; Gwede, C.; Cantor, A.; Patil, J.; Casey, L.; Zachariah, A.B.

    2001-01-01

    Purpose: The purpose of this study was to assess the need for obtaining weekly complete blood count (CBC) values and to identify the pattern of changes in CBC during regional conventional fractionated radiotherapy. Methods and Materials: A retrospective analysis of CBC data on 299 adult cancer patients who received definitive conventional radiotherapy to head and neck (n=95), chest (n=96), and pelvis (n=108) was performed. Temporal patterns and magnitude of change in white blood cells, neutrophils, lymphocytes, and platelets during radiotherapy were examined. Results: There were statistically significant declines in all counts, albeit not clinically significant. Notable differences between disease sites were found. The greatest weekly interval change in counts occurred during the first week of radiotherapy for all groups of patients. The mean WBC nadir values during treatment were 5.8 for head and neck, 6.8 for chest, and 5.4 for pelvis. The nadirs for all counts occurred toward the middle-to-end of radiotherapy. Lymphocytes were found to be more sensitive to radiotherapy than other leukocyte subcomponents. Conclusion: Our study suggests that weekly CBC monitoring is not necessary for all patients undergoing standard fractionated radiotherapy. Baseline blood counts may be used to determine an optimal schedule for monitoring CBCs in patients receiving conventional radiation alone. Reduced monitoring of CBC may result in significant financial savings

  11. Highly Sensitive Photon Counting Detectors for Deep Space Optical Communications, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — A new type of a photon-counting photodetector is proposed to advance the state-of the-art in deep space optical communications technology. The proposed detector...

  12. Monitoring Milk Somatic Cell Counts

    Directory of Open Access Journals (Sweden)

    Gheorghe Şteţca

    2014-11-01

    Full Text Available The presence of somatic cells in milk is a widely disputed issue in milk production sector. The somatic cell counts in raw milk are a marker for the specific cow diseases such as mastitis or swollen udder. The high level of somatic cells causes physical and chemical changes to milk composition and nutritional value, and as well to milk products. Also, the mastitic milk is not proper for human consumption due to its contribution to spreading of certain diseases and food poisoning. According to these effects, EU Regulations established the maximum threshold of admitted somatic cells in raw milk to 400000 cells / mL starting with 2014. The purpose of this study was carried out in order to examine the raw milk samples provided from small farms, industrial type farms and milk processing units. There are several ways to count somatic cells in milk but the reference accepted method is the microscopic method described by the SR EN ISO 13366-1/2008. Generally samples registered values in accordance with the admissible limit. By periodical monitoring of the somatic cell count, certain technological process issues are being avoided and consumer’s health ensured.

  13. A Frank mixture copula family for modeling higher-order correlations of neural spike counts

    International Nuclear Information System (INIS)

    Onken, Arno; Obermayer, Klaus

    2009-01-01

    In order to evaluate the importance of higher-order correlations in neural spike count codes, flexible statistical models of dependent multivariate spike counts are required. Copula families, parametric multivariate distributions that represent dependencies, can be applied to construct such models. We introduce the Frank mixture family as a new copula family that has separate parameters for all pairwise and higher-order correlations. In contrast to the Farlie-Gumbel-Morgenstern copula family that shares this property, the Frank mixture copula can model strong correlations. We apply spike count models based on the Frank mixture copula to data generated by a network of leaky integrate-and-fire neurons and compare the goodness of fit to distributions based on the Farlie-Gumbel-Morgenstern family. Finally, we evaluate the importance of using proper single neuron spike count distributions on the Shannon information. We find notable deviations in the entropy that increase with decreasing firing rates. Moreover, we find that the Frank mixture family increases the log likelihood of the fit significantly compared to the Farlie-Gumbel-Morgenstern family. This shows that the Frank mixture copula is a useful tool to assess the importance of higher-order correlations in spike count codes.

  14. Generalized linear models and point count data: statistical considerations for the design and analysis of monitoring studies

    Science.gov (United States)

    Nathaniel E. Seavy; Suhel Quader; John D. Alexander; C. John Ralph

    2005-01-01

    The success of avian monitoring programs to effectively guide management decisions requires that studies be efficiently designed and data be properly analyzed. A complicating factor is that point count surveys often generate data with non-normal distributional properties. In this paper we review methods of dealing with deviations from normal assumptions, and we focus...

  15. Fast radio burst event rate counts - I. Interpreting the observations

    Science.gov (United States)

    Macquart, J.-P.; Ekers, R. D.

    2018-02-01

    The fluence distribution of the fast radio burst (FRB) population (the `source count' distribution, N (>F) ∝Fα), is a crucial diagnostic of its distance distribution, and hence the progenitor evolutionary history. We critically reanalyse current estimates of the FRB source count distribution. We demonstrate that the Lorimer burst (FRB 010724) is subject to discovery bias, and should be excluded from all statistical studies of the population. We re-examine the evidence for flat, α > -1, source count estimates based on the ratio of single-beam to multiple-beam detections with the Parkes multibeam receiver, and show that current data imply only a very weak constraint of α ≲ -1.3. A maximum-likelihood analysis applied to the portion of the Parkes FRB population detected above the observational completeness fluence of 2 Jy ms yields α = -2.6_{-1.3}^{+0.7 }. Uncertainties in the location of each FRB within the Parkes beam render estimates of the Parkes event rate uncertain in both normalizing survey area and the estimated post-beam-corrected completeness fluence; this uncertainty needs to be accounted for when comparing the event rate against event rates measured at other telescopes.

  16. Time Evolving Fission Chain Theory and Fast Neutron and Gamma-Ray Counting Distributions

    International Nuclear Information System (INIS)

    Kim, K. S.; Nakae, L. F.; Prasad, M. K.; Snyderman, N. J.; Verbeke, J. M.

    2015-01-01

    Here, we solve a simple theoretical model of time evolving fission chains due to Feynman that generalizes and asymptotically approaches the point model theory. The point model theory has been used to analyze thermal neutron counting data. This extension of the theory underlies fast counting data for both neutrons and gamma rays from metal systems. Fast neutron and gamma-ray counting is now possible using liquid scintillator arrays with nanosecond time resolution. For individual fission chains, the differential equations describing three correlated probability distributions are solved: the time-dependent internal neutron population, accumulation of fissions in time, and accumulation of leaked neutrons in time. Explicit analytic formulas are given for correlated moments of the time evolving chain populations. The equations for random time gate fast neutron and gamma-ray counting distributions, due to randomly initiated chains, are presented. Correlated moment equations are given for both random time gate and triggered time gate counting. There are explicit formulas for all correlated moments are given up to triple order, for all combinations of correlated fast neutrons and gamma rays. The nonlinear differential equations for probabilities for time dependent fission chain populations have a remarkably simple Monte Carlo realization. A Monte Carlo code was developed for this theory and is shown to statistically realize the solutions to the fission chain theory probability distributions. Combined with random initiation of chains and detection of external quanta, the Monte Carlo code generates time tagged data for neutron and gamma-ray counting and from these data the counting distributions.

  17. Statistics and geometry of cosmic voids

    International Nuclear Information System (INIS)

    Gaite, José

    2009-01-01

    We introduce new statistical methods for the study of cosmic voids, focusing on the statistics of largest size voids. We distinguish three different types of distributions of voids, namely, Poisson-like, lognormal-like and Pareto-like distributions. The last two distributions are connected with two types of fractal geometry of the matter distribution. Scaling voids with Pareto distribution appear in fractal distributions with box-counting dimension smaller than three (its maximum value), whereas the lognormal void distribution corresponds to multifractals with box-counting dimension equal to three. Moreover, voids of the former type persist in the continuum limit, namely, as the number density of observable objects grows, giving rise to lacunar fractals, whereas voids of the latter type disappear in the continuum limit, giving rise to non-lacunar (multi)fractals. We propose both lacunar and non-lacunar multifractal models of the cosmic web structure of the Universe. A non-lacunar multifractal model is supported by current galaxy surveys as well as cosmological N-body simulations. This model suggests, in particular, that small dark matter halos and, arguably, faint galaxies are present in cosmic voids

  18. The effect of volume and quenching on estimation of counting efficiencies in liquid scintillation counting

    International Nuclear Information System (INIS)

    Knoche, H.W.; Parkhurst, A.M.; Tam, S.W.

    1979-01-01

    The effect of volume on the liquid scintillation counting performance of 14 C-samples has been investigated. A decrease in counting efficiency was observed for samples with volumes below about 6 ml and those above about 18 ml when unquenched samples were assayed. Two quench-correction methods, sample channels ratio and external standard channels ratio, and three different liquid scintillation counters, were used in an investigation to determine the magnitude of the error in predicting counting efficiencies when small volume samples (2 ml) with different levels of quenching were assayed. The 2 ml samples exhibited slightly greater standard deviations of the difference between predicted and determined counting efficiencies than did 15 ml samples. Nevertheless, the magnitude of the errors indicate that if the sample channels ratio method of quench correction is employed, 2 ml samples may be counted in conventional counting vials with little loss in counting precision. (author)

  19. Establishment of reference intervals for complete blood count parameters during normal pregnancy in Beijing.

    Science.gov (United States)

    Li, Aiwei; Yang, Shuo; Zhang, Jie; Qiao, Rui

    2017-11-01

    To observe the changes of complete blood count (CBC) parameters during pregnancy and establish appropriate reference intervals for healthy pregnant women. Healthy pregnant women took the blood tests at all trimesters. All blood samples were processed on Sysmex XE-2100. The following CBC parameters were analyzed: red blood cell count (RBC), hemoglobin (Hb), hematocrit (Hct), mean corpuscular volume (MCV), mean corpuscular hemoglobin (MCH), mean corpuscular hemoglobin concentration (MCHC), red blood cell distribution width (RDW), platelet count (PLT), mean platelet volume (MPV), platelet distribution width (PDW), white blood cell count (WBC), and leukocyte differential count. Reference intervals were established using the 2.5th and 97.5th percentile of the distribution. Complete blood count parameters showed dynamic changes during trimesters. RBC, Hb, Hct declined at trimester 1, reaching their lowest point at trimester 2, and began to rise again at trimester 3. WBC, neutrophil count (Neut), monocyte count (MONO), RDW, and PDW went up from trimester 1 to trimester 3. On the contrary, MCHC, lymphocyte count (LYMPH), PLT, and MPV gradually descended during pregnancy. There were statistical significances in all CBC parameters between pregnant women and normal women, regardless of the trimesters (Ppregnancy) as follows: RBC 4.50 vs 3.94×10 12 /L, Hb 137 vs 120 g/L, WBC 5.71 vs 9.06×10 9 /L, LYMPH% 32.2 vs 18.0, Neut% 58.7 vs 75.0, and PLT 251 vs 202×10 9 /L. The changes of CBC parameters during pregnancy are described, and reference intervals for Beijing pregnant women are demonstrated in this study. © 2017 Wiley Periodicals, Inc.

  20. Integrating chronological uncertainties for annually laminated lake sediments using layer counting, independent chronologies and Bayesian age modelling (Lake Ohau, South Island, New Zealand)

    Science.gov (United States)

    Vandergoes, Marcus J.; Howarth, Jamie D.; Dunbar, Gavin B.; Turnbull, Jocelyn C.; Roop, Heidi A.; Levy, Richard H.; Li, Xun; Prior, Christine; Norris, Margaret; Keller, Liz D.; Baisden, W. Troy; Ditchburn, Robert; Fitzsimons, Sean J.; Bronk Ramsey, Christopher

    2018-05-01

    Annually resolved (varved) lake sequences are important palaeoenvironmental archives as they offer a direct incremental dating technique for high-frequency reconstruction of environmental and climate change. Despite the importance of these records, establishing a robust chronology and quantifying its precision and accuracy (estimations of error) remains an essential but challenging component of their development. We outline an approach for building reliable independent chronologies, testing the accuracy of layer counts and integrating all chronological uncertainties to provide quantitative age and error estimates for varved lake sequences. The approach incorporates (1) layer counts and estimates of counting precision; (2) radiometric and biostratigrapic dating techniques to derive independent chronology; and (3) the application of Bayesian age modelling to produce an integrated age model. This approach is applied to a case study of an annually resolved sediment record from Lake Ohau, New Zealand. The most robust age model provides an average error of 72 years across the whole depth range. This represents a fractional uncertainty of ∼5%, higher than the <3% quoted for most published varve records. However, the age model and reported uncertainty represent the best fit between layer counts and independent chronology and the uncertainties account for both layer counting precision and the chronological accuracy of the layer counts. This integrated approach provides a more representative estimate of age uncertainty and therefore represents a statistically more robust chronology.

  1. Statistics associated with an elemental analysis system of particles induced by X-ray emission

    International Nuclear Information System (INIS)

    Romo K, C.M.

    1987-01-01

    In the quantitative elemental analysis by X-ray techniques one has to use data spectra which present fluctuations of statistical nature both from the energy and from the number of counts accumulated. While processing the results for the obtainment of a quantitative result, a detailed knowledge of the associated statistics distributions is needed. In this work, l) the statistics associated with the system photon's counting as well as 2) the distribution of the results as a function of the energy are analyzed. The first one is important for the definition of the expected values and uncertainties and for the spectra simulation (Mukoyama, 1975). The second one is fundamental for the determination of the contribution for each spectral line. (M.R.) [es

  2. Correlation of platelet count and acute ST-elevation myocardial infarction.

    Science.gov (United States)

    Paul, G K; Sen, B; Rahman, M Z; Ali, M; Rahman, M M; Rokonuzzaman, S M

    2014-10-01

    The study was conducted in the Department of cardiology, NICVD Dhaka during the period January 2006 to December 2007 to assess the impact of platelet on ST-elevation myocardial infarction (STEMI). To perform this prospective study 200 patients with STEMI within 72 hours of chest pain of both sexes were randomly selected and were evaluated by clinical history, physical examination and with the help of ECG, Echocardiography and others cardiac risk factors analysis. Heparin therapy before admission, previously documented thrombocytopenia (1.6mg/dl) and history of PCI & CABG were excluded in this study. Patient of Platelet count (PC) ≤200000/cubic millimeter (cmm) in Group I and patient of Group II, platelet counts were PC >200000/cmm. Follow up period was 3 days to 7 days after hospital admission. Primary outcome heart failure (any Killip class) was significantly more in Group II than Group I (40.0% vs. 23.0%; p=0.009). Though the incidence of Killip class I and cardiogenic shock were not significant between these two groups but Killip class II (18.0% vs. 8.0%; p=0.036) and Killip class III (15.0% vs. 6.0%; p=0.037) heart failure were significantly more among the patient with higher platelet counts. In-hospital mortality, one of the primary outcomes of this study, was significantly higher in Group II (13.0%) than Group I (5.0 %) and p value was 0.048. Re-infarction was more in patient with higher platelet counts group (Group II) than patients with lower platelet count (Group I) but statistically was not significant (16.0% vs.11.0%; p=0.300).

  3. Requirements on the Redshift Accuracy for future Supernova and Number Count Surveys

    International Nuclear Information System (INIS)

    Huterer, Dragan; Kim, Alex; Broderick, Tamara

    2004-01-01

    We investigate the required redshift accuracy of type Ia supernova and cluster number-count surveys in order for the redshift uncertainties not to contribute appreciably to the dark energy parameter error budget. For the SNAP supernova experiment, we find that, without the assistance of ground-based measurements, individual supernova redshifts would need to be determined to about 0.002 or better, which is a challenging but feasible requirement for a low-resolution spectrograph. However, we find that accurate redshifts for z < 0.1 supernovae, obtained with ground-based experiments, are sufficient to immunize the results against even relatively large redshift errors at high z. For the future cluster number-count surveys such as the South Pole Telescope, Planck or DUET, we find that the purely statistical error in photometric redshift is less important, and that the irreducible, systematic bias in redshift drives the requirements. The redshift bias will have to be kept below 0.001-0.005 per redshift bin (which is determined by the filter set), depending on the sky coverage and details of the definition of the minimal mass of the survey. Furthermore, we find that X-ray surveys have a more stringent required redshift accuracy than Sunyaev-Zeldovich (SZ) effect surveys since they use a shorter lever arm in redshift; conversely, SZ surveys benefit from their high redshift reach only so long as some redshift information is available for distant (zgtrsim1) clusters

  4. Counted Sb donors in Si quantum dots

    Science.gov (United States)

    Singh, Meenakshi; Pacheco, Jose; Bielejec, Edward; Perry, Daniel; Ten Eyck, Gregory; Bishop, Nathaniel; Wendt, Joel; Luhman, Dwight; Carroll, Malcolm; Lilly, Michael

    2015-03-01

    Deterministic control over the location and number of donors is critical for donor spin qubits in semiconductor based quantum computing. We have developed techniques using a focused ion beam and a diode detector integrated next to a silicon MOS single electron transistor to gain such control. With the diode detector operating in linear mode, the numbers of ions implanted have been counted and single ion implants have been detected. Poisson statistics in the number of ions implanted have been observed. Transport measurements performed on samples with counted number of implants have been performed and regular coulomb blockade and charge offsets observed. The capacitances to various gates are found to be in agreement with QCAD simulations for an electrostatically defined dot. This work was performed, in part, at the Center for Integrated Nanotechnologies, a U.S. DOE Office of Basic Energy Sciences user facility. The work was supported by Sandia National Laboratories Directed Research and Development Program. Sandia National Laboratories is a multi-program laboratory operated by Sandia Corporation, a Lockheed-Martin Company, for the U. S. Department of Energy under Contract No. DE-AC04-94AL85000.

  5. Why do statistics journals have low impact factors?

    NARCIS (Netherlands)

    van Nierop, Erjen

    In this paper, we answer the question why statistics journals get lower impact factors than journals in other disciplines. We analyze diffusion patterns of papers in several journals in various academic fields. To obtain insights into the diffusion of the citation counts of the papers, the data are

  6. HgCdTe APD-based linear-mode photon counting components and ladar receivers

    Science.gov (United States)

    Jack, Michael; Wehner, Justin; Edwards, John; Chapman, George; Hall, Donald N. B.; Jacobson, Shane M.

    2011-05-01

    Linear mode photon counting (LMPC) provides significant advantages in comparison with Geiger Mode (GM) Photon Counting including absence of after-pulsing, nanosecond pulse to pulse temporal resolution and robust operation in the present of high density obscurants or variable reflectivity objects. For this reason Raytheon has developed and previously reported on unique linear mode photon counting components and modules based on combining advanced APDs and advanced high gain circuits. By using HgCdTe APDs we enable Poisson number preserving photon counting. A metric of photon counting technology is dark count rate and detection probability. In this paper we report on a performance breakthrough resulting from improvement in design, process and readout operation enabling >10x reduction in dark counts rate to ~10,000 cps and >104x reduction in surface dark current enabling long 10 ms integration times. Our analysis of key dark current contributors suggest that substantial further reduction in DCR to ~ 1/sec or less can be achieved by optimizing wavelength, operating voltage and temperature.

  7. HEPS-BPIX, a single photon counting pixel detector with a high frame rate for the HEPS project

    Energy Technology Data Exchange (ETDEWEB)

    Wei, Wei, E-mail: weiw@ihep.ac.cn [Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049 (China); State Key Laboratory of Particle Detection and Electronics, Beijing 100049 (China); Zhang, Jie; Ning, Zhe; Lu, Yunpeng; Fan, Lei; Li, Huaishen; Jiang, Xiaoshan; Lan, Allan K.; Ouyang, Qun; Wang, Zheng; Zhu, Kejun; Chen, Yuanbo [Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049 (China); State Key Laboratory of Particle Detection and Electronics, Beijing 100049 (China); Liu, Peng [Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049 (China)

    2016-11-01

    China's next generation light source, named the High Energy Photon Source (HEPS), is currently under construction. HEPS-BPIX (HEPS-Beijing PIXel) is a dedicated pixel readout chip that operates in single photon counting mode for X-ray applications in HEPS. Designed using CMOS 0.13 µm technology, the chip contains a matrix of 104×72 pixels. Each pixel measures 150 µm×150 µm and has a counting depth of 20 bits. A bump-bonded prototyping detector module with a 300-µm thick silicon sensor was tested in the beamline of Beijing Synchrotron Radiation Facility. A fast stream of X-ray images was demonstrated, and a frame rate of 1.2 kHz was proven, with a negligible dead time. The test results showed an equivalent noise charge of 115 e{sup −} rms after bump bonding and a threshold dispersion of 55 e{sup −} rms after calibration.

  8. Nonparametric statistical inference

    CERN Document Server

    Gibbons, Jean Dickinson

    2014-01-01

    Thoroughly revised and reorganized, the fourth edition presents in-depth coverage of the theory and methods of the most widely used nonparametric procedures in statistical analysis and offers example applications appropriate for all areas of the social, behavioral, and life sciences. The book presents new material on the quantiles, the calculation of exact and simulated power, multiple comparisons, additional goodness-of-fit tests, methods of analysis of count data, and modern computer applications using MINITAB, SAS, and STATXACT. It includes tabular guides for simplified applications of tests and finding P values and confidence interval estimates.

  9. High absolute basophil count is a powerful independent predictor of inferior overall survival in patients with primary myelofibrosis.

    Science.gov (United States)

    Lucijanic, Marko; Livun, Ana; Stoos-Veic, Tajana; Pejsa, Vlatko; Jaksic, Ozren; Cicic, David; Lucijanic, Jelena; Romic, Zeljko; Orehovec, Biserka; Aralica, Gorana; Miletic, Marko; Kusec, Rajko

    2018-05-01

    To investigate the clinical and prognostic significance of absolute basophil count (ABC) in patients with primary myelofibrosis (PMF). We retrospectively investigated 58 patients with PMF treated in our institution in the period from 2006 to 2017. ABC was obtained in addition to other hematological and clinical parameters. Patients were separated into high and low ABC groups using the Receiver operating characteristic curve analysis. ABC was higher in PMF patients than in healthy controls (P constitutional symptoms (P = 0.030) and massive splenomegaly (P = 0.014). ABC was also positively correlated with absolute monocyte count (AMC) (P processes. Hence, both have a potential for improvement of current prognostic scores. Basophils represent a part of malignant clone in PMF and are associated with unfavorable disease features and poor prognosis which is independent of currently established prognostic scoring system and monocytosis.

  10. Comparison of probabilistic models of the distribution of counts

    International Nuclear Information System (INIS)

    Salma, I.; Zemplen-Papp, E.

    1992-01-01

    The binominal, Poisson and modified Poisson models for describing the statistical nature of the distribution of counts are compared theoretically, and conclusions for application are proposed. The validity of the Poisson and the modified Poisson distribution for observing k events in a short time interval is investigated experimentally for various measuring times. The experiments to measure the influence of the significant radioactive decay were performed with 89m Y (T 1/2 =16.06 s), using a multichannel analyser (4096 channels) in the multiscaling mode. According to the results, Poisson distribution describes the counting experiment for short measuring times (up to T=0.5 T 1/2 ) and its application is recommended. However, the analysis of the data demonstrated that for long measurements (T≥1 T 1/2 ) Poisson distribution is not valid and the modified Poisson distribution is preferable. The practical implications in calculating uncertainties and in optimizing the measuring time are discussed. (author) 20 refs.; 7 figs.; 1 tab

  11. Isospectral discrete and quantum graphs with the same flip counts and nodal counts

    Science.gov (United States)

    Juul, Jonas S.; Joyner, Christopher H.

    2018-06-01

    The existence of non-isomorphic graphs which share the same Laplace spectrum (to be referred to as isospectral graphs) leads naturally to the following question: what additional information is required in order to resolve isospectral graphs? It was suggested by Band, Shapira and Smilansky that this might be achieved by either counting the number of nodal domains or the number of times the eigenfunctions change sign (the so-called flip count) (Band et al 2006 J. Phys. A: Math. Gen. 39 13999–4014 Band and Smilansky 2007 Eur. Phys. J. Spec. Top. 145 171–9). Recent examples of (discrete) isospectral graphs with the same flip count and nodal count have been constructed by Ammann by utilising Godsil–McKay switching (Ammann private communication). Here, we provide a simple alternative mechanism that produces systematic examples of both discrete and quantum isospectral graphs with the same flip and nodal counts.

  12. The Top 100 Linked-To Pages on UK University Web Sites: High Inlink Counts Are Not Usually Associated with Quality Scholarly Content.

    Science.gov (United States)

    Thelwall, Mike

    2002-01-01

    Reports on an investigation into the most highly linked pages on United Kingdom university Web sites. Concludes that simple link counts are highly unreliable indicators of the average behavior of scholars, and that the most highly linked-to pages are those that facilitate access to a wide range of information rather than providing specific…

  13. A UNIFIED EMPIRICAL MODEL FOR INFRARED GALAXY COUNTS BASED ON THE OBSERVED PHYSICAL EVOLUTION OF DISTANT GALAXIES

    International Nuclear Information System (INIS)

    Béthermin, Matthieu; Daddi, Emanuele; Sargent, Mark T.; Elbaz, David; Mullaney, James; Pannella, Maurilio; Magdis, Georgios; Hezaveh, Yashar; Le Borgne, Damien; Buat, Véronique; Charmandaris, Vassilis; Lagache, Guilaine; Scott, Douglas

    2012-01-01

    We reproduce the mid-infrared to radio galaxy counts with a new empirical model based on our current understanding of the evolution of main-sequence (MS) and starburst (SB) galaxies. We rely on a simple spectral energy distribution (SED) library based on Herschel observations: a single SED for the MS and another one for SB, getting warmer with redshift. Our model is able to reproduce recent measurements of galaxy counts performed with Herschel, including counts per redshift slice. This agreement demonstrates the power of our 2-Star-Formation Modes (2SFM) decomposition in describing the statistical properties of infrared sources and their evolution with cosmic time. We discuss the relative contribution of MS and SB galaxies to the number counts at various wavelengths and flux densities. We also show that MS galaxies are responsible for a bump in the 1.4 GHz radio counts around 50 μJy. Material of the model (predictions, SED library, mock catalogs, etc.) is available online.

  14. Counting It Twice.

    Science.gov (United States)

    Schattschneider, Doris

    1991-01-01

    Provided are examples from many domains of mathematics that illustrate the Fubini Principle in its discrete version: the value of a summation over a rectangular array is independent of the order of summation. Included are: counting using partitions as in proof by pictures, combinatorial arguments, indirect counting as in the inclusion-exclusion…

  15. Low White Blood Cell Count

    Science.gov (United States)

    Symptoms Low white blood cell count By Mayo Clinic Staff A low white blood cell count (leukopenia) is a decrease ... of white blood cell (neutrophil). The definition of low white blood cell count varies from one medical ...

  16. A new method of quench monitoring in liquid scintillation counting

    International Nuclear Information System (INIS)

    Horrocks, D.L.

    1978-01-01

    The quench level of different liquid scintillation counting samples is measured by comparing the responses (pulse heights) produced by the same energy electrons in each sample. The electrons utilized in the measurements are those of the maximum energy (Esub(max)) which are produced by the single Compton scattering process for the same energy gamma-rays in each sample. The Esub(max) response produced in any sample is related to the Esub(max) response produced in an unquenched, sealed standard. The difference in response on a logarithm response scale is defined as the ''H Number''. The H number is related to the counting efficiency of the desired radionuclide by measurement of a set of standards of known amounts of the radionuclide and different amounts of quench (standard quench curve). The concept of the H number has been shown to be theoretically valid. Based upon this proof, the features of the H number concept as embodied in the Beckman LS-8000 Series Liquid Scintillation Systems have been demonstrated. It has been shown that one H number is unique; it provides a method of instrument calibration and wide dynamic quench range measurements. Further, it has been demonstrated that the H number concept provides a universal quench parameter. Counting efficiency vs. H number plots are repeatable within the statistical limits of +-1% counting efficiency. By the use of the H number concept a very accurate method of automatic quench compensation (A.Q.C.) is possible. (T.G.)

  17. Pile-up corrections for high-precision superallowed β decay half-life measurements via γ-ray photopeak counting

    Science.gov (United States)

    Grinyer, G. F.; Svensson, C. E.; Andreoiu, C.; Andreyev, A. N.; Austin, R. A. E.; Ball, G. C.; Bandyopadhyay, D.; Chakrawarthy, R. S.; Finlay, P.; Garrett, P. E.; Hackman, G.; Hyland, B.; Kulp, W. D.; Leach, K. G.; Leslie, J. R.; Morton, A. C.; Pearson, C. J.; Phillips, A. A.; Sarazin, F.; Schumaker, M. A.; Smith, M. B.; Valiente-Dobón, J. J.; Waddington, J. C.; Williams, S. J.; Wong, J.; Wood, J. L.; Zganjar, E. F.

    2007-09-01

    A general technique that corrects γ-ray gated β decay-curve data for detector pulse pile-up is presented. The method includes corrections for non-zero time-resolution and energy-threshold effects in addition to a special treatment of saturating events due to cosmic rays. This technique is verified through a Monte Carlo simulation and experimental data using radioactive beams of Na26 implanted at the center of the 8π γ-ray spectrometer at the ISAC facility at TRIUMF in Vancouver, Canada. The β-decay half-life of Na26 obtained from counting 1809-keV γ-ray photopeaks emitted by the daughter Mg26 was determined to be T=1.07167±0.00055 s following a 27σ correction for detector pulse pile-up. This result is in excellent agreement with the result of a previous measurement that employed direct β counting and demonstrates the feasibility of high-precision β-decay half-life measurements through the use of high-purity germanium γ-ray detectors. The technique presented here, while motivated by superallowed-Fermi β decay studies, is general and can be used for all half-life determinations (e.g. α-, β-, X-ray, fission) in which a γ-ray photopeak is used to select the decays of a particular isotope.

  18. AgNOR Count in Resting Cells (Resting NOR Is a New Prognostic Marker in Invasive Bladder Tumor

    Directory of Open Access Journals (Sweden)

    Mitsuro Tomobe

    2001-01-01

    Full Text Available Purpose: We have previously demonstrated that the AgNOR count in proliferating cells is a predictor of tumor recurrence in superficial bladder tumor (J. Urol. 162 (1999, 63–68. In the present study, we evaluate the type of AgNOR associated with cell cycles as a prognostic factor in invasive bladder tumor using a double staining technique employing both AgNOR and MIB-1 labelling. Materials and methods: Forty-four paraffin sections of invasive bladder tumors were stained simultaneously with AgNOR and MIB-1. The number of AgNORs in proliferating (MIB-1 positive or resting (MIB-1 negative cells were counted from a total of 100 nuclei. Correlations between MIB-1 associated AgNOR count and clinicopathological parameters were statistically analyzed. Results: The AgNOR count in proliferating cells (proliferating NOR was significantly higher than that in resting cells (resting NOR (p < 0.01. The resting NOR in tumors with distant metastases was significantly higher than that in tumors without metastases (p < 0.05. Patients with a low resting NOR tumor had a better prognosis than those with a high resting NOR tumor, whereas the proliferating NOR was not associated with survival. Survival analysis revealed that the resting NOR was the most powerful prognostic marker in patients with invasive bladder tumor (p < 0.05. Conclusions: Resting NOR had a predictive value in the prognosis of patients with invasive bladder tumor. Keywords: Transitional cell carcinoma, invasive, resting cell, AgNORs, MIB-1

  19. Grade Assignment by Ki-67 Proliferative Index, Mitotic Count, and Phosphohistone H3 Count in Surgically Resected Gastrointestinal and Pancreatic Neuroendocrine Tumors.

    Science.gov (United States)

    Murphy, Claire E; McCormick, Kinsey A; Shankaran, Veena; Reddi, Deepti M; Swanson, Paul E; Upton, Melissa P; Papanicolau-Sengos, Antonios; Khor, Sara; Westerhoff, Maria

    The aim of this study was to evaluate the concordance in grade assignment for gastroenteropancreatic neuroendocrine tumors using mitotic count (MC), Ki-67 proliferative index (KPI), and phosphohistone H3 count (PHH3C). Resected gastroenteropancreatic neuroendocrine tumors were graded based on MC, KPI, and PHH3C. Concordance was determined using a weighted κ statistic. Median survival across each grade category was determined using Kaplan-Meier methods. Of the 110 patients, the majority had gastrointestinal primaries and grade 1 or 2 tumors. Rates of discordance in grade assignment were 29% of cases for KPI versus MC (κW = 0.26), 32% for PHH3C versus MC (κW = 0.34), and 32% for PHH3C versus KPI (κW = 0.37). There was fair agreement between grading by KPI and MC. Relative to grade by KPI and MC, PHH3C tended to upgrade tumors. The proportion alive at 3 and 5 years was not significantly different for patients with grade 1 versus grade 2 tumors. The concordance between KPI and MC was fair. Phosphohistone H3 count tended to upgrade tumors using the cutoffs established by MC. Grade 1 and grade 2 tumors were associated with similar survival regardless of grading method. The overall relevance of the current cutoff values used in grading neuroendocrine tumors may need to be revisited.

  20. Counting Better? An Examination of the Impact of Quantitative Method Teaching on Statistical Anxiety and Confidence

    Science.gov (United States)

    Chamberlain, John Martyn; Hillier, John; Signoretta, Paola

    2015-01-01

    This article reports the results of research concerned with students' statistical anxiety and confidence to both complete and learn to complete statistical tasks. Data were collected at the beginning and end of a quantitative method statistics module. Students recognised the value of numeracy skills but felt they were not necessarily relevant for…

  1. Modeling Zero-Inflated and Overdispersed Count Data: An Empirical Study of School Suspensions

    Science.gov (United States)

    Desjardins, Christopher David

    2016-01-01

    The purpose of this article is to develop a statistical model that best explains variability in the number of school days suspended. Number of school days suspended is a count variable that may be zero-inflated and overdispersed relative to a Poisson model. Four models were examined: Poisson, negative binomial, Poisson hurdle, and negative…

  2. Reliability of spring interconnects for high channel-count polyimide electrode arrays

    Science.gov (United States)

    Khan, Sharif; Ordonez, Juan Sebastian; Stieglitz, Thomas

    2018-05-01

    Active neural implants with a high channel-count need robust and reliable operational assembly for the targeted environment in order to be classified as viable fully implantable systems. The discrete functionality of the electrode array and the implant electronics is vital for intact assembly. A critical interface exists at the interconnection sites between the electrode array and the implant electronics, especially in hybrid assemblies (e.g. retinal implants) where electrodes and electronics are not on the same substrate. Since the interconnects in such assemblies cannot be hermetically sealed, reliable protection against the physiological environment is essential for delivering high insulation resistance and low defusibility of salt ions, which are limited in complexity by current assembly techniques. This work reports on a combination of spring-type interconnects on a polyimide array with silicone rubber gasket insulation for chronically active implantable systems. The spring design of the interconnects on the backend of the electrode array compensates for the uniform thickness of the sandwiched gasket during bonding in assembly and relieves the propagation of extrinsic stresses to the bulk polyimide substrate. The contact resistance of the microflex-bonded spring interconnects with the underlying metallized ceramic test vehicles and insulation through the gasket between adjacent contacts was investigated against the MIL883 standard. The contact and insulation resistances remained stable in the exhausting environmental conditions.

  3. Mu-Spec - A High Performance Ultra-Compact Photon Counting spectrometer for Space Submillimeter Astronomy

    Science.gov (United States)

    Moseley, H.; Hsieh, W.-T.; Stevenson, T.; Wollack, E.; Brown, A.; Benford, D.; Sadleir; U-Yen, I.; Ehsan, N.; Zmuidzinas, J.; hide

    2011-01-01

    We have designed and are testing elements of a fully integrated submillimeter spectrometer based on superconducting microstrip technology. The instrument can offer resolving power R approximately 1500, and its high frequency cutoff is set by the gap of available high performance superconductors. All functions of the spectrometer are integrated - light is coupled to the microstrip circuit with a planar antenna, the spectra discrimination is achieved using a synthetic grating, orders are separated using planar filter, and detected using photon counting MKID detector. This spectrometer promises to revolutionize submillimeter spectroscopy from space. It replaces instruments with the scale of 1m with a spectrometer on a 10 cm Si wafer. The reduction in mass and volume promises a much higher performance system within available resource in a space mission. We will describe the system and the performance of the components that have been fabricated and tested.

  4. Experimental investigation of statistical density function of decaying radioactive sources

    International Nuclear Information System (INIS)

    Salma, I.; Zemplen-Papp, E.

    1991-01-01

    The validity of the Poisson and the λ P(k) modified Poisson statistical density functions of observing k events in a short time interval is investigated experimentally in radioactive decay detection for various measuring times. The experiments to measure radioactive decay were performed with 89m Y, using a multichannel analyzer. According to the results, Poisson statistics adequately describes the counting experiment for short measuring times. (author) 13 refs.; 4 figs

  5. New method for eliminating the statistical bias in highly turbulent flow measurements

    International Nuclear Information System (INIS)

    Nakao, S.I.; Terao, Y.; Hirata, K.I.; Kitakyushu Industrial Research Institute, Fukuoka, Japan)

    1987-01-01

    A simple method was developed for eliminating statistical bias which can be applied to highly turbulent flows with the sparse and nonuniform seeding conditions. Unlike the method proposed so far, a weighting function was determined based on the idea that the statistical bias could be eliminated if the asymmetric form of the probability density function of the velocity data were corrected. Moreover, the data more than three standard deviations away from the mean were discarded to remove the apparent turbulent intensity resulting from noise. The present method was applied to data obtained in the wake of a block, which provided local turbulent intensities up to about 120 percent, it was found to eliminate the statistical bias with high accuracy. 9 references

  6. Automatic Counting of Large Mammals from Very High Resolution Panchromatic Satellite Imagery

    NARCIS (Netherlands)

    Xue, Yifei; Wang, Tiejun; Skidmore, Andrew K.

    2017-01-01

    Estimating animal populations by direct counting is an essential component of wildlife conservation and management. However, conventional approaches (i.e., ground survey and aerial survey) have intrinsic constraints. Advances in image data capture and processing provide new opportunities for using

  7. Embodied finger counting in children with different cultural backgrounds and hand dominance

    Directory of Open Access Journals (Sweden)

    Liutsko L.

    2017-12-01

    Full Text Available Background. Embodied finger counting has been shown to have cross-cultural differences in previous studies (Lindemann, Alipour, & Fisher, 2011; Soto & Lalain, 2008. However, their results were contradictory in reference to Western populations with regard to the hand preferred: The first study showed that in Western countries — Europe and the United States — participants preferred to start with the left hand (whereas in the Middle East — Iran — they used the right hand; the second study showed that participants in France preferred the right hand. Objective. Our study aimed to observe these differences in two countries, Spain (Western Europe and Russia (Eastern Europe part, although taking into account the variety of cultural or ethnic groups who live there. Design. The observational/descriptive study, together with correlational analysis of the finger-counting pattern (from 1 to 10 used by children aged 10 to 12 who had not been taught to use their fingers for counting, considered factors of cultural origin and hand dominance. The possible effects of this action on cognition — in our case, math achievement — were considered also. Results and conclusion. The differences in the frequency of the finger-counting patterns might suggest cultural-individual differences in performance; however, the correlational analysis did not reveal that these differences were statistically significant, either for gender or for mark in math. However, hand dominance was a significant predictor of the preferred hand with which to start counting.

  8. Free parameter, figure of merit and ionization quench in liquid scintillation counting

    International Nuclear Information System (INIS)

    Carles, P. Grau; Malonda, A. Grau

    2001-01-01

    A statistical study of the detection process demonstrates that the free parameter is essential to compute the counting efficiency in both CIEMAT/NIST and TDCR methods. An analysis of the computed counting efficiencies shows the uselessness of old definition of the figure of merit. A new definition is required and we adopt the idea of taking quantities related with the output of the photomultiplier. In addition, we justify the application of the chemical quenching simulation with the electronic variation of the photomultiplier gain. Finally, we describe a new procedure to determine the figure of merit and the optimum ionization-quenching factor from the pulse spectrum of different radionuclides. The robustness of the new procedure is tested with three different sets of stopping power for low-energy electrons

  9. Comparative evaluation of platelet count and antimicrobial efficacy of injectable platelet-rich fibrin with other platelet concentrates: An in vitro study

    Directory of Open Access Journals (Sweden)

    Prerna Ashok Karde

    2017-01-01

    Full Text Available Background: Platelet concentrates are used in various medical procedures to promote soft- and hard-tissue regeneration. In recent times, their antimicrobial efficacy is also explored. However, various platelet concentrates have evolved which differ in the centrifugation protocols. One such recently introduced platelet concentrate is injectable platelet-rich fibrin (i-PRF concentrate. Hence, the aim was to evaluate the antimicrobial property, and platelet count of i-PRF in comparison to other platelet concentrates, i.e., PRF, platelet-rich plasma (PRP, and control (whole blood. Materials and Methods: Blood samples were obtained from 10 chronic generalized marginal gingivitis patients. Platelet concentrates were prepared using standardized centrifugation protocol. Platelet count was evaluated by manual counting method using smear preparation of each sample. Subsequently, antimicrobial activity against oral bacteria was examined on blood agar using disc diffusion method to quantify the inhibitory effects. Results: Statistical significance was analyzed by one-way analysis of variance (ANOVA. P 0.05. i-PRF showed statistically significant difference (P < 0.001 in platelet count when compared to control. It was also significant when compared to PRP (P < 0.01, PRF (P < 0.001. Conclusion: i-PRF has maximum antimicrobial efficacy and higher platelet count in comparison to other platelet concentrates, thereby indicating to have a better regenerative potential then others.

  10. High performance statistical computing with parallel R: applications to biology and climate modelling

    International Nuclear Information System (INIS)

    Samatova, Nagiza F; Branstetter, Marcia; Ganguly, Auroop R; Hettich, Robert; Khan, Shiraj; Kora, Guruprasad; Li, Jiangtian; Ma, Xiaosong; Pan, Chongle; Shoshani, Arie; Yoginath, Srikanth

    2006-01-01

    Ultrascale computing and high-throughput experimental technologies have enabled the production of scientific data about complex natural phenomena. With this opportunity, comes a new problem - the massive quantities of data so produced. Answers to fundamental questions about the nature of those phenomena remain largely hidden in the produced data. The goal of this work is to provide a scalable high performance statistical data analysis framework to help scientists perform interactive analyses of these raw data to extract knowledge. Towards this goal we have been developing an open source parallel statistical analysis package, called Parallel R, that lets scientists employ a wide range of statistical analysis routines on high performance shared and distributed memory architectures without having to deal with the intricacies of parallelizing these routines

  11. Hanford whole body counting manual

    International Nuclear Information System (INIS)

    Palmer, H.E.; Brim, C.P.; Rieksts, G.A.; Rhoads, M.C.

    1987-05-01

    This document, a reprint of the Whole Body Counting Manual, was compiled to train personnel, document operation procedures, and outline quality assurance procedures. The current manual contains information on: the location, availability, and scope of services of Hanford's whole body counting facilities; the administrative aspect of the whole body counting operation; Hanford's whole body counting facilities; the step-by-step procedure involved in the different types of in vivo measurements; the detectors, preamplifiers and amplifiers, and spectroscopy equipment; the quality assurance aspect of equipment calibration and recordkeeping; data processing, record storage, results verification, report preparation, count summaries, and unit cost accounting; and the topics of minimum detectable amount and measurement accuracy and precision. 12 refs., 13 tabs

  12. Thyroid dysfunction in human immunodeficiency virus-infected children and its correlation with CD4 + T lymphocyte count

    Directory of Open Access Journals (Sweden)

    Satyakumar Thongam

    2015-01-01

    Full Text Available Context: Thyroid dysfunction has been reported in human immunodeficiency virus (HIV-infected individuals including children. Some studies have reported that thyroid dysfunction may be a marker of severity or progression of HIV. Aims: The aim was to study thyroid function in HIV-infected children with and without highly active anti-retroviral therapy (HAART. Settings and Design: Cross-sectional study carried out at a teaching hospital with Anti-Retroviral Therapy Centre (Centre of Excellence of National AIDS Control Organization. Subjects and Methods: Thyroid stimulating hormone (TSH, total thyroxine (T4, and total tri-iodothyronine (T3 were analyzed in 60 pediatric HIV cases: 30 on HAART and 30 HAART naive. Correlation of T3, T4, and TSH with CD4 count was assessed. Statistical Analysis Used: Data reported as mean ± standard deviation and as the number of cases and percentages. Comparison between groups was done by independent sample t-test and χ2 -test. Spearman′s correlation coefficient is used to assess the association between thyroid dysfunction and CD4 count. Results: Thyroid function abnormality was seen in five out of 30 patients in both patients on HAART or without HAART therapy. Among patients on HAART, three had hypothyroidism, and two had biochemical feature of sick euthyroid syndrome. Among the HAART naive group, sub-clinical hypothyroisim was seen in four, and one had biochemical feature of sick euthyroid syndrome. None of the patients had clinical features of thyroid dysfunction. There is a highly significant correlation (P = 0.01 between TSH and CD4 count. Conclusions: Thyroid dysfunction is quite common among pediatric HIV cases. An inverse correlation is seen between TSH and CD4 count indicating trend for hypothyroidism as HIV disease progress.

  13. A flowrate measurement method by counting of radioactive particles suspended in a liquid

    International Nuclear Information System (INIS)

    Daniel, G.

    1983-04-01

    By external counting of fine #betta# emitting radioactive particles suspended in a liquid, the flowrate in a system of pipes can be measured. The study comprises three phases: 1. - The hydraulic validity of the method is demonstrated in laminar as well as in turbulent flow under certain conditions of particles size and density and of liquid viscosity. 2. - Radioactive labelling of microspheres of serumalbumin or ion exchange resins with indium 113m delivered by a generator Tin 113 → Indium 113m. 3. - Counting with a scintillation detector: a method of threshold overstepping is experimented with a mechanical or electronic simulator; the statistical study of particle superposition under the detector enables a correction for the resulting counting losses to be proposed. The method provides absolute measurements, but is particularly suitable to measure relative flowrates in a hydraulic network. It can be continuous and does not perturb the flow and the network. The accuracy of the method is analysed in details [fr

  14. 'Intelligent' approach to radioimmunoassay sample counting employing a microprocessor controlled sample counter

    International Nuclear Information System (INIS)

    Ekins, R.P.; Sufi, S.; Malan, P.G.

    1977-01-01

    The enormous impact on medical science in the last two decades of microanalytical techniques employing radioisotopic labels has, in turn, generated a large demand for automatic radioisotopic sample counters. Such instruments frequently comprise the most important item of capital equipment required in the use of radioimmunoassay and related techniques and often form a principle bottleneck in the flow of samples through a busy laboratory. It is therefore particularly imperitive that such instruments should be used 'intelligently' and in an optimal fashion to avoid both the very large capital expenditure involved in the unnecessary proliferation of instruments and the time delays arising from their sub-optimal use. The majority of the current generation of radioactive sample counters nevertheless rely on primitive control mechanisms based on a simplistic statistical theory of radioactive sample counting which preclude their efficient and rational use. The fundamental principle upon which this approach is based is that it is useless to continue counting a radioactive sample for a time longer than that required to yield a significant increase in precision of the measurement. Thus, since substantial experimental errors occur during sample preparation, these errors should be assessed and must be releted to the counting errors for that sample. It is the objective of this presentation to demonstrate that the combination of a realistic statistical assessment of radioactive sample measurement, together with the more sophisticated control mechanisms that modern microprocessor technology make possible, may often enable savings in counter usage of the order of 5-10 fold to be made. (orig.) [de

  15. Single electron counting using a dual MCP assembly

    International Nuclear Information System (INIS)

    Yang, Yuzhen; Liu, Shulin; Zhao, Tianchi; Yan, Baojun; Wang, Peiliang; Yu, Yang; Lei, Xiangcui; Yang, Luping; Wen, Kaile; Qi, Ming

    2016-01-01

    The gain, pulse height resolution and peak-to-valley ratio of single electrons detected by using a Chevron configured Microchannel Plate (MCP) assembly are studied. The two MCPs are separated by a 280 µm gap and are biased by four electrodes. The purpose of the study is to determine the optimum bias voltage arrangements for single electron counting. By comparing the results of various bias voltage combinations, we conclude that good performance for the electron counting can be achieved by operating the MCP assembly in saturation mode. In addition, by applying a small reverse bias voltage across the gap while adjusting the bias voltages of the MCPs, optimum performance of electron counting can be obtained. - Highlights: • Dual MCPs assembly with four electrodes using different voltage combinations has been investigated for single electron counting. • Both the MCP voltages and the gap voltage can affect the gain, pulse height resolution and P/V ratio. • A high gain of the first stage MCP, a saturation mode of the second stage MCP and an appropriately reverse gap voltage can improve the resolution greatly. • The optimum voltage arrangements is significant for the design of MCP detectors in single electron counting applications.

  16. Kansas Kids Count Data Book, 1998. A Project of Kansas Action for Children.

    Science.gov (United States)

    Kansas Action for Children, Inc., Topeka.

    The Kids Count Data Book examines trends in the well-being of Kansas children. The statistical portrait is based on trends in 22 indicators of child well-being, grouped into 5 areas: (1) economic well-being--births to single teens, free school meals, family economic assistance, child poverty rates; (2) physical health and safety--childhood deaths,…

  17. Kansas Kids Count Data Book, 1999. A Project of Kansas Action for Children.

    Science.gov (United States)

    Kansas Action for Children, Inc., Topeka.

    The Kids Count Data Book examines trends in the well-being of Kansas children. This statistical portrait is based on trends in 22 indicators of child well-being, grouped into 5 areas: (1) economic well-being--births to single teens, child poverty rates, free school meals, and family economic assistance; (2) physical health and safety--childhood…

  18. On Counting the Rational Numbers

    Science.gov (United States)

    Almada, Carlos

    2010-01-01

    In this study, we show how to construct a function from the set N of natural numbers that explicitly counts the set Q[superscript +] of all positive rational numbers using a very intuitive approach. The function has the appeal of Cantor's function and it has the advantage that any high school student can understand the main idea at a glance…

  19. The Fallacy of Word-Counts.

    Science.gov (United States)

    Engels, L.K.

    1968-01-01

    The greatest fallacy of word counts, the author maintains, lies in the fact that advocates of frequency lists stress the high percentage without telling the whole truth. It has become common to pretend that a frequency list of 3,000 words covers 95 percent of the language, that it enables a person to speak and understand a foreign language by…

  20. Precise method for correcting count-rate losses in scintillation cameras

    International Nuclear Information System (INIS)

    Madsen, M.T.; Nickles, R.J.

    1986-01-01

    Quantitative studies performed with scintillation detectors often require corrections for lost data because of the finite resolving time of the detector. Methods that monitor losses by means of a reference source or pulser have unacceptably large statistical fluctuations associated with their correction factors. Analytic methods that model the detector as a paralyzable system require an accurate estimate of the system resolving time. Because the apparent resolving time depends on many variables, including the window setting, source distribution, and the amount of scattering material, significant errors can be introduced by relying on a resolving time obtained from phantom measurements. These problems can be overcome by curve-fitting the data from a reference source to a paralyzable model in which the true total count rate in the selected window is estimated from the observed total rate. The resolving time becomes a free parameter in this method which is optimized to provide the best fit to the observed reference data. The fitted curve has the inherent accuracy of the reference source method with the precision associated with the observed total image count rate. Correction factors can be simply calculated from the ratio of the true reference source rate and the fitted curve. As a result, the statistical uncertainty of the data corrected by this method is not significantly increased

  1. Design Study of an Incinerator Ash Conveyor Counting System - 13323

    International Nuclear Information System (INIS)

    Jaederstroem, Henrik; Bronson, Frazier

    2013-01-01

    A design study has been performed for a system that should measure the Cs-137 activity in ash from an incinerator. Radioactive ash, expected to consist of both Cs-134 and Cs-137, will be transported on a conveyor belt at 0.1 m/s. The objective of the counting system is to determine the Cs-137 activity and direct the ash to the correct stream after a diverter. The decision levels are ranging from 8000 to 400000 Bq/kg and the decision error should be as low as possible. The decision error depends on the total measurement uncertainty which depends on the counting statistics and the uncertainty in the efficiency of the geometry. For the low activity decision it is necessary to know the efficiency to be able to determine if the signal from the Cs-137 is above the minimum detectable activity and that it generates enough counts to reach the desired precision. For the higher activity decision the uncertainty of the efficiency needs to be understood to minimize decision errors. The total efficiency of the detector is needed to be able to determine if the detector will be able operate at the count rate at the highest expected activity. The design study that is presented in this paper describes how the objectives of the monitoring systems were obtained, the choice of detector was made and how ISOCS (In Situ Object Counting System) mathematical modeling was used to calculate the efficiency. The ISOCS uncertainty estimator (IUE) was used to determine which parameters of the ash was important to know accurately in order to minimize the uncertainty of the efficiency. The examined parameters include the height of the ash on the conveyor belt, the matrix composition and density and relative efficiency of the detector. (authors)

  2. Design Study of an Incinerator Ash Conveyor Counting System - 13323

    Energy Technology Data Exchange (ETDEWEB)

    Jaederstroem, Henrik; Bronson, Frazier [Canberra Industries Inc., 800 Research Parkway Meriden CT 06450 (United States)

    2013-07-01

    A design study has been performed for a system that should measure the Cs-137 activity in ash from an incinerator. Radioactive ash, expected to consist of both Cs-134 and Cs-137, will be transported on a conveyor belt at 0.1 m/s. The objective of the counting system is to determine the Cs-137 activity and direct the ash to the correct stream after a diverter. The decision levels are ranging from 8000 to 400000 Bq/kg and the decision error should be as low as possible. The decision error depends on the total measurement uncertainty which depends on the counting statistics and the uncertainty in the efficiency of the geometry. For the low activity decision it is necessary to know the efficiency to be able to determine if the signal from the Cs-137 is above the minimum detectable activity and that it generates enough counts to reach the desired precision. For the higher activity decision the uncertainty of the efficiency needs to be understood to minimize decision errors. The total efficiency of the detector is needed to be able to determine if the detector will be able operate at the count rate at the highest expected activity. The design study that is presented in this paper describes how the objectives of the monitoring systems were obtained, the choice of detector was made and how ISOCS (In Situ Object Counting System) mathematical modeling was used to calculate the efficiency. The ISOCS uncertainty estimator (IUE) was used to determine which parameters of the ash was important to know accurately in order to minimize the uncertainty of the efficiency. The examined parameters include the height of the ash on the conveyor belt, the matrix composition and density and relative efficiency of the detector. (authors)

  3. Statistical inference for a class of multivariate negative binomial distributions

    DEFF Research Database (Denmark)

    Rubak, Ege Holger; Møller, Jesper; McCullagh, Peter

    This paper considers statistical inference procedures for a class of models for positively correlated count variables called α-permanental random fields, and which can be viewed as a family of multivariate negative binomial distributions. Their appealing probabilistic properties have earlier been...

  4. Alberta oil and gas industry: annual statistics for 1995

    International Nuclear Information System (INIS)

    1996-01-01

    Statistical data for 1995 concerning the supply and disposition of crude oil and equivalent, natural gas, ethane, butanes, natural gas liquids, and sulphur in the Province of Alberta, were provided. A list of new wells drilled during 1995, and an annual well count, were also included

  5. CalCOFI Egg Counts

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Fish egg counts and standardized counts for eggs captured in CalCOFI icthyoplankton nets (primarily vertical [Calvet or Pairovet], oblique [bongo or ring nets], and...

  6. Agreement of manual cell counts and automated counts of the scil Vet abc Plus(+) hematology analyzer for analysis of equine synovial fluid.

    Science.gov (United States)

    Van de Water, Eline; Oosterlinck, Maarten; Duchateau, Luc; Pille, Frederik

    2016-06-01

    The purpose of this study was to determine whether the scil Vet abc Plus(+) (SCIL Animal Care Company, Altorf, France), an impedance hematology analyzer, can accurately quantify and differentiate nucleated blood cells (NBCs) in equine synovial fluid. Synovial fluid samples (n=242) in different stages of experimentally induced inflammation were analyzed with and without hyaluronidase pretreatment and compared to manual hemocytometer counts and smear reviews. No significant effect of hyaluronidase pretreatment was observed. Total nucleated cell counts of the scil Vet abc Plus(+) were significantly higher compared to the manual method (P=0.02), yet the difference was small and clinically irrelevant (ratio manual/automated count equal to 0.97 with 95% CI [0.95, 1.00]). Differential cell counts of the scil Vet abc Plus(+) were not accurate. In conclusion, the scil Vet abc Plus(+) hematology analyzer is highly accurate for quantification, but not accurate for differentiation of NBCs in equine synovial fluid. Copyright © 2016 Elsevier Ltd. All rights reserved.

  7. Count-doubling time safety circuit

    International Nuclear Information System (INIS)

    Keefe, D.J.; McDowell, W.P.; Rusch, G.K.

    1981-01-01

    There is provided a nuclear reactor count-factor-increase time monitoring circuit which includes a pulse-type neutron detector, and means for counting the number of detected pulses during specific time periods. Counts are compared and the comparison is utilized to develop a reactor scram signal, if necessary

  8. Count-doubling time safety circuit

    Science.gov (United States)

    Rusch, Gordon K.; Keefe, Donald J.; McDowell, William P.

    1981-01-01

    There is provided a nuclear reactor count-factor-increase time monitoring circuit which includes a pulse-type neutron detector, and means for counting the number of detected pulses during specific time periods. Counts are compared and the comparison is utilized to develop a reactor scram signal, if necessary.

  9. Statistical inference of level densities from resolved resonance parameters

    International Nuclear Information System (INIS)

    Froehner, F.H.

    1983-08-01

    Level densities are most directly obtained by counting the resonances observed in the resolved resonance range. Even in the measurements, however, weak levels are invariably missed so that one has to estimate their number and add it to the raw count. The main categories of missinglevel estimators are discussed in the present review, viz. (I) ladder methods including those based on the theory of Hamiltonian matrix ensembles (Dyson-Mehta statistics), (II) methods based on comparison with artificial cross section curves (Monte Carlo simulation, Garrison's autocorrelation method), (III) methods exploiting the observed neutron width distribution by means of Bayesian or more approximate procedures such as maximum-likelihood, least-squares or moment methods, with various recipes for the treatment of detection thresholds and resolution effects. The language of mathematical statistics is employed to clarify the basis of, and the relationship between, the various techniques. Recent progress in the treatment of resolution effects, detection thresholds and p-wave admixture is described. (orig.) [de

  10. Construction and Test of Muon Drift Tube Chambers for High Counting Rates

    CERN Document Server

    Schwegler, Philipp; Dubbert, Jörg

    2010-01-01

    Since the start of operation of the Large Hadron Collider (LHC) at CERN on 20 November 2009, the instantaneous luminosity is steadily increasing. The muon spectrometer of the ATLAS detector at the LHC is instrumented with trigger and precision tracking chambers in a toroidal magnetic field. Monitored Drift-Tube (MDT) chambers are employed as precision tracking chambers, complemented by Cathode Strip Chambers (CSC) in the very forward region where the background counting rate due to neutrons and γ's produced in shielding material and detector components is too high for the MDT chambers. After several upgrades of the CERN accelerator system over the coming decade, the instantaneous luminosity is expected to be raised to about five times the LHC design luminosity. This necessitates replacement of the muon chambers in the regions with the highest background radiation rates in the so-called Small Wheels, which constitute the innermost layers of the muon spectrometer end-caps, by new detectors with higher rate cap...

  11. The Big Pumpkin Count.

    Science.gov (United States)

    Coplestone-Loomis, Lenny

    1981-01-01

    Pumpkin seeds are counted after students convert pumpkins to jack-o-lanterns. Among the activities involved, pupils learn to count by 10s, make estimates, and to construct a visual representation of 1,000. (MP)

  12. A Near-Infrared Photon Counting Camera for High Sensitivity Astronomical Observation, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — The innovation is a Near Infrared Photon-Counting Sensor (NIRPCS), an imaging device with sufficient sensitivity to capture the spectral signatures, in the...

  13. A Near-Infrared Photon Counting Camera for High Sensitivity Astronomical Observation, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — The innovation is a Near Infrared Photon-Counting Sensor (NIRPCS), an imaging device with sufficient sensitivity to capture the spectral signatures, in the...

  14. How Managed Care Affects Medicaid Utilization : A Synthetic Difference-in-Difference Zero-Inflated Count Model

    NARCIS (Netherlands)

    Freund, D.A.; Kniesner, T.J.; LoSasso, A.T.

    1996-01-01

    We develop a synthetic difference-in-differences statistical design to apply to experimental data for adult women living in Hennepin County, Minnesota, to estimate the impact of Medicaid managed care on various modes of medical care use.Because the outcomes of interest are utilization counts with

  15. Confidence Intervals for Asbestos Fiber Counts: Approximate Negative Binomial Distribution.

    Science.gov (United States)

    Bartley, David; Slaven, James; Harper, Martin

    2017-03-01

    The negative binomial distribution is adopted for analyzing asbestos fiber counts so as to account for both the sampling errors in capturing only a finite number of fibers and the inevitable human variation in identifying and counting sampled fibers. A simple approximation to this distribution is developed for the derivation of quantiles and approximate confidence limits. The success of the approximation depends critically on the use of Stirling's expansion to sufficient order, on exact normalization of the approximating distribution, on reasonable perturbation of quantities from the normal distribution, and on accurately approximating sums by inverse-trapezoidal integration. Accuracy of the approximation developed is checked through simulation and also by comparison to traditional approximate confidence intervals in the specific case that the negative binomial distribution approaches the Poisson distribution. The resulting statistics are shown to relate directly to early research into the accuracy of asbestos sampling and analysis. Uncertainty in estimating mean asbestos fiber concentrations given only a single count is derived. Decision limits (limits of detection) and detection limits are considered for controlling false-positive and false-negative detection assertions and are compared to traditional limits computed assuming normal distributions. Published by Oxford University Press on behalf of the British Occupational Hygiene Society 2017.

  16. Cherenkov radiation effects on counting efficiency in extremely quenched liquid scintillation samples

    International Nuclear Information System (INIS)

    Grau Carles, A.; Grau Malonda, A.; Rodriguez Barquero, L.

    1993-01-01

    The CIEMAT/NIST tracer method has successfully standardized nuclides with diverse quench values and decay schemes in liquid scintillation counting. However, the counting efficiency is computed inaccurately for extremely quenched samples. This article shows that when samples are extremely quenched, the counting efficiency in high-energy beta-ray nuclides depends principally on the Cherenkov effect. A new technique is described for quench determination, which makes the measurement of counting efficiency possible when scintillation counting approaches zero. A new efficiency computation model for pure beta-ray nuclides is also described. The results of the model are tested experimentally for 89 Sr, 90 Y, 36 Cl and 204 Tl nuclides with independence of the quench level. (orig.)

  17. Statistical Analysis and Time Series Modeling of Air Traffic Operations Data From Flight Service Stations and Terminal Radar Approach Control Facilities : Two Case Studies

    Science.gov (United States)

    1981-10-01

    Two statistical procedures have been developed to estimate hourly or daily aircraft counts. These counts can then be transformed into estimates of instantaneous air counts. The first procedure estimates the stable (deterministic) mean level of hourly...

  18. Examination of statistical noise in SPECT image and sampling pitch

    International Nuclear Information System (INIS)

    Takaki, Akihiro; Soma, Tsutomu; Murase, Kenya; Watanabe, Hiroyuki; Murakami, Tomonori; Kawakami, Kazunori; Teraoka, Satomi; Kojima, Akihiro; Matsumoto, Masanori

    2008-01-01

    Statistical noise in single photon emission computed tomography (SPECT) image was examined for its relation with total count and with sampling pitch by simulation and phantom experiment to obtain their projection data under defined conditions. The former SPECT simulation was performed on assumption of a virtual, homogeneous water column (20 cm diameter) as an absorbing mass. In the latter, used were 3D-Hoffman brain phantom (Data Spectrum Corp.) filled with 370 MBq of 99m Tc-pertechnetate solution and a facing 2-detector SPECT machine with a low-energy/high-resolution collimator, E-CAM (Siemens). Projected data by the two methods were reconstructed through the filtered back projection to make each transaxial image. The noise was evaluated by vision, by their root mean square uncertainty calculated from average count and standard deviation (SD) in the region of interest (ROI) defined in reconstructed images and by normalized mean squares calculated from the difference between the reference image obtained with common sampling pitch to and all of obtained slices of, the simulation and phantom. As a conclusion, the pitch was recommended to be set in the machine as to approximating the value calculated by the sampling theorem, though the projection counts per one angular direction were smaller with the same total time of data acquisition. (R.T.)

  19. Effect of warm-smoking on total microbial count of meat products

    Directory of Open Access Journals (Sweden)

    A Javadi

    2007-11-01

    Full Text Available The frankfurters are amongst the most famous and popular sausages in the world and beef and poultry meat are used in Iran for their preparation. The techniques of warm smoking at 42°c for two hours and then hot smoking together with steam cooking at 8°c for one hour are utilized in proportion of this product. In spite of its carcinogenic properties, smoke is used to create color, flavor and odor and to improve the preservative qualities of sausages. In this study, 14 sausage samples were taken from each of the stages of frankfurter production line including pre-smoking, post- warm smoking and post-hot smoking, their total microbial counts (aerobic mesophiles determined and the means of the three stages compared using the ANOVA statistical test. The results indicated that the total microbial count increased significantly (P

  20. Recent progress with digital coincidence counting

    International Nuclear Information System (INIS)

    Butcher, K.S.A.; Watt, G.C.; Alexiev, D.

    1999-01-01

    Digital Coincidence Counting (DCC) is a new technique, based on the older method of analogue coincidence counting. It has been developed by ANSTO as a faster more reliable means of determining the activity of ionising radiation samples. The technique employs a dual channel analogue to digital converter acquisition system for collecting pulse information from a 4Π beta detector and a NaI(Tl) gamma detector. The digitised pulse information is stored on a high speed hard disk and timing information for both channels is also stored. The data may subsequently be recalled and analysed using software based algorithms. The system is operational and results are now being routinely collected and analysed. Some of the early work is presented for Co-60, Na-22 and Sm-153

  1. The choice between Ruark-Devol and Poisson statistics in the measurement of short-lived radionuclides

    International Nuclear Information System (INIS)

    Kennedy, G.

    1981-01-01

    The question of whether to use Poisson or Ruark-Devol statistics for radioactivity measurements in which the counting time is long compared to the half-life is discussed. Experimental data are presented which are well described by Poisson statistics. The applications of Ruark-Devol statistics are found to be very limited, in disagreement with earlier publications. (author)

  2. Characterization of a new computer-ready photon counting system

    Science.gov (United States)

    Andor, Gyorgy

    1998-08-01

    The photon-counting system seems to be the best solution for extremely low optical power measurements. The Hamamatsu HC135 photon counting module has a built-in high-voltage power supply amplifier, discriminator, micro-controller with an RS232 serial output. It requires only a +5V supply voltage and an IBM PC or compatible computer to run. The system is supplied with an application software. This talk is about the testing of the device.

  3. Galaxy number counts: Pt. 2

    International Nuclear Information System (INIS)

    Metcalfe, N.; Shanks, T.; Fong, R.; Jones, L.R.

    1991-01-01

    Using the Prime Focus CCD Camera at the Isaac Newton Telescope we have determined the form of the B and R galaxy number-magnitude count relations in 12 independent fields for 21 m ccd m and 19 m ccd m 5. The average galaxy count relations lie in the middle of the wide range previously encompassed by photographic data. The field-to-field variation of the counts is small enough to define the faint (B m 5) galaxy count to ±10 per cent and this variation is consistent with that expected from galaxy clustering considerations. Our new data confirm that the B, and also the R, galaxy counts show evidence for strong galaxy luminosity evolution, and that the majority of the evolving galaxies are of moderately blue colour. (author)

  4. Statistical Origin of Black Hole Entropy in Matrix Theory

    International Nuclear Information System (INIS)

    Lowe, D.A.

    1998-01-01

    The statistical entropy of black holes in matrix theory is considered. Assuming matrix theory is the discretized light-cone quantization of a theory with eleven-dimensional Lorentz invariance, we map the counting problem onto the original Gibbons-Hawking calculations of the thermodynamic entropy. copyright 1998 The American Physical Society

  5. Multiplicity counting from fission chamber signals in the current mode

    Energy Technology Data Exchange (ETDEWEB)

    Pázsit, I. [Chalmers University of Technology, Department of Physics, Division of Subatomic and Plasma Physics, SE-412 96 Göteborg (Sweden); Pál, L. [Centre for Energy Research, Hungarian Academy of Sciences, 114, POB 49, H-1525 Budapest (Hungary); Nagy, L. [Chalmers University of Technology, Department of Physics, Division of Subatomic and Plasma Physics, SE-412 96 Göteborg (Sweden); Budapest University of Technology and Economics, Institute of Nuclear Techniques, H-1111 Budapest (Hungary)

    2016-12-11

    In nuclear safeguards, estimation of sample parameters using neutron-based non-destructive assay methods is traditionally based on multiplicity counting with thermal neutron detectors in the pulse mode. These methods in general require multi-channel analysers and various dead time correction methods. This paper proposes and elaborates on an alternative method, which is based on fast neutron measurements with fission chambers in the current mode. A theory of “multiplicity counting” with fission chambers is developed by incorporating Böhnel's concept of superfission [1] into a master equation formalism, developed recently by the present authors for the statistical theory of fission chamber signals [2,3]. Explicit expressions are derived for the first three central auto- and cross moments (cumulants) of the signals of up to three detectors. These constitute the generalisation of the traditional Campbell relationships for the case when the incoming events represent a compound Poisson distribution. Because now the expressions contain the factorial moments of the compound source, they contain the same information as the singles, doubles and triples rates of traditional multiplicity counting. The results show that in addition to the detector efficiency, the detector pulse shape also enters the formulas; hence, the method requires a more involved calibration than the traditional method of multiplicity counting. However, the method has some advantages by not needing dead time corrections, as well as having a simpler and more efficient data processing procedure, in particular for cross-correlations between different detectors, than the traditional multiplicity counting methods.

  6. Variations in neutrophil count in preterm infants with respiratory distress syndrome who subsequently developed chronic lung disease.

    Science.gov (United States)

    Kohelet, D; Arbel, E; Ballin, A; Goldberg, M

    2000-01-01

    Neutrophil counts were studied in 62 preterm infants receiving mechanical ventilation for neonatal respiratory distress syndrome (NRDS). Exploratory analysis indicated that the severity of NRDS, as demonstrated by fractional inspired oxygen (FiO2), mean airway pressure (MAP), arterial-alveolar PO2 ratio (a/APO2) and oxygenation index (OI), was correlated with percentage change of neutrophil counts during the first 5 days of life. Further analysis demonstrated that infants with NRDS who subsequently developed chronic lung disease (CLD) (n = 21) had statistically significant differences in variation of neutrophil counts when compared with the remainder (n = 41) without CLD (-35.0% +/- 4.3 vs. -16.9% +/- 5.8, p variations in neutrophil counts during the first 5 days of life may be found in infants with NRDS who subsequently develop CLD and that these changes may have predictive value regarding the development of CLD.

  7. A Scan Statistic for Continuous Data Based on the Normal Probability Model

    OpenAIRE

    Konty, Kevin; Kulldorff, Martin; Huang, Lan

    2009-01-01

    Abstract Temporal, spatial and space-time scan statistics are commonly used to detect and evaluate the statistical significance of temporal and/or geographical disease clusters, without any prior assumptions on the location, time period or size of those clusters. Scan statistics are mostly used for count data, such as disease incidence or mortality. Sometimes there is an interest in looking for clusters with respect to a continuous variable, such as lead levels in children or low birth weight...

  8. Gating circuit for single photon-counting fluorescence lifetime instruments using high repetition pulsed light sources

    International Nuclear Information System (INIS)

    Laws, W.R.; Potter, D.W.; Sutherland, J.C.

    1984-01-01

    We have constructed a circuit that permits conventional timing electronics to be used in single photon-counting fluorimeters with high repetition rate excitation sources (synchrotrons and mode-locked lasers). Most commercial time-to-amplitude and time-to-digital converters introduce errors when processing very short time intervals and when subjected to high-frequency signals. This circuit reduces the frequency of signals representing the pulsed light source (stops) to the rate of detected fluorescence events (starts). Precise timing between the start/stop pair is accomplished by using the second stop pulse after a start pulse. Important features of our design are that the circuit is insensitive to the simultaneous occurrence of start and stop signals and that the reduction in the stop frequency allows the start/stop time interval to be placed in linear regions of the response functions of commercial timing electronics

  9. Analysis of overdispersed count data: application to the Human Papillomavirus Infection in Men (HIM) Study.

    Science.gov (United States)

    Lee, J-H; Han, G; Fulp, W J; Giuliano, A R

    2012-06-01

    The Poisson model can be applied to the count of events occurring within a specific time period. The main feature of the Poisson model is the assumption that the mean and variance of the count data are equal. However, this equal mean-variance relationship rarely occurs in observational data. In most cases, the observed variance is larger than the assumed variance, which is called overdispersion. Further, when the observed data involve excessive zero counts, the problem of overdispersion results in underestimating the variance of the estimated parameter, and thus produces a misleading conclusion. We illustrated the use of four models for overdispersed count data that may be attributed to excessive zeros. These are Poisson, negative binomial, zero-inflated Poisson and zero-inflated negative binomial models. The example data in this article deal with the number of incidents involving human papillomavirus infection. The four models resulted in differing statistical inferences. The Poisson model, which is widely used in epidemiology research, underestimated the standard errors and overstated the significance of some covariates.

  10. Correlation between total lymphocyte count, hemoglobin, hematocrit and CD4 count in HIV patients in Nigeria.

    Science.gov (United States)

    Emuchay, Charles Iheanyichi; Okeniyi, Shemaiah Olufemi; Okeniyi, Joshua Olusegun

    2014-04-01

    The expensive and technology limited setting of CD4 count testing is a major setback to the initiation of HAART in a resource limited country like Nigeria. Simple and inexpensive tools such as Hemoglobin (Hb) measurement and Total Lymphocyte Count (TLC) are recommended as substitute marker. In order to assess the correlations of these parameters with CD4 count, 100 "apparently healthy" male volunteers tested HIV positive aged ≥ 20 years but ≤ 40 years were recruited and from whom Hb, Hct, TLC and CD4 count were obtained. The correlation coefficients, R, the Nash-Sutcliffe Coefficient of Efficiency (CoE) and the p-values of the ANOVA model of Hb, Hct and TLC with CD4 count were assessed. The assessments show that there is no significant relationship of any of these parameters with CD4 count and the correlation coefficients are very weak. This study shows that Hb, Hct and TLC cannot be substitute for CD4 count as this might lead to certain individuals' deprivation of required treatment.

  11. A Statistical Treatment of Bioassay Pour Fractions

    Science.gov (United States)

    Barengoltz, Jack; Hughes, David W.

    2014-01-01

    The binomial probability distribution is used to treat the statistics of a microbiological sample that is split into two parts, with only one part evaluated for spore count. One wishes to estimate the total number of spores in the sample based on the counts obtained from the part that is evaluated (pour fraction). Formally, the binomial distribution is recharacterized as a function of the observed counts (successes), with the total number (trials) an unknown. The pour fraction is the probability of success per spore (trial). This distribution must be renormalized in terms of the total number. Finally, the new renormalized distribution is integrated and mathematically inverted to yield the maximum estimate of the total number as a function of a desired level of confidence ( P(fraction. The extension to recovery efficiency corrections is also presented. Now the product of recovery efficiency and pour fraction may be small enough that the likely value may be much larger than the usual calculation: the number of spores divided by that product. The use of this analysis would not be limited to microbiological data.

  12. It counts who counts: an experimental evaluation of the importance of observer effects on spotlight count estimates

    DEFF Research Database (Denmark)

    Sunde, Peter; Jessen, Lonnie

    2013-01-01

    observers with respect to their ability to detect and estimate distance to realistic animal silhouettes at different distances. Detection probabilities were higher for observers experienced in spotlighting mammals than for inexperienced observers, higher for observers with a hunting background compared...... with non-hunters and decreased as function of age but were independent of sex or educational background. If observer-specific detection probabilities were applied to real counting routes, point count estimates from inexperienced observers without a hunting background would only be 43 % (95 % CI, 39...

  13. You can count on the motor cortex: Finger counting habits modulate motor cortex activation evoked by numbers

    Science.gov (United States)

    Tschentscher, Nadja; Hauk, Olaf; Fischer, Martin H.; Pulvermüller, Friedemann

    2012-01-01

    The embodied cognition framework suggests that neural systems for perception and action are engaged during higher cognitive processes. In an event-related fMRI study, we tested this claim for the abstract domain of numerical symbol processing: is the human cortical motor system part of the representation of numbers, and is organization of numerical knowledge influenced by individual finger counting habits? Developmental studies suggest a link between numerals and finger counting habits due to the acquisition of numerical skills through finger counting in childhood. In the present study, digits 1 to 9 and the corresponding number words were presented visually to adults with different finger counting habits, i.e. left- and right-starters who reported that they usually start counting small numbers with their left and right hand, respectively. Despite the absence of overt hand movements, the hemisphere contralateral to the hand used for counting small numbers was activated when small numbers were presented. The correspondence between finger counting habits and hemispheric motor activation is consistent with an intrinsic functional link between finger counting and number processing. PMID:22133748

  14. Meningiomas: Objective assessment of proliferative indices by immunohistochemistry and automated counting method.

    Science.gov (United States)

    Chavali, Pooja; Uppin, Megha S; Uppin, Shantveer G; Challa, Sundaram

    2017-01-01

    The most reliable histological correlate of recurrence risk in meningiomas is increased mitotic activity. Proliferative index with Ki-67 immunostaining is a helpful adjunct to manual counting. However, both show considerable inter-observer variability. A new immunohistochemical method for counting mitotic figures, using antibody against the phosphohistone H3 (PHH3) protein was introduced. Similarly, a computer based automated counting for Ki-67 labelling index (LI) is available. To study the use of these new techniques in the objective assessment of proliferation indices in meningiomas. This was a retrospective study of intracranial meningiomas diagnosed during the year 2013.The hematoxylin and eosin (H and E) sections and immunohistochemistry (IHC) with Ki-67 were reviewed by two pathologists. Photomicrographs of the representative areas were subjected to Ki-67 analysis by Immunoratio (IR) software. Mean Ki-67 LI, both manual and by IR were calculated. IHC with PHH3 was performed. PHH3 positive nuclei were counted and mean values calculated. Data analysis was done using SPSS software. A total of 64 intracranial meningiomas were diagnosed. Evaluation on H and E, PHH3, Ki-67 LI (both manual and IR) were done in 32 cases (22 grade I and 10 grade II meningiomas). Statistically significant correlation was seen between the mitotic count in each grade and PHH3 values and also between the grade of the tumor and values of Ki-67 and PHH3. Both the techniques used in the study had advantage over, as well as, correlated well with the existing techniques and hence, can be applied to routine use.

  15. Comparison of McMaster and FECPAKG2 methods for counting nematode eggs in the faeces of alpacas.

    Science.gov (United States)

    Rashid, Mohammed H; Stevenson, Mark A; Waenga, Shea; Mirams, Greg; Campbell, Angus J D; Vaughan, Jane L; Jabbar, Abdul

    2018-05-02

    This study aimed to compare the FECPAK G2 and the McMaster techniques for counting of gastrointestinal nematode eggs in the faeces of alpacas using two floatation solutions (saturated sodium chloride and sucrose solutions). Faecal eggs counts from both techniques were compared using the Lin's concordance correlation coefficient and Bland and Altman statistics. Results showed moderate to good agreement between the two methods, with better agreement achieved when saturated sugar is used as a floatation fluid, particularly when faecal egg counts are less than 1000 eggs per gram of faeces. To the best of our knowledge this is the first study to assess agreement of measurements between McMaster and FECPAK G2 methods for estimating faecal eggs in South American camelids.

  16. Total lymphocyte count and subpopulation lymphocyte counts in relation to dietary intake and nutritional status of peritoneal dialysis patients.

    Science.gov (United States)

    Grzegorzewska, Alicja E; Leander, Magdalena

    2005-01-01

    Dietary deficiency causes abnormalities in circulating lymphocyte counts. For the present paper, we evaluated correlations between total and subpopulation lymphocyte counts (TLC, SLCs) and parameters of nutrition in peritoneal dialysis (PD) patients. Studies were carried out in 55 patients treated with PD for 22.2 +/- 11.4 months. Parameters of nutritional status included total body mass, lean body mass (LBM), body mass index (BMI), and laboratory indices [total protein, albumin, iron, ferritin, and total iron binding capacity (TIBC)]. The SLCs were evaluated using flow cytometry. Positive correlations were seen between TLC and dietary intake of niacin; TLC and CD8 and CD16+56 counts and energy delivered from protein; CD4 count and beta-carotene and monounsaturated fatty acids 17:1 intake; and CD19 count and potassium, copper, vitamin A, and beta-carotene intake. Anorexia negatively influenced CD19 count. Serum albumin showed correlations with CD4 and CD19 counts, and LBM with CD19 count. A higher CD19 count was connected with a higher red blood cell count, hemoglobin, and hematocrit. Correlations were observed between TIBC and TLC and CD3 and CD8 counts, and between serum Fe and TLC and CD3 and CD4 counts. Patients with a higher CD19 count showed a better clinical-laboratory score, especially less weakness. Patients with a higher CD4 count had less expressed insomnia. Quantities of ingested vitamins and minerals influence lymphocyte counts in the peripheral blood of PD patients. Evaluation of TLC and SLCs is helpful in monitoring the effectiveness of nutrition in these patients.

  17. Eulerian and Lagrangian statistics from high resolution numerical simulations of weakly compressible turbulence

    NARCIS (Netherlands)

    Benzi, R.; Biferale, L.; Fisher, R.T.; Lamb, D.Q.; Toschi, F.

    2009-01-01

    We report a detailed study of Eulerian and Lagrangian statistics from high resolution Direct Numerical Simulations of isotropic weakly compressible turbulence. Reynolds number at the Taylor microscale is estimated to be around 600. Eulerian and Lagrangian statistics is evaluated over a huge data

  18. Risk factors associated with bulk tank standard plate count, bulk tank coliform count, and the presence of Staphylococcus aureus on organic and conventional dairy farms in the United States.

    Science.gov (United States)

    Cicconi-Hogan, K M; Gamroth, M; Richert, R; Ruegg, P L; Stiglbauer, K E; Schukken, Y H

    2013-01-01

    The purpose of this study was to assess the association of bulk tank milk standard plate counts, bulk tank coliform counts (CC), and the presence of Staphylococcus aureus in bulk tank milk with various management and farm characteristics on organic and conventional dairy farms throughout New York, Wisconsin, and Oregon. Data from size-matched organic farms (n=192), conventional nongrazing farms (n=64), and conventional grazing farms (n=36) were collected at a single visit for each farm. Of the 292 farms visited, 290 bulk tank milk samples were collected. Statistical models were created using data from all herds in the study, as well as exclusively for the organic subset of herds. Because of incomplete data, 267 of 290 herds were analyzed for total herd modeling, and 173 of 190 organic herds were analyzed for the organic herd modeling. Overall, more bulk tanks from organic farms had Staph. aureus cultured from them (62% of organic herds, 42% conventional nongrazing herds, and 43% of conventional grazing herds), whereas fewer organic herds had a high CC, defined as ≥50 cfu/mL, than conventional farms in the study. A high standard plate count (×1,000 cfu/mL) was associated with decreased body condition score of adult cows and decreased milk production in both models. Several variables were significant only in the model created using all herds or only in organic herds. The presence of Staph. aureus in the bulk tank milk was associated with fewer people treating mastitis, increased age of housing, and a higher percentage of cows with 3 or fewer teats in both the organic and total herd models. The Staph. aureus total herd model also showed a relationship with fewer first-lactation animals, higher hock scores, and less use of automatic takeoffs at milking. High bulk tank CC was related to feeding a total mixed ration and using natural service in nonlactating heifers in both models. Overall, attentive management and use of outside resources were useful with regard to CC

  19. Developments in statistical analysis in quantitative genetics

    DEFF Research Database (Denmark)

    Sorensen, Daniel

    2009-01-01

    of genetic means and variances, models for the analysis of categorical and count data, the statistical genetics of a model postulating that environmental variance is partly under genetic control, and a short discussion of models that incorporate massive genetic marker information. We provide an overview......A remarkable research impetus has taken place in statistical genetics since the last World Conference. This has been stimulated by breakthroughs in molecular genetics, automated data-recording devices and computer-intensive statistical methods. The latter were revolutionized by the bootstrap...... and by Markov chain Monte Carlo (McMC). In this overview a number of specific areas are chosen to illustrate the enormous flexibility that McMC has provided for fitting models and exploring features of data that were previously inaccessible. The selected areas are inferences of the trajectories over time...

  20. An investigation of the trade-off between the count level and image quality in myocardial perfusion SPECT using simulated images: the effects of statistical noise and object variability on defect detectability

    International Nuclear Information System (INIS)

    He Xin; Links, Jonathan M; Frey, Eric C

    2010-01-01

    Quantum noise as well as anatomic and uptake variability in patient populations limits observer performance on a defect detection task in myocardial perfusion SPECT (MPS). The goal of this study was to investigate the relative importance of these two effects by varying acquisition time, which determines the count level, and assessing the change in performance on a myocardial perfusion (MP) defect detection task using both mathematical and human observers. We generated ten sets of projections of a simulated patient population with count levels ranging from 1/128 to around 15 times a typical clinical count level to simulate different levels of quantum noise. For the simulated population we modeled variations in patient, heart and defect size, heart orientation and shape, defect location, organ uptake ratio, etc. The projection data were reconstructed using the OS-EM algorithm with no compensation or with attenuation, detector response and scatter compensation (ADS). The images were then post-filtered and reoriented to generate short-axis slices. A channelized Hotelling observer (CHO) was applied to the short-axis images, and the area under the receiver operating characteristics (ROC) curve (AUC) was computed. For each noise level and reconstruction method, we optimized the number of iterations and cutoff frequencies of the Butterworth filter to maximize the AUC. Using the images obtained with the optimal iteration and cutoff frequency and ADS compensation, we performed human observer studies for four count levels to validate the CHO results. Both CHO and human observer studies demonstrated that observer performance was dependent on the relative magnitude of the quantum noise and the patient variation. When the count level was high, the patient variation dominated, and the AUC increased very slowly with changes in the count level for the same level of anatomic variability. When the count level was low, however, quantum noise dominated, and changes in the count level

  1. The prognostic value of angiogenesis by Chalkley counting in a confirmatory study design on 836 breast cancer patients

    DEFF Research Database (Denmark)

    Hansen, S; Grabau, D A; Sørensen, Flemming Brandt

    2000-01-01

    This study addresses the prognostic value of estimating angiogenesis by Chalkley counting in breast cancer. A population-based group consisting of 836 patients with operated primary, unilateral invasive breast carcinomas was included from a predefined region and period of time. The median follow...... categories using predefined Chalkley cutoff points at five and seven. There were significant correlations between high Chalkley counts and axillary lymph node metastasis, large tumor size, high histological malignancy grade, and histological type. A high Chalkley count showed lower probabilities...... or =7 compared with counts between 5-7. The study confirmed that estimation of angiogenesis by Chalkley counting had independent prognostic value in breast cancer patients. The Chalkley count could be useful to stratify node-negative patients for adjuvant...

  2. Pyogenic arthritis, pyoderma gangrenosum, and acne (PAPA) syndrome: differential diagnosis of septic arthritis by regular detection of exceedingly high synovial cell counts.

    Science.gov (United States)

    Löffler, W; Lohse, P; Weihmayr, T; Widenmayer, W

    2017-08-01

    Pyogenic arthritis, pyoderma gangrenosum and acne syndrome was diagnosed in a 42-year-old patient, after an unusual persistency of high synovial cell counts had been noticed. Clinical peculiarities and problems with diagnosing septic versus non-septic arthritis are discussed.

  3. Statistical mechanics of complex neural systems and high dimensional data

    International Nuclear Information System (INIS)

    Advani, Madhu; Lahiri, Subhaneil; Ganguli, Surya

    2013-01-01

    Recent experimental advances in neuroscience have opened new vistas into the immense complexity of neuronal networks. This proliferation of data challenges us on two parallel fronts. First, how can we form adequate theoretical frameworks for understanding how dynamical network processes cooperate across widely disparate spatiotemporal scales to solve important computational problems? Second, how can we extract meaningful models of neuronal systems from high dimensional datasets? To aid in these challenges, we give a pedagogical review of a collection of ideas and theoretical methods arising at the intersection of statistical physics, computer science and neurobiology. We introduce the interrelated replica and cavity methods, which originated in statistical physics as powerful ways to quantitatively analyze large highly heterogeneous systems of many interacting degrees of freedom. We also introduce the closely related notion of message passing in graphical models, which originated in computer science as a distributed algorithm capable of solving large inference and optimization problems involving many coupled variables. We then show how both the statistical physics and computer science perspectives can be applied in a wide diversity of contexts to problems arising in theoretical neuroscience and data analysis. Along the way we discuss spin glasses, learning theory, illusions of structure in noise, random matrices, dimensionality reduction and compressed sensing, all within the unified formalism of the replica method. Moreover, we review recent conceptual connections between message passing in graphical models, and neural computation and learning. Overall, these ideas illustrate how statistical physics and computer science might provide a lens through which we can uncover emergent computational functions buried deep within the dynamical complexities of neuronal networks. (paper)

  4. A fast and high-sensitive dual-wavelength diffuse optical tomography system using digital lock-in photon-counting technique

    Science.gov (United States)

    Chen, Weiting; Yi, Xi; Zhao, Huijuan; Gao, Feng

    2014-09-01

    We presented a novel dual-wavelength diffuse optical imaging system which can perform 2-D or 3-D imaging fast and high-sensitively for monitoring the dynamic change of optical parameters. A newly proposed lock-in photon-counting detection method was adopted for week optical signal collection, which brought in excellent property as well as simplified geometry. Fundamental principles of the lock-in photon-counting detection were elaborately demonstrated, and the feasibility was strictly verified by the linearity experiment. Systemic performance of the prototype set up was experimentally accessed, including stray light rejection and inherent interference. Results showed that the system possessed superior anti-interference capability (under 0.58% in darkroom) compared with traditional photon-counting detection, and the crosstalk between two wavelengths was lower than 2.28%. For comprehensive assessment, 2-D phantom experiments towards relatively large dimension model (diameter of 4cm) were conducted. Different absorption targets were imaged to investigate detection sensitivity. Reconstruction image under all conditions was exciting, with a desirable SNR. Study on image quality v.s. integration time put forward a new method for accessing higher SNR with the sacrifice of measuring speed. In summary, the newly developed system showed great potential in promoting detection sensitivity as well as measuring speed. This will make substantial progress in dynamically tracking the blood concentration distribution in many clinical areas, such as small animal disease modeling, human brain activity research and thick tissues (for example, breast) diagnosis.

  5. Regression analysis of mixed recurrent-event and panel-count data with additive rate models.

    Science.gov (United States)

    Zhu, Liang; Zhao, Hui; Sun, Jianguo; Leisenring, Wendy; Robison, Leslie L

    2015-03-01

    Event-history studies of recurrent events are often conducted in fields such as demography, epidemiology, medicine, and social sciences (Cook and Lawless, 2007, The Statistical Analysis of Recurrent Events. New York: Springer-Verlag; Zhao et al., 2011, Test 20, 1-42). For such analysis, two types of data have been extensively investigated: recurrent-event data and panel-count data. However, in practice, one may face a third type of data, mixed recurrent-event and panel-count data or mixed event-history data. Such data occur if some study subjects are monitored or observed continuously and thus provide recurrent-event data, while the others are observed only at discrete times and hence give only panel-count data. A more general situation is that each subject is observed continuously over certain time periods but only at discrete times over other time periods. There exists little literature on the analysis of such mixed data except that published by Zhu et al. (2013, Statistics in Medicine 32, 1954-1963). In this article, we consider the regression analysis of mixed data using the additive rate model and develop some estimating equation-based approaches to estimate the regression parameters of interest. Both finite sample and asymptotic properties of the resulting estimators are established, and the numerical studies suggest that the proposed methodology works well for practical situations. The approach is applied to a Childhood Cancer Survivor Study that motivated this study. © 2014, The International Biometric Society.

  6. Factor V Leiden is associated with increased sperm count.

    Science.gov (United States)

    van Mens, T E; Joensen, U N; Bochdanovits, Z; Takizawa, A; Peter, J; Jørgensen, N; Szecsi, P B; Meijers, J C M; Weiler, H; Rajpert-De Meyts, E; Repping, S; Middeldorp, S

    2017-11-01

    Is the thrombophilia mutation factor V Leiden (FVL) associated with an increased total sperm count? Carriers of FVL have a higher total sperm count than non-FVL-carriers, which could not be explained by genetic linkage or by observations in a FVL-mouse model. FVL has a high prevalence in Caucasians despite detrimental health effects. Carriers have been shown to have higher fecundity, which might partly explain this evolutionary paradox. We determined FVL status in two cohorts (Dutch, n = 627; Danish, n = 854) of consecutively included men without known causes for spermatogenic failure, and performed an individual patient data meta-analysis of these two cohorts together with one previously published (Dutch, n = 908) cohort. We explored possible biological underpinnings for the relation between sperm count and FVL, by use of a FVL-mouse model and investigations of genetic linkage. Participants were male partners of subfertile couples (two Dutch cohorts) and young men from the general population (Danish cohort): FVL carrier rate was 4.0%, 4.6% and 7.3%, respectively. There were differences in smoking, abstinence time and age between the cohorts. We corrected for these in the primary analysis, which consisted of a mixed linear effects model, also incorporating unobjectified population differences. In public haplotype data from subjects of European descent, we explored linkage disequilibrium of FVL with all known single nucleotide polymorphisms in a 1.5 MB region around the F5 gene with an R2 cutoff of 0.8. We sequenced exons of four candidate genes hypothesized to be linked to FVL in a subgroup of FVL carriers with extreme sperm count values. The animal studies consisted of never mated 15-18-week-old C57BL/J6 mice heterozygous and homozygous for FVL and wild-type mice. We compared spermatogenesis parameters (normalized internal genitalia weights, epididymis sperm content and sperm motility) between FVL and wild-type mice. Human FVL carriers have a higher total sperm

  7. An 'intelligent' approach to radioimmunoassay sample counting employing a microprocessor-controlled sample counter

    International Nuclear Information System (INIS)

    Ekins, R.P.; Sufi, S.; Malan, P.G.

    1978-01-01

    The enormous impact on medical science in the last two decades of microanalytical techniques employing radioisotopic labels has, in turn, generated a large demand for automatic radioisotopic sample counters. Such instruments frequently comprise the most important item of capital equipment required in the use of radioimmunoassay and related techniques and often form a principle bottleneck in the flow of samples through a busy laboratory. It is therefore imperative that such instruments should be used 'intelligently' and in an optimal fashion to avoid both the very large capital expenditure involved in the unnecessary proliferation of instruments and the time delays arising from their sub-optimal use. Most of the current generation of radioactive sample counters nevertheless rely on primitive control mechanisms based on a simplistic statistical theory of radioactive sample counting which preclude their efficient and rational use. The fundamental principle upon which this approach is based is that it is useless to continue counting a radioactive sample for a time longer than that required to yield a significant increase in precision of the measurement. Thus, since substantial experimental errors occur during sample preparation, these errors should be assessed and must be related to the counting errors for that sample. The objective of the paper is to demonstrate that the combination of a realistic statistical assessment of radioactive sample measurement, together with the more sophisticated control mechanisms that modern microprocessor technology make possible, may often enable savings in counter usage of the order of 5- to 10-fold to be made. (author)

  8. Automatic vehicle counting system for traffic monitoring

    Science.gov (United States)

    Crouzil, Alain; Khoudour, Louahdi; Valiere, Paul; Truong Cong, Dung Nghy

    2016-09-01

    The article is dedicated to the presentation of a vision-based system for road vehicle counting and classification. The system is able to achieve counting with a very good accuracy even in difficult scenarios linked to occlusions and/or presence of shadows. The principle of the system is to use already installed cameras in road networks without any additional calibration procedure. We propose a robust segmentation algorithm that detects foreground pixels corresponding to moving vehicles. First, the approach models each pixel of the background with an adaptive Gaussian distribution. This model is coupled with a motion detection procedure, which allows correctly location of moving vehicles in space and time. The nature of trials carried out, including peak periods and various vehicle types, leads to an increase of occlusions between cars and between cars and trucks. A specific method for severe occlusion detection, based on the notion of solidity, has been carried out and tested. Furthermore, the method developed in this work is capable of managing shadows with high resolution. The related algorithm has been tested and compared to a classical method. Experimental results based on four large datasets show that our method can count and classify vehicles in real time with a high level of performance (>98%) under different environmental situations, thus performing better than the conventional inductive loop detectors.

  9. A new framework of statistical inferences based on the valid joint sampling distribution of the observed counts in an incomplete contingency table.

    Science.gov (United States)

    Tian, Guo-Liang; Li, Hui-Qiong

    2017-08-01

    Some existing confidence interval methods and hypothesis testing methods in the analysis of a contingency table with incomplete observations in both margins entirely depend on an underlying assumption that the sampling distribution of the observed counts is a product of independent multinomial/binomial distributions for complete and incomplete counts. However, it can be shown that this independency assumption is incorrect and can result in unreliable conclusions because of the under-estimation of the uncertainty. Therefore, the first objective of this paper is to derive the valid joint sampling distribution of the observed counts in a contingency table with incomplete observations in both margins. The second objective is to provide a new framework for analyzing incomplete contingency tables based on the derived joint sampling distribution of the observed counts by developing a Fisher scoring algorithm to calculate maximum likelihood estimates of parameters of interest, the bootstrap confidence interval methods, and the bootstrap testing hypothesis methods. We compare the differences between the valid sampling distribution and the sampling distribution under the independency assumption. Simulation studies showed that average/expected confidence-interval widths of parameters based on the sampling distribution under the independency assumption are shorter than those based on the new sampling distribution, yielding unrealistic results. A real data set is analyzed to illustrate the application of the new sampling distribution for incomplete contingency tables and the analysis results again confirm the conclusions obtained from the simulation studies.

  10. THE STATISTICS OF RADIO ASTRONOMICAL POLARIMETRY: BRIGHT SOURCES AND HIGH TIME RESOLUTION

    International Nuclear Information System (INIS)

    Van Straten, W.

    2009-01-01

    A four-dimensional statistical description of electromagnetic radiation is developed and applied to the analysis of radio pulsar polarization. The new formalism provides an elementary statistical explanation of the modal-broadening phenomenon in single-pulse observations. It is also used to argue that the degree of polarization of giant pulses has been poorly defined in past studies. Single- and giant-pulse polarimetry typically involves sources with large flux-densities and observations with high time-resolution, factors that necessitate consideration of source-intrinsic noise and small-number statistics. Self-noise is shown to fully explain the excess polarization dispersion previously noted in single-pulse observations of bright pulsars, obviating the need for additional randomly polarized radiation. Rather, these observations are more simply interpreted as an incoherent sum of covariant, orthogonal, partially polarized modes. Based on this premise, the four-dimensional covariance matrix of the Stokes parameters may be used to derive mode-separated pulse profiles without any assumptions about the intrinsic degrees of mode polarization. Finally, utilizing the small-number statistics of the Stokes parameters, it is established that the degree of polarization of an unresolved pulse is fundamentally undefined; therefore, previous claims of highly polarized giant pulses are unsubstantiated.

  11. Bayesian statistical evaluation of peak area measurements in gamma spectrometry

    International Nuclear Information System (INIS)

    Silva, L.; Turkman, A.; Paulino, C.D.

    2010-01-01

    We analyze results from determinations of peak areas for a radioactive source containing several radionuclides. The statistical analysis was performed using Bayesian methods based on the usual Poisson model for observed counts. This model does not appear to be a very good assumption for the counting system under investigation, even though it is not questioned as a whole by the inferential procedures adopted. We conclude that, in order to avoid incorrect inferences on relevant quantities, one must proceed to a further study that allows us to include missing influence parameters and to select a model explaining the observed data much better.

  12. Statistical issues in searches for new phenomena in High Energy Physics

    Science.gov (United States)

    Lyons, Louis; Wardle, Nicholas

    2018-03-01

    Many analyses of data in High Energy Physics are concerned with searches for New Physics. We review the statistical issues that arise in such searches, and then illustrate these using the specific example of the recent successful search for the Higgs boson, produced in collisions between high energy protons at CERN’s Large Hadron Collider.

  13. Statistical Mechanics and Black Hole Thermodynamics

    OpenAIRE

    Carlip, Steven

    1997-01-01

    Black holes are thermodynamic objects, but despite recent progress, the ultimate statistical mechanical origin of black hole temperature and entropy remains mysterious. Here I summarize an approach in which the entropy is viewed as arising from ``would-be pure gauge'' degrees of freedom that become dynamical at the horizon. For the (2+1)-dimensional black hole, these degrees of freedom can be counted, and yield the correct Bekenstein-Hawking entropy; the corresponding problem in 3+1 dimension...

  14. CERN{sub D}xCTA counting mode chip

    Energy Technology Data Exchange (ETDEWEB)

    Moraes, D. [CERN, CH-1211 Geneva 23 (Switzerland)], E-mail: danielle.moraes@cern.ch; Kaplon, J. [CERN, CH-1211 Geneva 23 (Switzerland); Nygard, E. [Interon AS, Asker, Norway and DX-ray Inc., Northridge, CA (United States)

    2008-06-11

    This ASIC is a counting mode front-end electronic optimized for the readout of CdZnTe/CdTe and silicon sensors, for possible use in applications where the flux of ionizing radiation is high. The chip is implemented in 0.25 {mu}m CMOS technology. The circuit comprises 128 channels equipped with a transimpedance amplifier followed by a gain shaper stage with 21 ns peaking time, two discriminators and two 18-bit counters. The channel architecture is optimized for the detector characteristics in order to achieve the best energy resolution at counting rates of up to 5 M counts/second. The amplifier shows a linear sensitivity of 118 mV/fC and an equivalent noise charge of about 711 e{sup -}, for a detector capacitance of 5 pF. Complete evaluation of the circuit is presented using electronic pulses and pixel detectors.

  15. LAWRENCE RADIATION LABORATORY COUNTING HANDBOOK

    Energy Technology Data Exchange (ETDEWEB)

    Group, Nuclear Instrumentation

    1966-10-01

    The Counting Handbook is a compilation of operational techniques and performance specifications on counting equipment in use at the Lawrence Radiation Laboratory, Berkeley. Counting notes have been written from the viewpoint of the user rather than that of the designer or maintenance man. The only maintenance instructions that have been included are those that can easily be performed by the experimenter to assure that the equipment is operating properly.

  16. Evaluation for asbestos exposure in lung cancer surgery cases. Relationships between asbestos body count and pleural plaques and between asbestos body count and pulmonary fibrosis

    International Nuclear Information System (INIS)

    Yusa, Toshikazu; Yasukawa, Tomohisa; Shioda, Hironobu; Hirai, Fumiko; Hiroshima, Kenzo; Kishimoto, Takumi

    2011-01-01

    This study was designed to examine the significance of pleural plaques and pulmonary fibrosis in the evaluation of asbestos exposure level. The subjects were patients who had undergone surgery for lung cancer. There were 64 patients who had pleural plaques based on surgical findings (pleural plaque group) and 9 patients who had neither a history of asbestos exposure nor pleural plaque (control group). An examination was performed regarding the extent of pleural plaques and the presence or absence of pulmonary fibrosis. The relationships between these findings and the asbestos body count in the resected lung were investigated. If chest CT showed no pleural plaque, the case was classified as class 0. If chest CT showed pleural plaques, the CT slice with the most extensive pleural plaque in either side was selected. If the plaque extended to less than one quarter of the inner chest wall, the case was classified as class 1. If the extent was one quarter or more, the case was classified as class 2. The cases were considered to have pulmonary fibrosis if the fibrotic findings were equivalent to those of asbestosis of type 1 or more by chest X-ray photography (XP) and if fibrosis was observed in CT. All other cases were considered not to have pulmonary fibrosis. The median asbestos body counts were 1,018 bodies per gram of dried lung in the pleural plaque group and 263 per gram of dried lung in the control group. There was a statistically significant difference between these groups (p=0.0034). There were 25 patients with class 0 pleural plaque, 17 patients with class 1, and 22 patients with class 2. Their median asbestos body counts were 612, 439, and 5,626 bodies, respectively. All class 0 or 1 patients had an asbestos body count of less than 5,000 bodies. There was no significant difference in the counts between patients with class 0 and 1. All class 2 patients had an asbestos body count of 1,000 bodies or more. The count of class 2 patients was significantly higher

  17. Overdispersion in nuclear statistics

    International Nuclear Information System (INIS)

    Semkow, Thomas M.

    1999-01-01

    The modern statistical distribution theory is applied to the development of the overdispersion theory in ionizing-radiation statistics for the first time. The physical nuclear system is treated as a sequence of binomial processes, each depending on a characteristic probability, such as probability of decay, detection, etc. The probabilities fluctuate in the course of a measurement, and the physical reasons for that are discussed. If the average values of the probabilities change from measurement to measurement, which originates from the random Lexis binomial sampling scheme, then the resulting distribution is overdispersed. The generating functions and probability distribution functions are derived, followed by a moment analysis. The Poisson and Gaussian limits are also given. The distribution functions belong to a family of generalized hypergeometric factorial moment distributions by Kemp and Kemp, and can serve as likelihood functions for the statistical estimations. An application to radioactive decay with detection is described and working formulae are given, including a procedure for testing the counting data for overdispersion. More complex experiments in nuclear physics (such as solar neutrino) can be handled by this model, as well as distinguishing between the source and background

  18. Eosinophil count is positively correlated with coronary artery calcification

    International Nuclear Information System (INIS)

    Tanaka, Muhei; Fukui, Michiaki; Yamasaki, Masahiro; Hasegawa, Goji; Oda, Yohei; Nakamura, Naoto; Tomiyasu, Ki-ichiro; Akabame, Satoshi; Nakano, Koji

    2012-01-01

    Recent studies suggested that allergic disorders and increased eosinophil count were associated with atherosclerosis. The purpose of this study was to assess the relationship between eosinophil count and coronary artery calcification (CAC). We performed a cross-sectional study in 1363 consecutive participants with clinical suspicion of coronary heart disease (CHD). We evaluated the relationships between CAC score determined by multislice CT and peripheral eosinophil count as well as major cardiovascular risk factors, including age, body mass index, smoking status, hypertension, dyslipidemia, diabetes mellitus (DM), high-sensitivity C-reactive protein and estimated glomerular filtration rate (eGFR). Sex (P=0.0004), hypertension (P=0.0002), dyslipidemia (P=0.0004) and DM (P=0.0061) were associated with log (CAC+1), respectively. Positive correlations were found between log (CAC+1), and age (r=0.325, P<0.0001) and eosinophil count (r=0.165, P<0.0001). Negative correlations were found between log (CAC+1) and eGFR (r=-0.166, P<0.0001). Multivariate linear regression analysis demonstrated that age (β=0.314, P<0.0001), sex (β=0.124, P<0.0001), hypertension (β=0.084, P=0.0008), DM (β=0.108, P<0.0001), eGFR (β=-0.079, P=0.0021) and eosinophil count (β=0.147, P<0.0001) were independent determinants of log (CAC+1). In conclusion, eosinophil count correlated positively with CAC in participants with clinical suspicion of CHD. (author)

  19. Periodontal status of HIV infected patients with special reference to CD4 cell count in West Bengal, India

    Directory of Open Access Journals (Sweden)

    Shallu Rozra

    2012-12-01

    Full Text Available Objective: To evaluate the periodontal status of HIV seropositive patients and to find out if any correlation exists between the severity of periodontal disease and the CD4 cell count in HIV patients. Methods: One hundred and thirty patients attending the Viral Diseases OPD, Calcutta School of Tropical Medicine, Kolkata were examined. They were grouped according to the CD4 cell count as Group A - Subjects with CD4 Cell count < 200/ 毺 L and Group B - Subjects with CD4 Cell count 曒 200/ 毺 L. Their community periodontal index of treatment needs (CPITN score were recorded. Results: It was found that most of the patients in each group were having score ‘2’ (i.e. presence of supra or subgingival calculus, as their highest score. A statistically significant association was found between immune status as depicted by CD4 cell count and periodontal status as shown by highest CPITN score in the present study. Conclusions: The present study confirms the effect of immunosuppression on periodontal diseases in HIV infected patients.

  20. High-Sensitivity Semiconductor Photocathodes for Space-Born UV Photon-Counting and Imaging, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — Many UV photon-counting and imaging applications, including space-borne astronomy, missile tracking and guidance, UV spectroscopy for chemical/biological...

  1. Statistical Inference for a Class of Multivariate Negative Binomial Distributions

    DEFF Research Database (Denmark)

    Rubak, Ege H.; Møller, Jesper; McCullagh, Peter

    This paper considers statistical inference procedures for a class of models for positively correlated count variables called -permanental random fields, and which can be viewed as a family of multivariate negative binomial distributions. Their appealing probabilistic properties have earlier been...... studied in the literature, while this is the first statistical paper on -permanental random fields. The focus is on maximum likelihood estimation, maximum quasi-likelihood estimation and on maximum composite likelihood estimation based on uni- and bivariate distributions. Furthermore, new results...

  2. Statistics for Radiation Measurement. Chapter 5

    Energy Technology Data Exchange (ETDEWEB)

    Lötter, M. G. [Department of Medical Physics, University of the Free State, Bloemfontein (South Africa)

    2014-12-15

    Measurement errors are of three general types: (i) blunders, (ii) systematic errors or accuracy of measurements, and (iii) random errors or precision of measurements. Blunders produce grossly inaccurate results and experienced observers easily detect their occurrence. Examples in radiation counting or measurements include the incorrect setting of the energy window, counting heavily contaminated samples, using contaminated detectors for imaging or counting, obtaining measurements of high activities, resulting in count rates that lead to excessive dead time effects, and selecting the wrong patient orientation during imaging. Although some blunders can be detected as outliers or by duplicate samples and measurements, blunders should be avoided by careful, meticulous and dedicated work. This is especially important where results will determine the diagnosis or treatment of patients.

  3. SPERM COUNT DISTRIBUTIONS IN FERTILE MEN

    Science.gov (United States)

    Sperm concentration and count are often used as indicators of environmental impacts on male reproductive health. Existing clinical databases may be biased towards subfertile men with low sperm counts and less is known about expected sperm count distributions in cohorts of fertil...

  4. Oral manifestations of human immunodeficiency virus/acquired immunodeficiency syndrome and their correlation to cluster of differentiation lymphocyte count in population of North-East India in highly active antiretroviral therapy era

    Directory of Open Access Journals (Sweden)

    Sarat Kumar Nayak

    2016-01-01

    Full Text Available Background: The human immunodeficiency virus (HIV infection which manifests as acquired immunodeficiency syndrome (AIDS is a disease involving the defects of the T-lymphocyte arm of the immune system. Certain laboratory parameters such as the cluster of differentiation (CD4 count and clinical parameters have long been used as markers of disease progression. In industrialized countries, many studies show a highly correlation between the incidence of oral lesions and immunosuppression and hence, can be used as a marker of immunosuppression. This might not be applicable to a developing country like India. In this study, efforts have been made to supplement the present knowledge on various aspects of oral manifestations in HIV patients in the Indian subcontinent. Aims: To correlate the oral manifestations in HIV/AIDS patients to the level of circulating CD4+ T-lymphocyte count and their effect in anti-retroviral therapy (ART. Subjects and Methods: A total of 104 HIV positive patients were examined for oral lesions. The CD4 count estimated on the same day by fluorescent activated cell sort count machine was then correlated with various oral lesions. Results: Oral manifestations appeared when CD4 count decreased below 500 cells/mm3. Moreover, oral lesions found at different stages showed very strong correlation to their respective CD4 count. Furthermore, there was considerable decline in the incidence of oral manifestations in patients undergoing highly active ART. Conclusions: Oral manifestations are highly predictive markers of severe immune deterioration and disease progression in HIV patients.

  5. Quantitative monitoring of the fluorination process by neutron counting

    International Nuclear Information System (INIS)

    Russo, P.A.; Appert, Q.D.; Biddle, R.S.; Kelley, T.A.; Martinez, M.M.; West, M.H.

    1993-01-01

    Plutonium metal is produced by reducing PuF 4 prepared from PuO 2 by fluorination. Both fluorination and reduction are batch processes at the Los Alamos Plutonium Facility. The conversion of plutonium oxide to fluoride greatly increases the neutron yield, a result of the high cross section for alpha-neutron (α,n) reactions on fluorine targets compared to the (more than 100 times) smaller α,n yield on oxygen targets. Because of the increase, total neutron counting can be used to monitor the conversion process. This monitoring ability can lead to an improved metal product, reduced scrap for recycle, waste reduction, minimized reagent usage, and reduce personnel radiation exposures. A new stirred-bed fluorination process has been developed simultaneously with a recent evaluation of an automated neutron-counting instrument for quantitative process monitoring. Neutrons are counted with polyethylene-moderated 3 He-gas proportional counters. Results include a calibration of the real-time neutron-count-rate indicator for the extent of fluorination using reference values obtained from destructive analysis of samples from the blended fluoroinated batch

  6. Statistical study of high-latitude plasma flow during magnetospheric substorms

    Directory of Open Access Journals (Sweden)

    G. Provan

    2004-11-01

    Full Text Available We have utilised the near-global imaging capabilities of the Northern Hemisphere SuperDARN radars, to perform a statistical superposed epoch analysis of high-latitude plasma flows during magnetospheric substorms. The study involved 67 substorms, identified using the IMAGE FUV space-borne auroral imager. A substorm co-ordinate system was developed, centred on the magnetic local time and magnetic latitude of substorm onset determined from the auroral images. The plasma flow vectors from all 67 intervals were combined, creating global statistical plasma flow patterns and backscatter occurrence statistics during the substorm growth and expansion phases. The commencement of the substorm growth phase was clearly observed in the radar data 18-20min before substorm onset, with an increase in the anti-sunward component of the plasma velocity flowing across dawn sector of the polar cap and a peak in the dawn-to-dusk transpolar voltage. Nightside backscatter moved to lower latitudes as the growth phase progressed. At substorm onset a flow suppression region was observed on the nightside, with fast flows surrounding the suppressed flow region. The dawn-to-dusk transpolar voltage increased from ~40kV just before substorm onset to ~75kV 12min after onset. The low-latitude return flow started to increase at substorm onset and continued to increase until 8min after onset. The velocity flowing across the polar-cap peaked 12-14min after onset. This increase in the flux of the polar cap and the excitation of large-scale plasma flow occurred even though the IMF Bz component was increasing (becoming less negative during most of this time. This study is the first to statistically prove that nightside reconnection creates magnetic flux and excites high-latitude plasma flow in a similar way to dayside reconnection and that dayside and nightside reconnection, are two separate time-dependent processes.

  7. Bose and his statistics

    International Nuclear Information System (INIS)

    Venkataraman, G.

    1992-01-01

    Treating radiation gas as a classical gas, Einstein derived Planck's law of radiation by considering the dynamic equilibrium between atoms and radiation. Dissatisfied with this treatment, S.N. Bose derived Plank's law by another original way. He treated the problem in generality: he counted how many cells were available for the photon gas in phase space and distributed the photons into these cells. In this manner of distribution, there were three radically new ideas: The indistinguishability of particles, the spin of the photon (with only two possible orientations) and the nonconservation of photon number. This gave rise to a new discipline of quantum statistical mechanics. Physics underlying Bose's discovery, its significance and its role in development of the concept of ideal gas, spin-statistics theorem and spin particles are described. The book has been written in a simple and direct language in an informal style aiming to stimulate the curiosity of a reader. (M.G.B.)

  8. Bayesian analysis of energy and count rate data for detection of low count rate radioactive sources.

    Science.gov (United States)

    Klumpp, John; Brandl, Alexander

    2015-03-01

    A particle counting and detection system is proposed that searches for elevated count rates in multiple energy regions simultaneously. The system analyzes time-interval data (e.g., time between counts), as this was shown to be a more sensitive technique for detecting low count rate sources compared to analyzing counts per unit interval (Luo et al. 2013). Two distinct versions of the detection system are developed. The first is intended for situations in which the sample is fixed and can be measured for an unlimited amount of time. The second version is intended to detect sources that are physically moving relative to the detector, such as a truck moving past a fixed roadside detector or a waste storage facility under an airplane. In both cases, the detection system is expected to be active indefinitely; i.e., it is an online detection system. Both versions of the multi-energy detection systems are compared to their respective gross count rate detection systems in terms of Type I and Type II error rates and sensitivity.

  9. High energy resolution and high count rate gamma spectrometry measurement of primary coolant of generation 4 sodium-cooled fast reactor

    International Nuclear Information System (INIS)

    Coulon, R.

    2010-01-01

    Sodium-cooled Fast Reactors are under development for the fourth generation of nuclear reactor. Breeders reactors could gives solutions for the need of energy and the preservation of uranium resources. An other purpose is the radioactive wastes production reduction by transmutation and the control of non-proliferation using a closed-cycle. These thesis shows safety and profit advantages that could be obtained by a new generation of gamma spectrometry system for SFR. Now, the high count rate abilities, allow us to study new methods of accurate power measurement and fast clad failure detection. Simulations have been done and an experimental test has been performed at the French Phenix SFR of the CEA Marcoule showing promising results for these new measurements. (author) [fr

  10. Statistics of high-level scene context.

    Science.gov (United States)

    Greene, Michelle R

    2013-01-01

    CONTEXT IS CRITICAL FOR RECOGNIZING ENVIRONMENTS AND FOR SEARCHING FOR OBJECTS WITHIN THEM: contextual associations have been shown to modulate reaction time and object recognition accuracy, as well as influence the distribution of eye movements and patterns of brain activations. However, we have not yet systematically quantified the relationships between objects and their scene environments. Here I seek to fill this gap by providing descriptive statistics of object-scene relationships. A total of 48, 167 objects were hand-labeled in 3499 scenes using the LabelMe tool (Russell et al., 2008). From these data, I computed a variety of descriptive statistics at three different levels of analysis: the ensemble statistics that describe the density and spatial distribution of unnamed "things" in the scene; the bag of words level where scenes are described by the list of objects contained within them; and the structural level where the spatial distribution and relationships between the objects are measured. The utility of each level of description for scene categorization was assessed through the use of linear classifiers, and the plausibility of each level for modeling human scene categorization is discussed. Of the three levels, ensemble statistics were found to be the most informative (per feature), and also best explained human patterns of categorization errors. Although a bag of words classifier had similar performance to human observers, it had a markedly different pattern of errors. However, certain objects are more useful than others, and ceiling classification performance could be achieved using only the 64 most informative objects. As object location tends not to vary as a function of category, structural information provided little additional information. Additionally, these data provide valuable information on natural scene redundancy that can be exploited for machine vision, and can help the visual cognition community to design experiments guided by statistics

  11. A compact 7-cell Si-drift detector module for high-count rate X-ray spectroscopy.

    Science.gov (United States)

    Hansen, K; Reckleben, C; Diehl, I; Klär, H

    2008-05-01

    A new Si-drift detector module for fast X-ray spectroscopy experiments was developed and realized. The Peltier-cooled module comprises a sensor with 7 × 7-mm 2 active area, an integrated circuit for amplification, shaping and detection, storage, and derandomized readout of signal pulses in parallel, and amplifiers for line driving. The compactness and hexagonal shape of the module with a wrench size of 16mm allow very short distances to the specimen and multi-module arrangements. The power dissipation is 186mW. At a shaper peaking time of 190 ns and an integration time of 450 ns an electronic rms noise of ~11 electrons was achieved. When operated at 7 °C, FWHM line widths around 260 and 460 eV (Cu-K α ) were obtained at low rates and at sum-count rates of 1.7 MHz, respectively. The peak shift is below 1% for a broad range of count rates. At 1.7-MHz sum-count rate the throughput loss amounts to 30%.

  12. Statistical analysis and digital processing of the Mössbauer spectra

    International Nuclear Information System (INIS)

    Prochazka, Roman; Tucek, Jiri; Mashlan, Miroslav; Pechousek, Jiri; Tucek, Pavel; Marek, Jaroslav

    2010-01-01

    This work is focused on using the statistical methods and development of the filtration procedures for signal processing in Mössbauer spectroscopy. Statistical tools for noise filtering in the measured spectra are used in many scientific areas. The use of a pure statistical approach in accumulated Mössbauer spectra filtration is described. In Mössbauer spectroscopy, the noise can be considered as a Poisson statistical process with a Gaussian distribution for high numbers of observations. This noise is a superposition of the non-resonant photons counting with electronic noise (from γ-ray detection and discrimination units), and the velocity system quality that can be characterized by the velocity nonlinearities. The possibility of a noise-reducing process using a new design of statistical filter procedure is described. This mathematical procedure improves the signal-to-noise ratio and thus makes it easier to determine the hyperfine parameters of the given Mössbauer spectra. The filter procedure is based on a periodogram method that makes it possible to assign the statistically important components in the spectral domain. The significance level for these components is then feedback-controlled using the correlation coefficient test results. The estimation of the theoretical correlation coefficient level which corresponds to the spectrum resolution is performed. Correlation coefficient test is based on comparison of the theoretical and the experimental correlation coefficients given by the Spearman method. The correctness of this solution was analyzed by a series of statistical tests and confirmed by many spectra measured with increasing statistical quality for a given sample (absorber). The effect of this filter procedure depends on the signal-to-noise ratio and the applicability of this method has binding conditions

  13. Statistical analysis and digital processing of the Mössbauer spectra

    Science.gov (United States)

    Prochazka, Roman; Tucek, Pavel; Tucek, Jiri; Marek, Jaroslav; Mashlan, Miroslav; Pechousek, Jiri

    2010-02-01

    This work is focused on using the statistical methods and development of the filtration procedures for signal processing in Mössbauer spectroscopy. Statistical tools for noise filtering in the measured spectra are used in many scientific areas. The use of a pure statistical approach in accumulated Mössbauer spectra filtration is described. In Mössbauer spectroscopy, the noise can be considered as a Poisson statistical process with a Gaussian distribution for high numbers of observations. This noise is a superposition of the non-resonant photons counting with electronic noise (from γ-ray detection and discrimination units), and the velocity system quality that can be characterized by the velocity nonlinearities. The possibility of a noise-reducing process using a new design of statistical filter procedure is described. This mathematical procedure improves the signal-to-noise ratio and thus makes it easier to determine the hyperfine parameters of the given Mössbauer spectra. The filter procedure is based on a periodogram method that makes it possible to assign the statistically important components in the spectral domain. The significance level for these components is then feedback-controlled using the correlation coefficient test results. The estimation of the theoretical correlation coefficient level which corresponds to the spectrum resolution is performed. Correlation coefficient test is based on comparison of the theoretical and the experimental correlation coefficients given by the Spearman method. The correctness of this solution was analyzed by a series of statistical tests and confirmed by many spectra measured with increasing statistical quality for a given sample (absorber). The effect of this filter procedure depends on the signal-to-noise ratio and the applicability of this method has binding conditions.

  14. Strengthening Children's Math Skills with Enhanced Instruction: The Impacts of Making Pre-K Count and High 5s on Kindergarten Outcomes

    Science.gov (United States)

    Mattera, Shira K.; Jacob, Robin; Morris, Pamela A.

    2018-01-01

    Early math skills are a strong predictor of later achievement for young children, not only in math, but in other domains as well. Exhibiting strong math skills in elementary school is predictive of later high school completion and college attendance. To that end, the Making Pre-K Count and High 5s studies set out to rigorously assess whether…

  15. Rainflow counting revisited

    Energy Technology Data Exchange (ETDEWEB)

    Soeker, H [Deutsches Windenergie-Institut (Germany)

    1996-09-01

    As state of the art method the rainflow counting technique is presently applied everywhere in fatigue analysis. However, the author feels that the potential of the technique is not fully recognized in wind energy industries as it is used, most of the times, as a mere data reduction technique disregarding some of the inherent information of the rainflow counting results. The ideas described in the following aim at exploitation of this information and making it available for use in the design and verification process. (au)

  16. Platelet Count and Plateletcrit

    African Journals Online (AJOL)

    strated that neonates with late onset sepsis (bacteremia after 3 days of age) had a dramatic increase in MPV and. PDW18. We hypothesize that as the MPV and PDW increase and platelet count and PCT decrease in sick children, intui- tively, the ratio of MPV to PCT; MPV to Platelet count,. PDW to PCT, PDW to platelet ...

  17. A hierarchical spatiotemporal analog forecasting model for count data.

    Science.gov (United States)

    McDermott, Patrick L; Wikle, Christopher K; Millspaugh, Joshua

    2018-01-01

    Analog forecasting is a mechanism-free nonlinear method that forecasts a system forward in time by examining how past states deemed similar to the current state moved forward. Previous applications of analog forecasting has been successful at producing robust forecasts for a variety of ecological and physical processes, but it has typically been presented in an empirical or heuristic procedure, rather than as a formal statistical model. The methodology presented here extends the model-based analog method of McDermott and Wikle (Environmetrics, 27, 2016, 70) by placing analog forecasting within a fully hierarchical statistical framework that can accommodate count observations. Using a Bayesian approach, the hierarchical analog model is able to quantify rigorously the uncertainty associated with forecasts. Forecasting waterfowl settling patterns in the northwestern United States and Canada is conducted by applying the hierarchical analog model to a breeding population survey dataset. Sea surface temperature (SST) in the Pacific Ocean is used to help identify potential analogs for the waterfowl settling patterns.

  18. The Bayesian count rate probability distribution in measurement of ionizing radiation by use of a ratemeter

    Energy Technology Data Exchange (ETDEWEB)

    Weise, K.

    2004-06-01

    Recent metrological developments concerning measurement uncertainty, founded on Bayesian statistics, give rise to a revision of several parts of the DIN 25482 and ISO 11929 standard series. These series stipulate detection limits and decision thresholds for ionizing-radiation measurements. Part 3 and, respectively, part 4 of them deal with measurements by use of linear-scale analogue ratemeters. A normal frequency distribution of the momentary ratemeter indication for a fixed count rate value is assumed. The actual distribution, which is first calculated numerically by solving an integral equation, differs, however, considerably from the normal distribution although this one represents an approximation of it for sufficiently large values of the count rate to be measured. As is shown, this similarly holds true for the Bayesian probability distribution of the count rate for sufficiently large given measured values indicated by the ratemeter. This distribution follows from the first one mentioned by means of the Bayes theorem. Its expectation value and variance are needed for the standards to be revised on the basis of Bayesian statistics. Simple expressions are given by the present standards for estimating these parameters and for calculating the detection limit and the decision threshold. As is also shown, the same expressions can similarly be used as sufficient approximations by the revised standards if, roughly, the present indicated value exceeds the reciprocal ratemeter relaxation time constant. (orig.)

  19. Estimating the standard deviation for 222Rn scintillation counting - a note concerning the paper by Sarmiento et al

    International Nuclear Information System (INIS)

    Key, R.M.

    1977-01-01

    In a recent report Sarmiento et al.(1976) presented a method for estimating the statistical error associated with 222 Rn scintillation counting. Because of certain approximations, the method is less accurate than that of an earlier work by Lucas and Woodward (1964). The Sarmiento method and the Lucas method are compared, and the magnitude of errors incurred using the approximations are determined. For counting times greater than 300 minutes, the disadvantage of the slight inaccuracies of the Sarmiento method are outweighed by the advantage of easier calculation. (Auth.)

  20. Finite-size effects in transcript sequencing count distribution: its power-law correction necessarily precedes downstream normalization and comparative analysis.

    Science.gov (United States)

    Wong, Wing-Cheong; Ng, Hong-Kiat; Tantoso, Erwin; Soong, Richie; Eisenhaber, Frank

    2018-02-12

    signal-to-noise ratio by 50% and the statistical/detection sensitivity by as high as 30% regardless of the downstream mapping and normalization methods. Most importantly, the power-law correction improves concordance in significant calls among different normalization methods of a data series averagely by 22%. When presented with a higher sequence depth (4 times difference), the improvement in concordance is asymmetrical (32% for the higher sequencing depth instance versus 13% for the lower instance) and demonstrates that the simple power-law correction can increase significant detection with higher sequencing depths. Finally, the correction dramatically enhances the statistical conclusions and eludes the metastasis potential of the NUGC3 cell line against AGS of our dilution analysis. The finite-size effects due to undersampling generally plagues transcript count data with reproducibility issues but can be minimized through a simple power-law correction of the count distribution. This distribution correction has direct implication on the biological interpretation of the study and the rigor of the scientific findings. This article was reviewed by Oliviero Carugo, Thomas Dandekar and Sandor Pongor.