Effect of sample stratification on dairy GWAS results
Background Artificial insemination and genetic selection are major factors contributing to population stratification in dairy cattle. In this study, we analyzed the effect of sample stratification and the effect of stratification correction on results of a dairy genome-wide association study (GWAS)....
Desu, M M
2012-01-01
One of the most important problems in designing an experiment or a survey is sample size determination and this book presents the currently available methodology. It includes both random sampling from standard probability distributions and from finite populations. Also discussed is sample size determination for estimating parameters in a Bayesian setting by considering the posterior distribution of the parameter and specifying the necessary requirements. The determination of the sample size is considered for ranking and selection problems as well as for the design of clinical trials. Appropria
Rakesh R. Pathak
2012-02-01
Full Text Available Based on the law of large numbers which is derived from probability theory, we tend to increase the sample size to the maximum. Central limit theorem is another inference from the same probability theory which approves largest possible number as sample size for better validity of measuring central tendencies like mean and median. Sometimes increase in sample-size turns only into negligible betterment or there is no increase at all in statistical relevance due to strong dependence or systematic error. If we can afford a little larger sample, statistically power of 0.90 being taken as acceptable with medium Cohen's d (<0.5 and for that we can take a sample size of 175 very safely and considering problem of attrition 200 samples would suffice. [Int J Basic Clin Pharmacol 2012; 1(1.000: 43-44
Sampling efficiency of national, EU and global stratifications : exploring by using CL2000
Metzger, M.J.; Brus, D.J.; Ortega, M.
2012-01-01
Stratification, dividing the statistical population into less heterogeneous subgroups before sampling, can help improve sampling efficiency by improving representativeness and reducing sampling error. This report explores the added sampling efficiency that is achieved by using the European Environme
PAN Genxing; WU Laosheng; LI Lianqing; ZHANG Xuhui; GONG Wei; WOOD Yvonne
2008-01-01
Developing realistic soil carbon (C) sequestration strategies for China's sustainable agriculture relies on accurate estimates of the amount, retention and turnover rates of C stored in paddy soils. Available C estimates to date are predominantly for the tilled and flood-irrigated surface topsoil (ca. 30 cm). Such estimates cannot be used to extrapolate to soil depths of 100 cm since soil organic carbon (SOC) generally shows a sharp decrease with depth. In this research, composite soil samples were collected at several depths to 100 cm from three representative paddy soils in the Taihu Lake region, China. Soil organic carbon distribution in the profiles and in aggregate-size fractions was determined. Results showed that while SOC decreased exponentially with depth to 100 cm, a substantial proportion of the total SOC (30%-40%) is stored below the 30 cm depth. In the carbon-enriched paddy topsoils, SOC was found to accumulate preferentially in the 2-0.25 and 0.25-0.02 mm aggregate size fractions. d13C analysis of the coarse micro-aggregate fraction showed that the high degree of C stratification in the paddy topsoil was in agreement with the occurrence of lighter d1313C in the upper 30 cm depth. These results suggest that SOC stratification within profiles varies with different pedogenetical types of paddy soils with regards to clay and iron oxyhydrates distributions. Sand-sized fractions of aggregates in paddy soil systems may play a very important role in carbon sequestration and turnover, dissimilar to other studied agricultural systems.
Roberge, Cornelia; Wulff, Sören; Reese, Heather; Ståhl, Göran
2016-04-01
Many countries have a national forest inventory (NFI) designed to produce statistically sound estimates of forest parameters. However, this type of inventory may not provide reliable results for forest damage which usually affects only small parts of the forest in a country. For this reason, specially designed forest damage inventories are performed in many countries, sometimes in coordination with the NFIs. In this study, we evaluated a new approach for damage inventory where existing NFI data form the basis for two-phase sampling for stratification and remotely sensed auxiliary data are applied for further improvement of precision through post-stratification. We applied Monte Carlo sampling simulation to evaluate different sampling strategies linked to different damage scenarios. The use of existing NFI data in a two-phase sampling for stratification design resulted in a relative efficiency of 50 % or lower, i.e., the variance was at least halved compared to a simple random sample of the same size. With post-stratification based on simulated remotely sensed auxiliary data, there was additional improvement, which depended on the accuracy of the auxiliary data and the properties of the forest damage. In many cases, the relative efficiency was further reduced by as much as one-half. In conclusion, the results show that substantial gains in precision can be obtained by utilizing auxiliary information in forest damage surveys, through two-phase sampling, through post-stratification, and through the combination of these two approaches, i.e., post-stratified two-phase sampling for stratification.
Sample size determination and power
Ryan, Thomas P, Jr
2013-01-01
THOMAS P. RYAN, PhD, teaches online advanced statistics courses for Northwestern University and The Institute for Statistics Education in sample size determination, design of experiments, engineering statistics, and regression analysis.
Size definitions for particle sampling
1981-05-01
The recommendations of an ad hoc working group appointed by Committee TC 146 of the International Standards Organization on size definitions for particle sampling are reported. The task of the group was to collect the various definitions of 'respirable dust' and to propose a practical definition on recommendations for handling standardization on this matter. One of two proposed cut-sizes in regard to division at the larynx will be adopted after a ballot.
Michael S. Williams
2001-01-01
A number of different estimators can be used when forest inventory plots cover two or more distinctly different condition classes. In this article the properties of two approximate Horvitz- Thompson (HT) estimators, a ratio of means (RM), and a mean of ratios (MR) estimator are explored in the framework of double sampling for stratification. Relevant theoretical...
[Clinical research V. Sample size].
Talavera, Juan O; Rivas-Ruiz, Rodolfo; Bernal-Rosales, Laura Paola
2011-01-01
In clinical research it is impossible and inefficient to study all patients with a specific pathology, so it is necessary to study a sample of them. The estimation of the sample size before starting a study guarantees the stability of the results and allows us to foresee the feasibility of the study depending on the availability of patients and cost. The basic structure of sample size estimation is based on the premise that seeks to demonstrate, among other cases, that the observed difference between two or more maneuvers in the subsequent state is real. Initially, it requires knowing the value of the expected difference (δ) and its data variation (standard deviation). These data are usually obtained from previous studies. Then, other components must be considered: a (alpha), percentage of error in the assertion that the difference between means is real, usually 5 %; and β, error rate accepting the claim that the no-difference between the means is real, usually ranging from 15 to 20 %. Finally, these values are substituted into the formula or in an electronic program for estimating sample size. While summary and dispersion measures vary with the type of variable according to the outcome, the basic structure is the same.
Principal Stratification in sample selection problems with non normal error terms
Rocci, Roberto; Mellace, Giovanni
The aim of the paper is to relax distributional assumptions on the error terms, often imposed in parametric sample selection models to estimate causal effects, when plausible exclusion restrictions are not available. Within the principal stratification framework, we approximate the true distribut......The aim of the paper is to relax distributional assumptions on the error terms, often imposed in parametric sample selection models to estimate causal effects, when plausible exclusion restrictions are not available. Within the principal stratification framework, we approximate the true...... distribution of the error terms with a mixture of Gaussian. We propose an EM type algorithm for ML estimation. In a simulation study we show that our estimator has lower MSE than the ML and two-step Heckman estimators with any non normal distribution considered for the error terms. Finally we provide...... an application to the Job Corps training program....
PIXE-PIGE analysis of size-segregated aerosol samples from remote areas
Calzolai, G.; Chiari, M.; Lucarelli, F.; Nava, S.; Taccetti, F.; Becagli, S.; Frosini, D.; Traversi, R.; Udisti, R.
2014-01-01
The chemical characterization of size-segregated samples is helpful to study the aerosol effects on both human health and environment. The sampling with multi-stage cascade impactors (e.g., Small Deposit area Impactor, SDI) produces inhomogeneous samples, with a multi-spot geometry and a non-negligible particle stratification.
Sample size in qualitative interview studies
Malterud, Kirsti; Siersma, Volkert Dirk; Guassora, Ann Dorrit Kristiane
2016-01-01
Sample sizes must be ascertained in qualitative studies like in quantitative studies but not by the same means. The prevailing concept for sample size in qualitative studies is “saturation.” Saturation is closely tied to a specific methodology, and the term is inconsistently applied. We propose...... the concept “information power” to guide adequate sample size for qualitative studies. Information power indicates that the more information the sample holds, relevant for the actual study, the lower amount of participants is needed. We suggest that the size of a sample with sufficient information power...... and during data collection of a qualitative study is discussed....
How sample size influences research outcomes
Jorge Faber
2014-08-01
Full Text Available Sample size calculation is part of the early stages of conducting an epidemiological, clinical or lab study. In preparing a scientific paper, there are ethical and methodological indications for its use. Two investigations conducted with the same methodology and achieving equivalent results, but different only in terms of sample size, may point the researcher in different directions when it comes to making clinical decisions. Therefore, ideally, samples should not be small and, contrary to what one might think, should not be excessive. The aim of this paper is to discuss in clinical language the main implications of the sample size when interpreting a study.
Size stratification in a Gilbert delta due to a varying base level: flume experiments.
Chavarrias, Victor; Orru, Clara; Viparelli, Enrica; Vide, Juan Pedro Martin; Blom, Astrid
2014-05-01
mobile armor that covered the fluvial reach. This led to an initial coarsening of the brinkpoint load (and foreset deposit). Once the mobile armour was eroded, base level fall led to degradation of the finer substrate, which resulted in a fining of the brinkpoint load and foreset deposit. The relation between the sediment size stratification and the base level change may be used for the reconstruction of the paleo sea level from the stratigraphy of ancient Gilbert deltas.
Hanike, Yusrianti; Sadik, Kusman; Kurnia, Anang
2016-02-01
This research implemented unemployment rate in Indonesia that based on Poisson distribution. It would be estimated by modified the post-stratification and Small Area Estimation (SAE) model. Post-stratification was one of technique sampling that stratified after collected survey data. It's used when the survey data didn't serve for estimating the interest area. Interest area here was the education of unemployment which separated in seven category. The data was obtained by Labour Employment National survey (Sakernas) that's collected by company survey in Indonesia, BPS, Statistic Indonesia. This company served the national survey that gave too small sample for level district. Model of SAE was one of alternative to solved it. According the problem above, we combined this post-stratification sampling and SAE model. This research gave two main model of post-stratification sampling. Model I defined the category of education was the dummy variable and model II defined the category of education was the area random effect. Two model has problem wasn't complied by Poisson assumption. Using Poisson-Gamma model, model I has over dispersion problem was 1.23 solved to 0.91 chi square/df and model II has under dispersion problem was 0.35 solved to 0.94 chi square/df. Empirical Bayes was applied to estimate the proportion of every category education of unemployment. Using Bayesian Information Criteria (BIC), Model I has smaller mean square error (MSE) than model II.
Biostatistics Series Module 5: Determining Sample Size.
Hazra, Avijit; Gogtay, Nithya
2016-01-01
Determining the appropriate sample size for a study, whatever be its type, is a fundamental aspect of biomedical research. An adequate sample ensures that the study will yield reliable information, regardless of whether the data ultimately suggests a clinically important difference between the interventions or elements being studied. The probability of Type 1 and Type 2 errors, the expected variance in the sample and the effect size are the essential determinants of sample size in interventional studies. Any method for deriving a conclusion from experimental data carries with it some risk of drawing a false conclusion. Two types of false conclusion may occur, called Type 1 and Type 2 errors, whose probabilities are denoted by the symbols σ and β. A Type 1 error occurs when one concludes that a difference exists between the groups being compared when, in reality, it does not. This is akin to a false positive result. A Type 2 error occurs when one concludes that difference does not exist when, in reality, a difference does exist, and it is equal to or larger than the effect size defined by the alternative to the null hypothesis. This may be viewed as a false negative result. When considering the risk of Type 2 error, it is more intuitive to think in terms of power of the study or (1 - β). Power denotes the probability of detecting a difference when a difference does exist between the groups being compared. Smaller α or larger power will increase sample size. Conventional acceptable values for power and α are 80% or above and 5% or below, respectively, when calculating sample size. Increasing variance in the sample tends to increase the sample size required to achieve a given power level. The effect size is the smallest clinically important difference that is sought to be detected and, rather than statistical convention, is a matter of past experience and clinical judgment. Larger samples are required if smaller differences are to be detected. Although the
Basic Statistical Concepts for Sample Size Estimation
Vithal K Dhulkhed
2008-01-01
Full Text Available For grant proposals the investigator has to include an estimation of sample size .The size of the sample should be adequate enough so that there is sufficient data to reliably answer the research question being addressed by the study. At the very planning stage of the study the investigator has to involve the statistician. To have meaningful dialogue with the statistician every research worker should be familiar with the basic concepts of statistics. This paper is concerned with simple principles of sample size calculation. Concepts are explained based on logic rather than rigorous mathematical calculations to help him assimilate the fundamentals.
Sample size planning for classification models.
Beleites, Claudia; Neugebauer, Ute; Bocklitz, Thomas; Krafft, Christoph; Popp, Jürgen
2013-01-14
In biospectroscopy, suitably annotated and statistically independent samples (e.g. patients, batches, etc.) for classifier training and testing are scarce and costly. Learning curves show the model performance as function of the training sample size and can help to determine the sample size needed to train good classifiers. However, building a good model is actually not enough: the performance must also be proven. We discuss learning curves for typical small sample size situations with 5-25 independent samples per class. Although the classification models achieve acceptable performance, the learning curve can be completely masked by the random testing uncertainty due to the equally limited test sample size. In consequence, we determine test sample sizes necessary to achieve reasonable precision in the validation and find that 75-100 samples will usually be needed to test a good but not perfect classifier. Such a data set will then allow refined sample size planning on the basis of the achieved performance. We also demonstrate how to calculate necessary sample sizes in order to show the superiority of one classifier over another: this often requires hundreds of statistically independent test samples or is even theoretically impossible. We demonstrate our findings with a data set of ca. 2550 Raman spectra of single cells (five classes: erythrocytes, leukocytes and three tumour cell lines BT-20, MCF-7 and OCI-AML3) as well as by an extensive simulation that allows precise determination of the actual performance of the models in question. Copyright © 2012 Elsevier B.V. All rights reserved.
Experimental determination of size distributions: analyzing proper sample sizes
Buffo, A.; Alopaeus, V.
2016-04-01
The measurement of various particle size distributions is a crucial aspect for many applications in the process industry. Size distribution is often related to the final product quality, as in crystallization or polymerization. In other cases it is related to the correct evaluation of heat and mass transfer, as well as reaction rates, depending on the interfacial area between the different phases or to the assessment of yield stresses of polycrystalline metals/alloys samples. The experimental determination of such distributions often involves laborious sampling procedures and the statistical significance of the outcome is rarely investigated. In this work, we propose a novel rigorous tool, based on inferential statistics, to determine the number of samples needed to obtain reliable measurements of size distribution, according to specific requirements defined a priori. Such methodology can be adopted regardless of the measurement technique used.
Sample size calculation in metabolic phenotyping studies.
Billoir, Elise; Navratil, Vincent; Blaise, Benjamin J
2015-09-01
The number of samples needed to identify significant effects is a key question in biomedical studies, with consequences on experimental designs, costs and potential discoveries. In metabolic phenotyping studies, sample size determination remains a complex step. This is due particularly to the multiple hypothesis-testing framework and the top-down hypothesis-free approach, with no a priori known metabolic target. Until now, there was no standard procedure available to address this purpose. In this review, we discuss sample size estimation procedures for metabolic phenotyping studies. We release an automated implementation of the Data-driven Sample size Determination (DSD) algorithm for MATLAB and GNU Octave. Original research concerning DSD was published elsewhere. DSD allows the determination of an optimized sample size in metabolic phenotyping studies. The procedure uses analytical data only from a small pilot cohort to generate an expanded data set. The statistical recoupling of variables procedure is used to identify metabolic variables, and their intensity distributions are estimated by Kernel smoothing or log-normal density fitting. Statistically significant metabolic variations are evaluated using the Benjamini-Yekutieli correction and processed for data sets of various sizes. Optimal sample size determination is achieved in a context of biomarker discovery (at least one statistically significant variation) or metabolic exploration (a maximum of statistically significant variations). DSD toolbox is encoded in MATLAB R2008A (Mathworks, Natick, MA) for Kernel and log-normal estimates, and in GNU Octave for log-normal estimates (Kernel density estimates are not robust enough in GNU octave). It is available at http://www.prabi.fr/redmine/projects/dsd/repository, with a tutorial at http://www.prabi.fr/redmine/projects/dsd/wiki. © The Author 2015. Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.
Predicting sample size required for classification performance
Figueroa Rosa L
2012-02-01
Full Text Available Abstract Background Supervised learning methods need annotated data in order to generate efficient models. Annotated data, however, is a relatively scarce resource and can be expensive to obtain. For both passive and active learning methods, there is a need to estimate the size of the annotated sample required to reach a performance target. Methods We designed and implemented a method that fits an inverse power law model to points of a given learning curve created using a small annotated training set. Fitting is carried out using nonlinear weighted least squares optimization. The fitted model is then used to predict the classifier's performance and confidence interval for larger sample sizes. For evaluation, the nonlinear weighted curve fitting method was applied to a set of learning curves generated using clinical text and waveform classification tasks with active and passive sampling methods, and predictions were validated using standard goodness of fit measures. As control we used an un-weighted fitting method. Results A total of 568 models were fitted and the model predictions were compared with the observed performances. Depending on the data set and sampling method, it took between 80 to 560 annotated samples to achieve mean average and root mean squared error below 0.01. Results also show that our weighted fitting method outperformed the baseline un-weighted method (p Conclusions This paper describes a simple and effective sample size prediction algorithm that conducts weighted fitting of learning curves. The algorithm outperformed an un-weighted algorithm described in previous literature. It can help researchers determine annotation sample size for supervised machine learning.
Selbig, William R; Cox, Amanda; Bannerman, Roger T
2012-04-01
A new water sample collection system was developed to improve representation of solids entrained in urban stormwater by integrating water-quality samples from the entire water column, rather than a single, fixed point. The depth-integrated sample arm (DISA) was better able to characterize suspended-sediment concentration and particle size distribution compared to fixed-point methods when tested in a controlled laboratory environment. Median suspended-sediment concentrations overestimated the actual concentration by 49 and 7% when sampling the water column at 3- and 4-points spaced vertically throughout the water column, respectively. Comparatively, sampling only at the bottom of the pipe, the fixed-point overestimated the actual concentration by 96%. The fixed-point sampler also showed a coarser particle size distribution compared to the DISA which was better able to reproduce the average distribution of particles in the water column over a range of hydraulic conditions. These results emphasize the need for a water sample collection system that integrates the entire water column, rather than a single, fixed point to properly characterize the concentration and distribution of particles entrained in stormwater pipe flow.
Xiao, Hongyi; Deng, Zhekai; Umbanhowar, Paul; Ottino, Julio; Lueptow, Richard
2016-11-01
Segregation of disperse granular materials in unsteady flows is ubiquitous in nature and industry, yet remains largely unexplored. In this study, unsteady flows are generated by feeding size-bidisperse granular mixtures onto a quasi-2D bounded heap using alternating feed rates, which results in stratified layers of large and small particles. The mechanism of stratification is investigated in detail using Discrete Element Method (DEM) simulations of the flow. During the transition from the slow to the fast feed rate, a segregating wedge propagates downstream and forms a large particle layer extending upstream. During the opposite transition, upstream segregated small particles relax downstream and form a small particle layer extending downstream. The transient kinematics from DEM simulations are quantified and used to inform a time-dependent continuum model that captures the interplay of advection, diffusion, and segregation in the flowing layer. The continuum model reproduces the principle characteristics of the stratification patterns observed in experiments and simulations. Funded by NSF Grant CBET-1511450.
Selbig, William R.
2017-01-01
Collection of water-quality samples that accurately characterize average particle concentrations and distributions in channels can be complicated by large sources of variability. The U.S. Geological Survey (USGS) developed a fully automated Depth-Integrated Sample Arm (DISA) as a way to reduce bias and improve accuracy in water-quality concentration data. The DISA was designed to integrate with existing autosampler configurations commonly used for the collection of water-quality samples in vertical profile thereby providing a better representation of average suspended sediment and sediment-associated pollutant concentrations and distributions than traditional fixed-point samplers. In controlled laboratory experiments, known concentrations of suspended sediment ranging from 596 to 1,189 mg/L were injected into a 3 foot diameter closed channel (circular pipe) with regulated flows ranging from 1.4 to 27.8 ft3 /s. Median suspended sediment concentrations in water-quality samples collected using the DISA were within 7 percent of the known, injected value compared to 96 percent for traditional fixed-point samplers. Field evaluation of this technology in open channel fluvial systems showed median differences between paired DISA and fixed-point samples to be within 3 percent. The range of particle size measured in the open channel was generally that of clay and silt. Differences between the concentration and distribution measured between the two sampler configurations could potentially be much larger in open channels that transport larger particles, such as sand.
Defining sample size and sampling strategy for dendrogeomorphic rockfall reconstructions
Morel, Pauline; Trappmann, Daniel; Corona, Christophe; Stoffel, Markus
2015-05-01
Optimized sampling strategies have been recently proposed for dendrogeomorphic reconstructions of mass movements with a large spatial footprint, such as landslides, snow avalanches, and debris flows. Such guidelines have, by contrast, been largely missing for rockfalls and cannot be transposed owing to the sporadic nature of this process and the occurrence of individual rocks and boulders. Based on a data set of 314 European larch (Larix decidua Mill.) trees (i.e., 64 trees/ha), growing on an active rockfall slope, this study bridges this gap and proposes an optimized sampling strategy for the spatial and temporal reconstruction of rockfall activity. Using random extractions of trees, iterative mapping, and a stratified sampling strategy based on an arbitrary selection of trees, we investigate subsets of the full tree-ring data set to define optimal sample size and sampling design for the development of frequency maps of rockfall activity. Spatially, our results demonstrate that the sampling of only 6 representative trees per ha can be sufficient to yield a reasonable mapping of the spatial distribution of rockfall frequencies on a slope, especially if the oldest and most heavily affected individuals are included in the analysis. At the same time, however, sampling such a low number of trees risks causing significant errors especially if nonrepresentative trees are chosen for analysis. An increased number of samples therefore improves the quality of the frequency maps in this case. Temporally, we demonstrate that at least 40 trees/ha are needed to obtain reliable rockfall chronologies. These results will facilitate the design of future studies, decrease the cost-benefit ratio of dendrogeomorphic studies and thus will permit production of reliable reconstructions with reasonable temporal efforts.
Sample size estimation and sampling techniques for selecting a representative sample
Aamir Omair
2014-01-01
Introduction: The purpose of this article is to provide a general understanding of the concepts of sampling as applied to health-related research. Sample Size Estimation: It is important to select a representative sample in quantitative research in order to be able to generalize the results to the target population. The sample should be of the required sample size and must be selected using an appropriate probability sampling technique. There are many hidden biases which can adversely affect ...
A web application for sample size and power calculation in case-control microbiome studies.
Mattiello, Federico; Verbist, Bie; Faust, Karoline; Raes, Jeroen; Shannon, William D; Bijnens, Luc; Thas, Olivier
2016-07-01
: When designing a case-control study to investigate differences in microbial composition, it is fundamental to assess the sample sizes needed to detect an hypothesized difference with sufficient statistical power. Our application includes power calculation for (i) a recoded version of the two-sample generalized Wald test of the 'HMP' R-package for comparing community composition, and (ii) the Wilcoxon-Mann-Whitney test for comparing operational taxonomic unit-specific abundances between two samples (optional). The simulation-based power calculations make use of the Dirichlet-Multinomial model to describe and generate abundances. The web interface allows for easy specification of sample and effect sizes. As an illustration of our application, we compared the statistical power of the two tests, with and without stratification of samples. We observed that statistical power increases considerably when stratification is employed, meaning that less samples are needed to detect the same effect size with the same power. The web interface is written in R code using Shiny (RStudio Inc., 2016) and it is available at https://fedematt.shinyapps.io/shinyMB The R code for the recoded generalized Wald test can be found at https://github.com/mafed/msWaldHMP CONTACT: Federico.Mattiello@UGent.be. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Maniatis Danae
2010-12-01
Full Text Available Abstract Background Developing countries that are willing to participate in the recently adopted (16th Session of the Conference of Parties (COP in Cancun mitigation mechanism of Reducing emissions from Deforestation and Forest Degradation - and the role of conservation, sustainable management of forests and enhancement of forest carbon stocks (REDD+ - will have to establish a national forest monitoring system in order to assess anthropogenic forest-related greenhouse gas emissions by sources and removals by sinks. Such a system should support the Measurement, Reporting and Verification (MRV requirement of the United Nations Framework Convention on Climate Change (UNFCCC as the REDD+ mechanism is results-based. A national forest inventory (NFI is one potential key component of such an MRV system. Following the Decision adopted during the 15th Session of the COP in Copenhagen, the most recent Intergovernmental Panel on Climate Change (IPCC Guidance and Guidelines should be used as a basis for estimating anthropogenic forest-related greenhouse gas emissions by sources and removals by sinks and changes in forest carbon stocks and area. Results First, we present the key indispensable elements of the IPCC Guidance and Guidelines that have been developed to fulfil the UNFCCC reporting requirements. This is done in order to set the framework to develop the MRV requirement in which a NFI for REDD+ implementation could be developed. Second, within this framework, we develop and propose a novel scheme for the stratification of forest land for REDD+. Finally, we present some non-exhaustive optional elements within this framework that a country could consider to successfully operationalise and implement its REDD+ NFI. Conclusion Evidently, both the methodological guidance and political decisions on REDD+ under the UNFCCC will continue to evolve. Even so, and considering that there exists decades of experience in setting up traditional NFIs, developing a NFI
Maniatis, Danae; Mollicone, Danilo
2010-12-27
Developing countries that are willing to participate in the recently adopted (16th Session of the Conference of Parties (COP) in Cancun) mitigation mechanism of Reducing emissions from Deforestation and Forest Degradation - and the role of conservation, sustainable management of forests and enhancement of forest carbon stocks (REDD+) - will have to establish a national forest monitoring system in order to assess anthropogenic forest-related greenhouse gas emissions by sources and removals by sinks. Such a system should support the Measurement, Reporting and Verification (MRV) requirement of the United Nations Framework Convention on Climate Change (UNFCCC) as the REDD+ mechanism is results-based. A national forest inventory (NFI) is one potential key component of such an MRV system. Following the Decision adopted during the 15th Session of the COP in Copenhagen, the most recent Intergovernmental Panel on Climate Change (IPCC) Guidance and Guidelines should be used as a basis for estimating anthropogenic forest-related greenhouse gas emissions by sources and removals by sinks and changes in forest carbon stocks and area. First, we present the key indispensable elements of the IPCC Guidance and Guidelines that have been developed to fulfil the UNFCCC reporting requirements. This is done in order to set the framework to develop the MRV requirement in which a NFI for REDD+ implementation could be developed. Second, within this framework, we develop and propose a novel scheme for the stratification of forest land for REDD+. Finally, we present some non-exhaustive optional elements within this framework that a country could consider to successfully operationalise and implement its REDD+ NFI. Evidently, both the methodological guidance and political decisions on REDD+ under the UNFCCC will continue to evolve. Even so, and considering that there exists decades of experience in setting up traditional NFIs, developing a NFI that a country may use to directly support REDD
Sample size estimation and sampling techniques for selecting a representative sample
Aamir Omair
2014-01-01
Full Text Available Introduction: The purpose of this article is to provide a general understanding of the concepts of sampling as applied to health-related research. Sample Size Estimation: It is important to select a representative sample in quantitative research in order to be able to generalize the results to the target population. The sample should be of the required sample size and must be selected using an appropriate probability sampling technique. There are many hidden biases which can adversely affect the outcome of the study. Important factors to consider for estimating the sample size include the size of the study population, confidence level, expected proportion of the outcome variable (for categorical variables/standard deviation of the outcome variable (for numerical variables, and the required precision (margin of accuracy from the study. The more the precision required, the greater is the required sample size. Sampling Techniques: The probability sampling techniques applied for health related research include simple random sampling, systematic random sampling, stratified random sampling, cluster sampling, and multistage sampling. These are more recommended than the nonprobability sampling techniques, because the results of the study can be generalized to the target population.
Heckmann, Tobias; Gegg, Katharina; Becht, Michael
2013-04-01
Statistical approaches to landslide susceptibility modelling on the catchment and regional scale are used very frequently compared to heuristic and physically based approaches. In the present study, we deal with the problem of the optimal sample size for a logistic regression model. More specifically, a stepwise approach has been chosen in order to select those independent variables (from a number of derivatives of a digital elevation model and landcover data) that explain best the spatial distribution of debris flow initiation zones in two neighbouring central alpine catchments in Austria (used mutually for model calculation and validation). In order to minimise problems arising from spatial autocorrelation, we sample a single raster cell from each debris flow initiation zone within an inventory. In addition, as suggested by previous work using the "rare events logistic regression" approach, we take a sample of the remaining "non-event" raster cells. The recommendations given in the literature on the size of this sample appear to be motivated by practical considerations, e.g. the time and cost of acquiring data for non-event cases, which do not apply to the case of spatial data. In our study, we aim at finding empirically an "optimal" sample size in order to avoid two problems: First, a sample too large will violate the independent sample assumption as the independent variables are spatially autocorrelated; hence, a variogram analysis leads to a sample size threshold above which the average distance between sampled cells falls below the autocorrelation range of the independent variables. Second, if the sample is too small, repeated sampling will lead to very different results, i.e. the independent variables and hence the result of a single model calculation will be extremely dependent on the choice of non-event cells. Using a Monte-Carlo analysis with stepwise logistic regression, 1000 models are calculated for a wide range of sample sizes. For each sample size
Schulze, A K S; Weisbjerg, M R; Storm, A C; Nørgaard, P
2014-06-01
The objective of this study was to investigate the effect of NDF content in highly digestible grass/clover silage on particle size reduction, ruminal stratification, and selective retention in dairy heifers. The reduction in particle size from feed to feces was evaluated and related to feed intake, chewing activity, and apparent digestibility. Four grass/clover harvests (Mixtures of Lolium perenne, Trifolium pratense, and Trifolium repens) were performed from early May to late August at different maturities, at different regrowth stages, and with different clover proportions, resulting in silages with NDF contents of 312, 360, 371, and 446 g/kg DM, respectively, and decreasing NDF digestibility with greater NDF content. Four rumen-fistulated dairy heifers were fed silage at 90% of ad libitum level as the only feed source in a 4 × 4 Latin square design. Silage, ingested feed boluses, medial and ventral ruminal digesta, and feces samples were washed with neutral detergent in nylon bags of 10-μm pore size, freeze dried, and divided into small (1 mm) particles by dry-sieving. Chewing activity, rumen pool size, and apparent digestibility were measured. Intake of NDF increased linearly from 2.3 to 2.8 kg/d with greater NDF content of forages (P = 0.01), but silages were exposed to similar eating time (P = 0.55) and rumination time per kg NDF (P = 0.35). No linear effect of NDF content was found on proportion of LP in ingested feed boluses (P = 0.31), medial rumen digesta (P = 0.95), ventral rumen digesta (P = 0.84), and feces (P = 0.09). Greater proportions of DM (P silages (P > 0.13). The LP proportion was >30% of particles in the ventral and medial rumen, whereas in the feces, the LP proportion was silages, stressing that the retention mechanism of large undigested particles lies elsewhere than with particle entrapment in the rumen mat. In this study, forage particle breakdown, ruminal stratification, and retention of particles in the rumen were not affected by NDF
Kühberger, Anton; Fritz, Astrid; Scherndl, Thomas
2014-01-01
.... We extracted p values, effect sizes, and sample sizes of all empirical papers, and calculated the correlation between effect size and sample size, and investigated the distribution of p values...
7 CFR 52.803 - Sample unit size.
2010-01-01
... 7 Agriculture 2 2010-01-01 2010-01-01 false Sample unit size. 52.803 Section 52.803 Agriculture... United States Standards for Grades of Frozen Red Tart Pitted Cherries Sample Unit Size § 52.803 Sample unit size. Compliance with requirements for size and the various quality factors is based on the...
7 CFR 52.775 - Sample unit size.
2010-01-01
... 7 Agriculture 2 2010-01-01 2010-01-01 false Sample unit size. 52.775 Section 52.775 Agriculture... United States Standards for Grades of Canned Red Tart Pitted Cherries 1 Sample Unit Size § 52.775 Sample unit size. Compliance with requirements for the size and the various quality factors is based on the...
De Preter, Katleen; Mestdagh, Pieter; Vermeulen, Joëlle; Zeka, Fjoralba; Naranjo, Arlene; Bray, Isabella; Castel, Victoria; Chen, Caifu; Drozynska, Elzbieta; Eggert, Angelika; Hogarty, Michael D; Izycka-Swieszewska, Ewa; London, Wendy B; Noguera, Rosa; Piqueras, Marta; Bryan, Kenneth; Schowe, Benjamin; van Sluis, Peter; Molenaar, Jan J; Schramm, Alexander; Schulte, Johannes H; Stallings, Raymond L; Versteeg, Rogier; Laureys, Geneviève; Van Roy, Nadine; Speleman, Frank; Vandesompele, Jo
2011-12-15
More accurate assessment of prognosis is important to further improve the choice of risk-related therapy in neuroblastoma (NB) patients. In this study, we aimed to establish and validate a prognostic miRNA signature for children with NB and tested it in both fresh frozen and archived formalin-fixed paraffin-embedded (FFPE) samples. Four hundred-thirty human mature miRNAs were profiled in two patient subgroups with maximally divergent clinical courses. Univariate logistic regression analysis was used to select miRNAs correlating with NB patient survival. A 25-miRNA gene signature was built using 51 training samples, tested on 179 test samples, and validated on an independent set of 304 fresh frozen tumor samples and 75 archived FFPE samples. The 25-miRNA signature significantly discriminates the test patients with respect to progression-free and overall survival (P risk patients. Multivariate analysis indicates that the miRNA signature is an independent predictor of patient survival after controlling for current risk factors. The results were confirmed in an external validation set. In contrast to a previously published mRNA classifier, the 25-miRNA signature was found to be predictive for patient survival in a set of 75 FFPE neuroblastoma samples. In this study, we present the largest NB miRNA expression study so far, including more than 500 NB patients. We established and validated a robust miRNA classifier, able to identify a cohort of high-risk NB patients at greater risk for adverse outcome using both fresh frozen and archived material. ©2011 AACR.
PIXE–PIGE analysis of size-segregated aerosol samples from remote areas
Calzolai, G., E-mail: calzolai@fi.infn.it [Department of Physics and Astronomy, University of Florence and National Institute of Nuclear Physics (INFN), Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); Chiari, M.; Lucarelli, F.; Nava, S.; Taccetti, F. [Department of Physics and Astronomy, University of Florence and National Institute of Nuclear Physics (INFN), Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); Becagli, S.; Frosini, D.; Traversi, R.; Udisti, R. [Department of Chemistry, University of Florence, Via della Lastruccia 3, 50019 Sesto Fiorentino (Italy)
2014-01-01
The chemical characterization of size-segregated samples is helpful to study the aerosol effects on both human health and environment. The sampling with multi-stage cascade impactors (e.g., Small Deposit area Impactor, SDI) produces inhomogeneous samples, with a multi-spot geometry and a non-negligible particle stratification. At LABEC (Laboratory of nuclear techniques for the Environment and the Cultural Heritage), an external beam line is fully dedicated to PIXE–PIGE analysis of aerosol samples. PIGE is routinely used as a sidekick of PIXE to correct the underestimation of PIXE in quantifying the concentration of the lightest detectable elements, like Na or Al, due to X-ray absorption inside the individual aerosol particles. In this work PIGE has been used to study proper attenuation correction factors for SDI samples: relevant attenuation effects have been observed also for stages collecting smaller particles, and consequent implications on the retrieved aerosol modal structure have been evidenced.
Sampling large landscapes with small-scale stratification-User's Manual
Bart, Jonathan
2011-01-01
This manual explains procedures for partitioning a large landscape into plots, assigning the plots to strata, and selecting plots in each stratum to be surveyed. These steps are referred to as the "sampling large landscapes (SLL) process." We assume that users of the manual have a moderate knowledge of ArcGIS and Microsoft ® Excel. The manual is written for a single user but in many cases, some steps will be carried out by a biologist designing the survey and some steps will be carried out by a quantitative assistant. Thus, the manual essentially may be passed back and forth between these users. The SLL process primarily has been used to survey birds, and we refer to birds as subjects of the counts. The process, however, could be used to count any objects. ®
Comparison of Bayesian Sample Size Criteria: ACC, ALC, and WOC.
Cao, Jing; Lee, J Jack; Alber, Susan
2009-12-01
A challenge for implementing performance based Bayesian sample size determination is selecting which of several methods to use. We compare three Bayesian sample size criteria: the average coverage criterion (ACC) which controls the coverage rate of fixed length credible intervals over the predictive distribution of the data, the average length criterion (ALC) which controls the length of credible intervals with a fixed coverage rate, and the worst outcome criterion (WOC) which ensures the desired coverage rate and interval length over all (or a subset of) possible datasets. For most models, the WOC produces the largest sample size among the three criteria, and sample sizes obtained by the ACC and the ALC are not the same. For Bayesian sample size determination for normal means and differences between normal means, we investigate, for the first time, the direction and magnitude of differences between the ACC and ALC sample sizes. For fixed hyperparameter values, we show that the difference of the ACC and ALC sample size depends on the nominal coverage, and not on the nominal interval length. There exists a threshold value of the nominal coverage level such that below the threshold the ALC sample size is larger than the ACC sample size, and above the threshold the ACC sample size is larger. Furthermore, the ACC sample size is more sensitive to changes in the nominal coverage. We also show that for fixed hyperparameter values, there exists an asymptotic constant ratio between the WOC sample size and the ALC (ACC) sample size. Simulation studies are conducted to show that similar relationships among the ACC, ALC, and WOC may hold for estimating binomial proportions. We provide a heuristic argument that the results can be generalized to a larger class of models.
Andersen, Elsa; Furbo, Simon
2008-01-01
heating system. High temperatures in the top of the storage tank established by the energy from the solar collector reduce the use of auxiliary energy. Low temperatures in the bottom of the storage tank improve the operation conditions for the solar collector. Using thermal stratified heat storages...... results in longer operation periods and improved utilization of the solar collector. Thermal stratification can be achieved, for example by using inlet stratification devices at all inlets to the storage tank. This paper presents how thermal stratification is established and utilized by means of inlet......Thermal stratification in the storage tank is extremely important in order to achieve high thermal performance of a solar heating system. High temperatures in the top of the storage tank and low temperatures in the bottom of the storage tank lead to the best operation conditions for any solar...
On bootstrap sample size in extreme value theory
J.L. Geluk (Jaap); L.F.M. de Haan (Laurens)
2002-01-01
textabstractIt has been known for a long time that for bootstrapping the probability distribution of the maximum of a sample consistently, the bootstrap sample size needs to be of smaller order than the original sample size. See Jun Shao and Dongsheng Tu (1995), Ex. 3.9,p. 123. We show that the same
Sample size determination in clinical trials with multiple endpoints
Sozu, Takashi; Hamasaki, Toshimitsu; Evans, Scott R
2015-01-01
This book integrates recent methodological developments for calculating the sample size and power in trials with more than one endpoint considered as multiple primary or co-primary, offering an important reference work for statisticians working in this area. The determination of sample size and the evaluation of power are fundamental and critical elements in the design of clinical trials. If the sample size is too small, important effects may go unnoticed; if the sample size is too large, it represents a waste of resources and unethically puts more participants at risk than necessary. Recently many clinical trials have been designed with more than one endpoint considered as multiple primary or co-primary, creating a need for new approaches to the design and analysis of these clinical trials. The book focuses on the evaluation of power and sample size determination when comparing the effects of two interventions in superiority clinical trials with multiple endpoints. Methods for sample size calculation in clin...
How Small Is Big: Sample Size and Skewness.
Piovesana, Adina; Senior, Graeme
2016-09-21
Sample sizes of 50 have been cited as sufficient to obtain stable means and standard deviations in normative test data. The influence of skewness on this minimum number, however, has not been evaluated. Normative test data with varying levels of skewness were compiled for 12 measures from 7 tests collected as part of ongoing normative studies in Brisbane, Australia. Means and standard deviations were computed from sample sizes of 10 to 100 drawn with replacement from larger samples of 272 to 973 cases. The minimum sample size was determined by the number at which both mean and standard deviation estimates remained within the 90% confidence intervals surrounding the population estimates. Sample sizes of greater than 85 were found to generate stable means and standard deviations regardless of the level of skewness, with smaller samples required in skewed distributions. A formula was derived to compute recommended sample size at differing levels of skewness.
Cutoff sample size estimation for survival data: a simulation study
2014-01-01
This thesis demonstrates the possible cutoff sample size point that balances goodness of es-timation and study expenditure by a practical cancer case. As it is crucial to determine the sample size in designing an experiment, researchers attempt to find the suitable sample size that achieves desired power and budget efficiency at the same time. The thesis shows how simulation can be used for sample size and precision calculations with survival data. The pre-sentation concentrates on the simula...
Improved Design of High Viscous Crude Oil Stratification Sampling Facility%超稠原油大罐分层取样装置改进型设计
宋传阳
2014-01-01
Stratification sampling facility is always adopted in high viscous liquid production at home and abroad. In this paper, based on practical application, aiming at problems of old stratification sampling facility, a new improved stratification sampling facility was designed. The improved facility is excellent in convenient use, easy operation and high sampling efficiency.%超稠原油大罐分层取样装置一直是国内外高粘液体生产中常用的大罐取样装置，本文结合现场实际应用，针对原有分层取样装置提手、取样尺宜断裂、取样瓶宜掉落等问题，设计出一种改进型大罐分层取样装置，该装置使用方便、易于操作、取样效果好，有效满足了油田及化工日常生产的需要。
Downs, Timothy J.; Ogneva-Himmelberger, Yelena; Aupont, Onesky; Wang, Yangyang; Raj, Ann; Zimmerman, Paula; Goble, Robert; Taylor, Octavia; Churchill, Linda; Lemay, Celeste; McLaughlin, Thomas; Felice, Marianne
2010-01-01
Background The National Children’s Study is the most ambitious study ever attempted in the United States to assess how environmental factors impact child health and development. It aims to follow 100,000 children from gestation until 21 years of age. Success requires breaking new interdisciplinary ground, starting with how to select the sample of > 1,000 children in each of 105 study sites; no standardized protocol exists for stratification of the target population by factoring in the diverse environments it inhabits. Worcester County, Massachusetts, like other sites, stratifies according to local conditions and local knowledge, subject to probability sampling rules. Objectives We answer the following questions: How do we divide Worcester County into viable strata that represent its health-relevant environmental and sociodemographic heterogeneity, subject to sampling rules? What potential does our approach have to inform stratification at other sites? Results We developed a multivariable, vulnerability-based method for spatial sampling consisting of two descriptive indices: a hazards/stressors exposure index (comprising three proxy variables), and an adaptive capacity/sociodemographic character index (five variables). Multivariable, health-relevant stratification at the start of the study may improve detection power for environment–child health associations down the line. Eighteen strata capture countywide heterogeneity in the indices and have optimal relative homogeneity within each. They achieve comparable expected birth counts and conform to local concepts of space. Conclusion The approach offers moderate to high potential to inform other sites, limited by intersite differences in data availability, geodemographics, and technical capacity. Energetic community engagement from the start promotes local stratification coherence, plus vital researcher–community trust and co-ownership for sustainability. PMID:20211802
Anton Kühberger
Full Text Available The p value obtained from a significance test provides no information about the magnitude or importance of the underlying phenomenon. Therefore, additional reporting of effect size is often recommended. Effect sizes are theoretically independent from sample size. Yet this may not hold true empirically: non-independence could indicate publication bias.We investigate whether effect size is independent from sample size in psychological research. We randomly sampled 1,000 psychological articles from all areas of psychological research. We extracted p values, effect sizes, and sample sizes of all empirical papers, and calculated the correlation between effect size and sample size, and investigated the distribution of p values.We found a negative correlation of r = -.45 [95% CI: -.53; -.35] between effect size and sample size. In addition, we found an inordinately high number of p values just passing the boundary of significance. Additional data showed that neither implicit nor explicit power analysis could account for this pattern of findings.The negative correlation between effect size and samples size, and the biased distribution of p values indicate pervasive publication bias in the entire field of psychology.
Kühberger, Anton; Fritz, Astrid; Scherndl, Thomas
2014-01-01
Background The p value obtained from a significance test provides no information about the magnitude or importance of the underlying phenomenon. Therefore, additional reporting of effect size is often recommended. Effect sizes are theoretically independent from sample size. Yet this may not hold true empirically: non-independence could indicate publication bias. Methods We investigate whether effect size is independent from sample size in psychological research. We randomly sampled 1,000 psychological articles from all areas of psychological research. We extracted p values, effect sizes, and sample sizes of all empirical papers, and calculated the correlation between effect size and sample size, and investigated the distribution of p values. Results We found a negative correlation of r = −.45 [95% CI: −.53; −.35] between effect size and sample size. In addition, we found an inordinately high number of p values just passing the boundary of significance. Additional data showed that neither implicit nor explicit power analysis could account for this pattern of findings. Conclusion The negative correlation between effect size and samples size, and the biased distribution of p values indicate pervasive publication bias in the entire field of psychology. PMID:25192357
40 CFR 80.127 - Sample size guidelines.
2010-07-01
... 40 Protection of Environment 16 2010-07-01 2010-07-01 false Sample size guidelines. 80.127 Section 80.127 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) REGULATION OF FUELS AND FUEL ADDITIVES Attest Engagements § 80.127 Sample size guidelines. In performing...
Approaches to sample size determination for multivariate data
Saccenti, Edoardo; Timmerman, Marieke E.
2016-01-01
Sample size determination is a fundamental step in the design of experiments. Methods for sample size determination are abundant for univariate analysis methods, but scarce in the multivariate case. Omics data are multivariate in nature and are commonly investigated using multivariate statistical
Sample Size Requirements for Estimating Pearson, Spearman and Kendall Correlations.
Bonett, Douglas G.; Wright, Thomas A.
2000-01-01
Reviews interval estimates of the Pearson, Kendall tau-alpha, and Spearman correlates and proposes an improved standard error for the Spearman correlation. Examines the sample size required to yield a confidence interval having the desired width. Findings show accurate results from a two-stage approximation to the sample size. (SLD)
Determination of sample size in genome-scale RNAi screens.
Zhang, Xiaohua Douglas; Heyse, Joseph F
2009-04-01
For genome-scale RNAi research, it is critical to investigate sample size required for the achievement of reasonably low false negative rate (FNR) and false positive rate. The analysis in this article reveals that current design of sample size contributes to the occurrence of low signal-to-noise ratio in genome-scale RNAi projects. The analysis suggests that (i) an arrangement of 16 wells per plate is acceptable and an arrangement of 20-24 wells per plate is preferable for a negative control to be used for hit selection in a primary screen without replicates; (ii) in a confirmatory screen or a primary screen with replicates, a sample size of 3 is not large enough, and there is a large reduction in FNRs when sample size increases from 3 to 4. To search a tradeoff between benefit and cost, any sample size between 4 and 11 is a reasonable choice. If the main focus is the selection of siRNAs with strong effects, a sample size of 4 or 5 is a good choice. If we want to have enough power to detect siRNAs with moderate effects, sample size needs to be 8, 9, 10 or 11. These discoveries about sample size bring insight to the design of a genome-scale RNAi screen experiment.
Power Analysis and Sample Size Determination in Metabolic Phenotyping.
Blaise, Benjamin J; Correia, Gonçalo; Tin, Adrienne; Young, J Hunter; Vergnaud, Anne-Claire; Lewis, Matthew; Pearce, Jake T M; Elliott, Paul; Nicholson, Jeremy K; Holmes, Elaine; Ebbels, Timothy M D
2016-05-17
Estimation of statistical power and sample size is a key aspect of experimental design. However, in metabolic phenotyping, there is currently no accepted approach for these tasks, in large part due to the unknown nature of the expected effect. In such hypothesis free science, neither the number or class of important analytes nor the effect size are known a priori. We introduce a new approach, based on multivariate simulation, which deals effectively with the highly correlated structure and high-dimensionality of metabolic phenotyping data. First, a large data set is simulated based on the characteristics of a pilot study investigating a given biomedical issue. An effect of a given size, corresponding either to a discrete (classification) or continuous (regression) outcome is then added. Different sample sizes are modeled by randomly selecting data sets of various sizes from the simulated data. We investigate different methods for effect detection, including univariate and multivariate techniques. Our framework allows us to investigate the complex relationship between sample size, power, and effect size for real multivariate data sets. For instance, we demonstrate for an example pilot data set that certain features achieve a power of 0.8 for a sample size of 20 samples or that a cross-validated predictivity QY(2) of 0.8 is reached with an effect size of 0.2 and 200 samples. We exemplify the approach for both nuclear magnetic resonance and liquid chromatography-mass spectrometry data from humans and the model organism C. elegans.
[Effect sizes, statistical power and sample sizes in "the Japanese Journal of Psychology"].
Suzukawa, Yumi; Toyoda, Hideki
2012-04-01
This study analyzed the statistical power of research studies published in the "Japanese Journal of Psychology" in 2008 and 2009. Sample effect sizes and sample statistical powers were calculated for each statistical test and analyzed with respect to the analytical methods and the fields of the studies. The results show that in the fields like perception, cognition or learning, the effect sizes were relatively large, although the sample sizes were small. At the same time, because of the small sample sizes, some meaningful effects could not be detected. In the other fields, because of the large sample sizes, meaningless effects could be detected. This implies that researchers who could not get large enough effect sizes would use larger samples to obtain significant results.
Fearon, Elizabeth; Chabata, Sungai T; Thompson, Jennifer A; Cowan, Frances M; Hargreaves, James R
2017-09-14
While guidance exists for obtaining population size estimates using multiplier methods with respondent-driven sampling surveys, we lack specific guidance for making sample size decisions. To guide the design of multiplier method population size estimation studies using respondent-driven sampling surveys to reduce the random error around the estimate obtained. The population size estimate is obtained by dividing the number of individuals receiving a service or the number of unique objects distributed (M) by the proportion of individuals in a representative survey who report receipt of the service or object (P). We have developed an approach to sample size calculation, interpreting methods to estimate the variance around estimates obtained using multiplier methods in conjunction with research into design effects and respondent-driven sampling. We describe an application to estimate the number of female sex workers in Harare, Zimbabwe. There is high variance in estimates. Random error around the size estimate reflects uncertainty from M and P, particularly when the estimate of P in the respondent-driven sampling survey is low. As expected, sample size requirements are higher when the design effect of the survey is assumed to be greater. We suggest a method for investigating the effects of sample size on the precision of a population size estimate obtained using multipler methods and respondent-driven sampling. Uncertainty in the size estimate is high, particularly when P is small, so balancing against other potential sources of bias, we advise researchers to consider longer service attendance reference periods and to distribute more unique objects, which is likely to result in a higher estimate of P in the respondent-driven sampling survey.
Venkateswara Rao, Gottumukkala; Markandeya, Ravvala; Sharma, Satish Kumar
2017-04-01
Experiments were carried out with two different sizes of (-30 + 6 and -6 + 1 mm) sub grade iron ore sample from Deposit No. 10 and 11A, Bacheli Complex, Bailadila, India to study the stratification behaviour at optimised parameters in a under bed air pulsed jig at 1, 2, 5, 10, 15 and 20 minutes residence time. This paper deals with the rate at which stratification takes place and determines the optimum stratification time (residence time) for above two size fractions. Average apparent density along with Jig Stratification Index (JSI) of both the size fractions was calculated. It was observed that the stratification rate is high for fines (-6 + 1 mm) and stratification index was higher for lump (-30 + 6 mm) when compared with the other size fraction. The maximum JSI observed was 0.35 for lump (-30 + 6 mm) and 0.30 for fines (-6 + 1 mm).
Sample size calculation for comparing two negative binomial rates.
Zhu, Haiyuan; Lakkis, Hassan
2014-02-10
Negative binomial model has been increasingly used to model the count data in recent clinical trials. It is frequently chosen over Poisson model in cases of overdispersed count data that are commonly seen in clinical trials. One of the challenges of applying negative binomial model in clinical trial design is the sample size estimation. In practice, simulation methods have been frequently used for sample size estimation. In this paper, an explicit formula is developed to calculate sample size based on the negative binomial model. Depending on different approaches to estimate the variance under null hypothesis, three variations of the sample size formula are proposed and discussed. Important characteristics of the formula include its accuracy and its ability to explicitly incorporate dispersion parameter and exposure time. The performance of the formula with each variation is assessed using simulations.
Determination of the optimal sample size for a clinical trial accounting for the population size.
Stallard, Nigel; Miller, Frank; Day, Simon; Hee, Siew Wan; Madan, Jason; Zohar, Sarah; Posch, Martin
2017-07-01
The problem of choosing a sample size for a clinical trial is a very common one. In some settings, such as rare diseases or other small populations, the large sample sizes usually associated with the standard frequentist approach may be infeasible, suggesting that the sample size chosen should reflect the size of the population under consideration. Incorporation of the population size is possible in a decision-theoretic approach either explicitly by assuming that the population size is fixed and known, or implicitly through geometric discounting of the gain from future patients reflecting the expected population size. This paper develops such approaches. Building on previous work, an asymptotic expression is derived for the sample size for single and two-arm clinical trials in the general case of a clinical trial with a primary endpoint with a distribution of one parameter exponential family form that optimizes a utility function that quantifies the cost and gain per patient as a continuous function of this parameter. It is shown that as the size of the population, N, or expected size, N∗ in the case of geometric discounting, becomes large, the optimal trial size is O(N1/2) or O(N∗1/2). The sample size obtained from the asymptotic expression is also compared with the exact optimal sample size in examples with responses with Bernoulli and Poisson distributions, showing that the asymptotic approximations can also be reasonable in relatively small sample sizes. © 2016 The Author. Biometrical Journal published by WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
On power and sample size calculation in ethnic sensitivity studies.
Zhang, Wei; Sethuraman, Venkat
2011-01-01
In ethnic sensitivity studies, it is of interest to know whether the same dose has the same effect over populations in different regions. Glasbrenner and Rosenkranz (2006) proposed a criterion for ethnic sensitivity studies in the context of different dose-exposure models. Their method is liberal in the sense that their sample size will not achieve the target power. We will show that the power function can be easily calculated by numeric integration, and the sample size can be determined by bisection.
Sample Size Determination: A Comparison of Attribute, Continuous Variable, and Cell Size Methods.
Clark, Philip M.
1984-01-01
Describes three methods of sample size determination, each having its use in investigation of social science problems: Attribute method; Continuous Variable method; Galtung's Cell Size method. Statistical generalization, benefits of cell size method (ease of use, trivariate analysis and trichotyomized variables), and choice of method are…
Size selective sampling using mobile, 3D nanoporous membranes.
Randall, Christina L; Gillespie, Aubri; Singh, Siddarth; Leong, Timothy G; Gracias, David H
2009-02-01
We describe the fabrication of 3D membranes with precisely patterned surface nanoporosity and their utilization in size selective sampling. The membranes were self-assembled as porous cubes from lithographically fabricated 2D templates (Leong et al., Langmuir 23:8747-8751, 2007) with face dimensions of 200 microm, volumes of 8 nL, and monodisperse pores ranging in size from approximately 10 microm to 100 nm. As opposed to conventional sampling and filtration schemes where fluid is moved across a static membrane, we demonstrate sampling by instead moving the 3D nanoporous membrane through the fluid. This new scheme allows for straightforward sampling in small volumes, with little to no loss. Membranes with five porous faces and one open face were moved through fluids to sample and retain nanoscale beads and cells based on pore size. Additionally, cells retained within the membranes were subsequently cultured and multiplied using standard cell culture protocols upon retrieval.
Calculating sample size in trials using historical controls.
Zhang, Song; Cao, Jing; Ahn, Chul
2010-08-01
Makuch and Simon [Sample size considerations for non-randomised comparative studies. J Chronic Dis 1980; 33: 175-81.] developed a sample size formula for historical control trials. When assessing power, they assumed the true control treatment effect to be equal to the observed effect from the historical control group. Many researchers have pointed out that the Makuch-Simon approach does not preserve the nominal power and type I error when considering the uncertainty in the true historical control treatment effect. To develop a sample size formula that properly accounts for the underlying randomness in the observations from the historical control group. We reveal the extremely skewed nature in the distributions of power and type I error, obtained over all the random realizations of the historical control data. The skewness motivates us to derive a sample size formula that controls the percentiles, instead of the means, of the power and type I error. A closed-form sample size formula is developed to control arbitrary percentiles of power and type I error for historical control trials. A simulation study further demonstrates that this approach preserves the operational characteristics in a more realistic scenario where the population variances are unknown and replaced by sample variances. The closed-form sample size formula is derived for continuous outcomes. The formula is more complicated for binary or survival time outcomes. We have derived a closed-form sample size formula that controls the percentiles instead of means of power and type I error in historical control trials, which have extremely skewed distributions over all the possible realizations of historical control data.
Sample Size Requirements for Traditional and Regression-Based Norms.
Oosterhuis, Hannah E M; van der Ark, L Andries; Sijtsma, Klaas
2016-04-01
Test norms enable determining the position of an individual test taker in the group. The most frequently used approach to obtain test norms is traditional norming. Regression-based norming may be more efficient than traditional norming and is rapidly growing in popularity, but little is known about its technical properties. A simulation study was conducted to compare the sample size requirements for traditional and regression-based norming by examining the 95% interpercentile ranges for percentile estimates as a function of sample size, norming method, size of covariate effects on the test score, test length, and number of answer categories in an item. Provided the assumptions of the linear regression model hold in the data, for a subdivision of the total group into eight equal-size subgroups, we found that regression-based norming requires samples 2.5 to 5.5 times smaller than traditional norming. Sample size requirements are presented for each norming method, test length, and number of answer categories. We emphasize that additional research is needed to establish sample size requirements when the assumptions of the linear regression model are violated.
Conservative Sample Size Determination for Repeated Measures Analysis of Covariance.
Morgan, Timothy M; Case, L Douglas
2013-07-05
In the design of a randomized clinical trial with one pre and multiple post randomized assessments of the outcome variable, one needs to account for the repeated measures in determining the appropriate sample size. Unfortunately, one seldom has a good estimate of the variance of the outcome measure, let alone the correlations among the measurements over time. We show how sample sizes can be calculated by making conservative assumptions regarding the correlations for a variety of covariance structures. The most conservative choice for the correlation depends on the covariance structure and the number of repeated measures. In the absence of good estimates of the correlations, the sample size is often based on a two-sample t-test, making the 'ultra' conservative and unrealistic assumption that there are zero correlations between the baseline and follow-up measures while at the same time assuming there are perfect correlations between the follow-up measures. Compared to the case of taking a single measurement, substantial savings in sample size can be realized by accounting for the repeated measures, even with very conservative assumptions regarding the parameters of the assumed correlation matrix. Assuming compound symmetry, the sample size from the two-sample t-test calculation can be reduced at least 44%, 56%, and 61% for repeated measures analysis of covariance by taking 2, 3, and 4 follow-up measures, respectively. The results offer a rational basis for determining a fairly conservative, yet efficient, sample size for clinical trials with repeated measures and a baseline value.
Anton Kühberger; Astrid Fritz; Thomas Scherndl
2014-01-01
.... We extracted p values, effect sizes, and sample sizes of all empirical papers, and calculated the correlation between effect size and sample size, and investigated the distribution of p values...
Estimating hidden population size using Respondent-Driven Sampling data.
Handcock, Mark S; Gile, Krista J; Mar, Corinne M
Respondent-Driven Sampling (RDS) is n approach to sampling design and inference in hard-to-reach human populations. It is often used in situations where the target population is rare and/or stigmatized in the larger population, so that it is prohibitively expensive to contact them through the available frames. Common examples include injecting drug users, men who have sex with men, and female sex workers. Most analysis of RDS data has focused on estimating aggregate characteristics, such as disease prevalence. However, RDS is often conducted in settings where the population size is unknown and of great independent interest. This paper presents an approach to estimating the size of a target population based on data collected through RDS. The proposed approach uses a successive sampling approximation to RDS to leverage information in the ordered sequence of observed personal network sizes. The inference uses the Bayesian framework, allowing for the incorporation of prior knowledge. A flexible class of priors for the population size is used that aids elicitation. An extensive simulation study provides insight into the performance of the method for estimating population size under a broad range of conditions. A further study shows the approach also improves estimation of aggregate characteristics. Finally, the method demonstrates sensible results when used to estimate the size of known networked populations from the National Longitudinal Study of Adolescent Health, and when used to estimate the size of a hard-to-reach population at high risk for HIV.
Anderson, Samantha F; Kelley, Ken; Maxwell, Scott E
2017-09-01
The sample size necessary to obtain a desired level of statistical power depends in part on the population value of the effect size, which is, by definition, unknown. A common approach to sample-size planning uses the sample effect size from a prior study as an estimate of the population value of the effect to be detected in the future study. Although this strategy is intuitively appealing, effect-size estimates, taken at face value, are typically not accurate estimates of the population effect size because of publication bias and uncertainty. We show that the use of this approach often results in underpowered studies, sometimes to an alarming degree. We present an alternative approach that adjusts sample effect sizes for bias and uncertainty, and we demonstrate its effectiveness for several experimental designs. Furthermore, we discuss an open-source R package, BUCSS, and user-friendly Web applications that we have made available to researchers so that they can easily implement our suggested methods.
Sample size considerations for historical control studies with survival outcomes
Zhu, Hong; Zhang, Song; Ahn, Chul
2015-01-01
Historical control trials (HCTs) are frequently conducted to compare an experimental treatment with a control treatment from a previous study, when they are applicable and favored over a randomized clinical trial (RCT) due to feasibility, ethics and cost concerns. Makuch and Simon developed a sample size formula for historical control (HC) studies with binary outcomes, assuming that the observed response rate in the HC group is the true response rate. This method was extended by Dixon and Simon to specify sample size for HC studies comparing survival outcomes. For HC studies with binary and continuous outcomes, many researchers have shown that the popular Makuch and Simon method does not preserve the nominal power and type I error, and suggested alternative approaches. For HC studies with survival outcomes, we reveal through simulation that the conditional power and type I error over all the random realizations of the HC data have highly skewed distributions. Therefore, the sampling variability of the HC data needs to be appropriately accounted for in determining sample size. A flexible sample size formula that controls arbitrary percentiles, instead of means, of the conditional power and type I error, is derived. Although an explicit sample size formula with survival outcomes is not available, the computation is straightforward. Simulations demonstrate that the proposed method preserves the operational characteristics in a more realistic scenario where the true hazard rate of the HC group is unknown. A real data application of an advanced non-small cell lung cancer (NSCLC) clinical trial is presented to illustrate sample size considerations for HC studies in comparison of survival outcomes. PMID:26098200
Current sample size conventions: Flaws, harms, and alternatives
Bacchetti Peter
2010-03-01
Full Text Available Abstract Background The belief remains widespread that medical research studies must have statistical power of at least 80% in order to be scientifically sound, and peer reviewers often question whether power is high enough. Discussion This requirement and the methods for meeting it have severe flaws. Notably, the true nature of how sample size influences a study's projected scientific or practical value precludes any meaningful blanket designation of value of information methods, simple choices based on cost or feasibility that have recently been justified, sensitivity analyses that examine a meaningful array of possible findings, and following previous analogous studies. To promote more rational approaches, research training should cover the issues presented here, peer reviewers should be extremely careful before raising issues of "inadequate" sample size, and reports of completed studies should not discuss power. Summary Common conventions and expectations concerning sample size are deeply flawed, cause serious harm to the research process, and should be replaced by more rational alternatives.
Power and Sample Size Calculations for Contrast Analysis in ANCOVA.
Shieh, Gwowen
2017-01-01
Analysis of covariance (ANCOVA) is commonly used in behavioral and educational research to reduce the error variance and improve the power of analysis of variance by adjusting the covariate effects. For planning and evaluating randomized ANCOVA designs, a simple sample-size formula has been proposed to account for the variance deflation factor in the comparison of two treatment groups. The objective of this article is to highlight an overlooked and potential problem of the exiting approximation and to provide an alternative and exact solution of power and sample size assessments for testing treatment contrasts. Numerical investigations are conducted to reveal the relative performance of the two procedures as a reliable technique to accommodate the covariate features that make ANCOVA design particularly distinctive. The described approach has important advantages over the current method in general applicability, methodological justification, and overall accuracy. To enhance the practical usefulness, computer algorithms are presented to implement the recommended power calculations and sample-size determinations.
Sample size in psychological research over the past 30 years.
Marszalek, Jacob M; Barber, Carolyn; Kohlhart, Julie; Holmes, Cooper B
2011-04-01
The American Psychological Association (APA) Task Force on Statistical Inference was formed in 1996 in response to a growing body of research demonstrating methodological issues that threatened the credibility of psychological research, and made recommendations to address them. One issue was the small, even dramatically inadequate, size of samples used in studies published by leading journals. The present study assessed the progress made since the Task Force's final report in 1999. Sample sizes reported in four leading APA journals in 1955, 1977, 1995, and 2006 were compared using nonparametric statistics, while data from the last two waves were fit to a hierarchical generalized linear growth model for more in-depth analysis. Overall, results indicate that the recommendations for increasing sample sizes have not been integrated in core psychological research, although results slightly vary by field. This and other implications are discussed in the context of current methodological critique and practice.
Heidel, R. Eric
2016-01-01
Statistical power is the ability to detect a significant effect, given that the effect actually exists in a population. Like most statistical concepts, statistical power tends to induce cognitive dissonance in hepatology researchers. However, planning for statistical power by an a priori sample size calculation is of paramount importance when designing a research study. There are five specific empirical components that make up an a priori sample size calculation: the scale of measurement of the outcome, the research design, the magnitude of the effect size, the variance of the effect size, and the sample size. A framework grounded in the phenomenon of isomorphism, or interdependencies amongst different constructs with similar forms, will be presented to understand the isomorphic effects of decisions made on each of the five aforementioned components of statistical power. PMID:27073717
R. Eric Heidel
2016-01-01
Full Text Available Statistical power is the ability to detect a significant effect, given that the effect actually exists in a population. Like most statistical concepts, statistical power tends to induce cognitive dissonance in hepatology researchers. However, planning for statistical power by an a priori sample size calculation is of paramount importance when designing a research study. There are five specific empirical components that make up an a priori sample size calculation: the scale of measurement of the outcome, the research design, the magnitude of the effect size, the variance of the effect size, and the sample size. A framework grounded in the phenomenon of isomorphism, or interdependencies amongst different constructs with similar forms, will be presented to understand the isomorphic effects of decisions made on each of the five aforementioned components of statistical power.
Sample size in orthodontic randomized controlled trials: are numbers justified?
Koletsi, Despina; Pandis, Nikolaos; Fleming, Padhraig S
2014-02-01
Sample size calculations are advocated by the Consolidated Standards of Reporting Trials (CONSORT) group to justify sample sizes in randomized controlled trials (RCTs). This study aimed to analyse the reporting of sample size calculations in trials published as RCTs in orthodontic speciality journals. The performance of sample size calculations was assessed and calculations verified where possible. Related aspects, including number of authors; parallel, split-mouth, or other design; single- or multi-centre study; region of publication; type of data analysis (intention-to-treat or per-protocol basis); and number of participants recruited and lost to follow-up, were considered. Of 139 RCTs identified, complete sample size calculations were reported in 41 studies (29.5 per cent). Parallel designs were typically adopted (n = 113; 81 per cent), with 80 per cent (n = 111) involving two arms and 16 per cent having three arms. Data analysis was conducted on an intention-to-treat (ITT) basis in a small minority of studies (n = 18; 13 per cent). According to the calculations presented, overall, a median of 46 participants were required to demonstrate sufficient power to highlight meaningful differences (typically at a power of 80 per cent). The median number of participants recruited was 60, with a median of 4 participants being lost to follow-up. Our finding indicates good agreement between projected numbers required and those verified (median discrepancy: 5.3 per cent), although only a minority of trials (29.5 per cent) could be examined. Although sample size calculations are often reported in trials published as RCTs in orthodontic speciality journals, presentation is suboptimal and in need of significant improvement.
On an Approach to Bayesian Sample Sizing in Clinical Trials
Muirhead, Robb J
2012-01-01
This paper explores an approach to Bayesian sample size determination in clinical trials. The approach falls into the category of what is often called "proper Bayesian", in that it does not mix frequentist concepts with Bayesian ones. A criterion for a "successful trial" is defined in terms of a posterior probability, its probability is assessed using the marginal distribution of the data, and this probability forms the basis for choosing sample sizes. We illustrate with a standard problem in clinical trials, that of establishing superiority of a new drug over a control.
Sample Size Calculations for Precise Interval Estimation of the Eta-Squared Effect Size
Shieh, Gwowen
2015-01-01
Analysis of variance is one of the most frequently used statistical analyses in the behavioral, educational, and social sciences, and special attention has been paid to the selection and use of an appropriate effect size measure of association in analysis of variance. This article presents the sample size procedures for precise interval estimation…
Sample size considerations for clinical research studies in nuclear cardiology.
Chiuzan, Cody; West, Erin A; Duong, Jimmy; Cheung, Ken Y K; Einstein, Andrew J
2015-12-01
Sample size calculation is an important element of research design that investigators need to consider in the planning stage of the study. Funding agencies and research review panels request a power analysis, for example, to determine the minimum number of subjects needed for an experiment to be informative. Calculating the right sample size is crucial to gaining accurate information and ensures that research resources are used efficiently and ethically. The simple question "How many subjects do I need?" does not always have a simple answer. Before calculating the sample size requirements, a researcher must address several aspects, such as purpose of the research (descriptive or comparative), type of samples (one or more groups), and data being collected (continuous or categorical). In this article, we describe some of the most frequent methods for calculating the sample size with examples from nuclear cardiology research, including for t tests, analysis of variance (ANOVA), non-parametric tests, correlation, Chi-squared tests, and survival analysis. For the ease of implementation, several examples are also illustrated via user-friendly free statistical software.
Consultants' forum: should post hoc sample size calculations be done?
Walters, Stephen J
2009-01-01
Pre-study sample size calculations for clinical trial research protocols are now mandatory. When an investigator is designing a study to compare the outcomes of an intervention, an essential step is the calculation of sample sizes that will allow a reasonable chance (power) of detecting a pre-determined difference (effect size) in the outcome variable, at a given level of statistical significance. Frequently studies will recruit fewer patients than the initial pre-study sample size calculation suggested. Investigators are faced with the fact that their study may be inadequately powered to detect the pre-specified treatment effect and the statistical analysis of the collected outcome data may or may not report a statistically significant result. If the data produces a "non-statistically significant result" then investigators are frequently tempted to ask the question "Given the actual final study size, what is the power of the study, now, to detect a treatment effect or difference?" The aim of this article is to debate whether or not it is desirable to answer this question and to undertake a power calculation, after the data have been collected and analysed.
Sample size calculation for meta-epidemiological studies.
Giraudeau, Bruno; Higgins, Julian P T; Tavernier, Elsa; Trinquart, Ludovic
2016-01-30
Meta-epidemiological studies are used to compare treatment effect estimates between randomized clinical trials with and without a characteristic of interest. To our knowledge, there is presently nothing to help researchers to a priori specify the required number of meta-analyses to be included in a meta-epidemiological study. We derived a theoretical power function and sample size formula in the framework of a hierarchical model that allows for variation in the impact of the characteristic between trials within a meta-analysis and between meta-analyses. A simulation study revealed that the theoretical function overestimated power (because of the assumption of equal weights for each trial within and between meta-analyses). We also propose a simulation approach that allows for relaxing the constraints used in the theoretical approach and is more accurate. We illustrate that the two variables that mostly influence power are the number of trials per meta-analysis and the proportion of trials with the characteristic of interest. We derived a closed-form power function and sample size formula for estimating the impact of trial characteristics in meta-epidemiological studies. Our analytical results can be used as a 'rule of thumb' for sample size calculation for a meta-epidemiologic study. A more accurate sample size can be derived with a simulation study.
(Sample) Size Matters: Defining Error in Planktic Foraminiferal Isotope Measurement
Lowery, C.; Fraass, A. J.
2015-12-01
Planktic foraminifera have been used as carriers of stable isotopic signals since the pioneering work of Urey and Emiliani. In those heady days, instrumental limitations required hundreds of individual foraminiferal tests to return a usable value. This had the fortunate side-effect of smoothing any seasonal to decadal changes within the planktic foram population, which generally turns over monthly, removing that potential noise from each sample. With the advent of more sensitive mass spectrometers, smaller sample sizes have now become standard. This has been a tremendous advantage, allowing longer time series with the same investment of time and energy. Unfortunately, the use of smaller numbers of individuals to generate a data point has lessened the amount of time averaging in the isotopic analysis and decreased precision in paleoceanographic datasets. With fewer individuals per sample, the differences between individual specimens will result in larger variation, and therefore error, and less precise values for each sample. Unfortunately, most workers (the authors included) do not make a habit of reporting the error associated with their sample size. We have created an open-source model in R to quantify the effect of sample sizes under various realistic and highly modifiable parameters (calcification depth, diagenesis in a subset of the population, improper identification, vital effects, mass, etc.). For example, a sample in which only 1 in 10 specimens is diagenetically altered can be off by >0.3‰ δ18O VPDB or ~1°C. Additionally, and perhaps more importantly, we show that under unrealistically ideal conditions (perfect preservation, etc.) it takes ~5 individuals from the mixed-layer to achieve an error of less than 0.1‰. Including just the unavoidable vital effects inflates that number to ~10 individuals to achieve ~0.1‰. Combining these errors with the typical machine error inherent in mass spectrometers make this a vital consideration moving forward.
Rock sampling. [method for controlling particle size distribution
Blum, P. (Inventor)
1971-01-01
A method for sampling rock and other brittle materials and for controlling resultant particle sizes is described. The method involves cutting grooves in the rock surface to provide a grouping of parallel ridges and subsequently machining the ridges to provide a powder specimen. The machining step may comprise milling, drilling, lathe cutting or the like; but a planing step is advantageous. Control of the particle size distribution is effected primarily by changing the height and width of these ridges. This control exceeds that obtainable by conventional grinding.
Svenkrtova, Andrea; Belicova, Lenka; Volejnikova, Andrea; Sigler, Karel; Jazwinski, S Michal; Pichova, Alena
2016-04-01
Cells of the budding yeast Saccharomyces cerevisiae undergo a process akin to differentiation during prolonged culture without medium replenishment. Various methods have been used to separate and determine the potential role and fate of the different cell species. We have stratified chronologically-aged yeast cultures into cells of different sizes, using centrifugal elutriation, and characterized these subpopulations physiologically. We distinguish two extreme cell types, very small (XS) and very large (L) cells. L cells display higher viability based on two separate criteria. They respire much more actively, but produce lower levels of reactive oxygen species (ROS). L cells are capable of dividing, albeit slowly, giving rise to XS cells which do not divide. L cells are more resistant to osmotic stress and they have higher trehalose content, a storage carbohydrate often connected to stress resistance. Depletion of trehalose by deletion of TPS2 does not affect the vital characteristics of L cells, but it improves some of these characteristics in XS cells. Therefore, we propose that the response of L and XS cells to the trehalose produced in the former differs in a way that lowers the vitality of the latter. We compare our XS- and L-fraction cell characteristics with those of cells isolated from stationary cultures by others based on density. This comparison suggests that the cells have some similarities but also differences that may prove useful in addressing whether it is the segregation or the response to trehalose that may play the predominant role in cell division from stationary culture.
Sample size cognizant detection of signals in white noise
Rao, N Raj
2007-01-01
The detection and estimation of signals in noisy, limited data is a problem of interest to many scientific and engineering communities. We present a computationally simple, sample eigenvalue based procedure for estimating the number of high-dimensional signals in white noise when there are relatively few samples. We highlight a fundamental asymptotic limit of sample eigenvalue based detection of weak high-dimensional signals from a limited sample size and discuss its implication for the detection of two closely spaced signals. This motivates our heuristic definition of the 'effective number of identifiable signals.' Numerical simulations are used to demonstrate the consistency of the algorithm with respect to the effective number of signals and the superior performance of the algorithm with respect to Wax and Kailath's "asymptotically consistent" MDL based estimator.
Power and sample size in cost-effectiveness analysis.
Laska, E M; Meisner, M; Siegel, C
1999-01-01
For resource allocation under a constrained budget, optimal decision rules for mutually exclusive programs require that the treatment with the highest incremental cost-effectiveness ratio (ICER) below a willingness-to-pay (WTP) criterion be funded. This is equivalent to determining the treatment with the smallest net health cost. The designer of a cost-effectiveness study needs to select a sample size so that the power to reject the null hypothesis, the equality of the net health costs of two treatments, is high. A recently published formula derived under normal distribution theory overstates sample-size requirements. Using net health costs, the authors present simple methods for power analysis based on conventional normal and on nonparametric statistical theory.
Estimation of individual reference intervals in small sample sizes
Hansen, Ase Marie; Garde, Anne Helene; Eller, Nanna Hurwitz
2007-01-01
of that order of magnitude for all topics in question. Therefore, new methods to estimate reference intervals for small sample sizes are needed. We present an alternative method based on variance component models. The models are based on data from 37 men and 84 women taking into account biological variation...... presented in this study. The presented method enables occupational health researchers to calculate reference intervals for specific groups, i.e. smokers versus non-smokers, etc. In conclusion, the variance component models provide an appropriate tool to estimate reference intervals based on small sample...
Hydrophobicity of soil samples and soil size fractions
Lowen, H.A.; Dudas, M.J. [Alberta Univ., Edmonton, AB (Canada). Dept. of Renewable Resources; Roy, J.L. [Imperial Oil Resources Canada, Calgary, AB (Canada); Johnson, R.L. [Alberta Research Council, Vegreville, AB (Canada); McGill, W.B. [Alberta Univ., Edmonton, AB (Canada). Dept. of Renewable Resources
2001-07-01
The inability of dry soil to absorb water droplets within 10 seconds or less is defined as soil hydrophobicity. The severity, persistence and circumstances causing it vary greatly. There is a possibility that hydrophobicity in Alberta is a symptom of crude oil spills. In this study, the authors investigated the severity of soil hydrophobicity, as determined by the molarity of ethanol droplet test (MED) and dichloromethane extractable organic (DEO) concentration. The soil samples were collected from pedons within 12 hydrophobic soil sites, located northeast from Calgary to Cold Lake, Alberta. All the sites were located at an elevation ranging from 450 metres to 990 metres above sea level. The samples contained compounds from the Chernozemic, Gleysolic, Luvisolic, and Solonetzic soil orders. The results obtained indicated that the MED and DEO were positively correlated in whole soil samples. No relationships were found between MED and DEO in soil samples divided in soil fractions. More severe hydrophobicity and lower DEO concentrations were exhibited in clay- and silt-sized particles in the less than 53 micrometres, when compared to the samples in the other fraction (between 53 and 2000 micrometres). It was concluded that hydrophobicity was not restricted to a particular soil particle size class. 5 refs., 4 figs.
Simple and multiple linear regression: sample size considerations.
Hanley, James A
2016-11-01
The suggested "two subjects per variable" (2SPV) rule of thumb in the Austin and Steyerberg article is a chance to bring out some long-established and quite intuitive sample size considerations for both simple and multiple linear regression. This article distinguishes two of the major uses of regression models that imply very different sample size considerations, neither served well by the 2SPV rule. The first is etiological research, which contrasts mean Y levels at differing "exposure" (X) values and thus tends to focus on a single regression coefficient, possibly adjusted for confounders. The second research genre guides clinical practice. It addresses Y levels for individuals with different covariate patterns or "profiles." It focuses on the profile-specific (mean) Y levels themselves, estimating them via linear compounds of regression coefficients and covariates. By drawing on long-established closed-form variance formulae that lie beneath the standard errors in multiple regression, and by rearranging them for heuristic purposes, one arrives at quite intuitive sample size considerations for both research genres. Copyright Â© 2016 Elsevier Inc. All rights reserved.
Sample size of the reference sample in a case-augmented study.
Ghosh, Palash; Dewanji, Anup
2017-05-01
The case-augmented study, in which a case sample is augmented with a reference (random) sample from the source population with only covariates information known, is becoming popular in different areas of applied science such as pharmacovigilance, ecology, and econometrics. In general, the case sample is available from some source (for example, hospital database, case registry, etc.); however, the reference sample is required to be drawn from the corresponding source population. The required minimum size of the reference sample is an important issue in this regard. In this work, we address the minimum sample size calculation and discuss related issues. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.
Sample size for monitoring sirex populations and their natural enemies
Susete do Rocio Chiarello Penteado
2016-09-01
Full Text Available The woodwasp Sirex noctilio Fabricius (Hymenoptera: Siricidae was introduced in Brazil in 1988 and became the main pest in pine plantations. It has spread to about 1.000.000 ha, at different population levels, in the states of Rio Grande do Sul, Santa Catarina, Paraná, São Paulo and Minas Gerais. Control is done mainly by using a nematode, Deladenus siricidicola Bedding (Nematoda: Neothylenchidae. The evaluation of the efficiency of natural enemies has been difficult because there are no appropriate sampling systems. This study tested a hierarchical sampling system to define the sample size to monitor the S. noctilio population and the efficiency of their natural enemies, which was found to be perfectly adequate.
Sample size for monitoring sirex populations and their natural enemies
Susete do Rocio Chiarello Penteado
2016-09-01
Full Text Available The woodwasp Sirex noctilio Fabricius (Hymenoptera: Siricidae was introduced in Brazil in 1988 and became the main pest in pine plantations. It has spread to about 1.000.000 ha, at different population levels, in the states of Rio Grande do Sul, Santa Catarina, Paraná, São Paulo and Minas Gerais. Control is done mainly by using a nematode, Deladenus siricidicola Bedding (Nematoda: Neothylenchidae. The evaluation of the efficiency of natural enemies has been difficult because there are no appropriate sampling systems. This study tested a hierarchical sampling system to define the sample size to monitor the S. noctilio population and the efficiency of their natural enemies, which was found to be perfectly adequate.
Sample size and precision in NIH peer review.
David Kaplan
Full Text Available The Working Group on Peer Review of the Advisory Committee to the Director of NIH has recommended that at least 4 reviewers should be used to assess each grant application. A sample size analysis of the number of reviewers needed to evaluate grant applications reveals that a substantially larger number of evaluators are required to provide the level of precision that is currently mandated. NIH should adjust their peer review system to account for the number of reviewers needed to provide adequate precision in their evaluations.
Enhancing sampling design in mist-net bat surveys by accounting for sample size optimization
Trevelin, Leonardo Carreira; Novaes, Roberto Leonan Morim; Colas-Rosas, Paul François; Benathar, Thayse Cristhina Melo; Peres, Carlos A.
2017-01-01
The advantages of mist-netting, the main technique used in Neotropical bat community studies to date, include logistical implementation, standardization and sampling representativeness. Nonetheless, study designs still have to deal with issues of detectability related to how different species behave and use the environment. Yet there is considerable sampling heterogeneity across available studies in the literature. Here, we approach the problem of sample size optimization. We evaluated the common sense hypothesis that the first six hours comprise the period of peak night activity for several species, thereby resulting in a representative sample for the whole night. To this end, we combined re-sampling techniques, species accumulation curves, threshold analysis, and community concordance of species compositional data, and applied them to datasets of three different Neotropical biomes (Amazonia, Atlantic Forest and Cerrado). We show that the strategy of restricting sampling to only six hours of the night frequently results in incomplete sampling representation of the entire bat community investigated. From a quantitative standpoint, results corroborated the existence of a major Sample Area effect in all datasets, although for the Amazonia dataset the six-hour strategy was significantly less species-rich after extrapolation, and for the Cerrado dataset it was more efficient. From the qualitative standpoint, however, results demonstrated that, for all three datasets, the identity of species that are effectively sampled will be inherently impacted by choices of sub-sampling schedule. We also propose an alternative six-hour sampling strategy (at the beginning and the end of a sample night) which performed better when resampling Amazonian and Atlantic Forest datasets on bat assemblages. Given the observed magnitude of our results, we propose that sample representativeness has to be carefully weighed against study objectives, and recommend that the trade-off between
Enhancing sampling design in mist-net bat surveys by accounting for sample size optimization.
Trevelin, Leonardo Carreira; Novaes, Roberto Leonan Morim; Colas-Rosas, Paul François; Benathar, Thayse Cristhina Melo; Peres, Carlos A
2017-01-01
The advantages of mist-netting, the main technique used in Neotropical bat community studies to date, include logistical implementation, standardization and sampling representativeness. Nonetheless, study designs still have to deal with issues of detectability related to how different species behave and use the environment. Yet there is considerable sampling heterogeneity across available studies in the literature. Here, we approach the problem of sample size optimization. We evaluated the common sense hypothesis that the first six hours comprise the period of peak night activity for several species, thereby resulting in a representative sample for the whole night. To this end, we combined re-sampling techniques, species accumulation curves, threshold analysis, and community concordance of species compositional data, and applied them to datasets of three different Neotropical biomes (Amazonia, Atlantic Forest and Cerrado). We show that the strategy of restricting sampling to only six hours of the night frequently results in incomplete sampling representation of the entire bat community investigated. From a quantitative standpoint, results corroborated the existence of a major Sample Area effect in all datasets, although for the Amazonia dataset the six-hour strategy was significantly less species-rich after extrapolation, and for the Cerrado dataset it was more efficient. From the qualitative standpoint, however, results demonstrated that, for all three datasets, the identity of species that are effectively sampled will be inherently impacted by choices of sub-sampling schedule. We also propose an alternative six-hour sampling strategy (at the beginning and the end of a sample night) which performed better when resampling Amazonian and Atlantic Forest datasets on bat assemblages. Given the observed magnitude of our results, we propose that sample representativeness has to be carefully weighed against study objectives, and recommend that the trade-off between
Size Matters: FTIR Spectral Analysis of Apollo Regolith Samples Exhibits Grain Size Dependence.
Martin, Dayl; Joy, Katherine; Pernet-Fisher, John; Wogelius, Roy; Morlok, Andreas; Hiesinger, Harald
2017-04-01
The Mercury Thermal Infrared Spectrometer (MERTIS) on the upcoming BepiColombo mission is designed to analyse the surface of Mercury in thermal infrared wavelengths (7-14 μm) to investigate the physical properties of the surface materials [1]. Laboratory analyses of analogue materials are useful for investigating how various sample properties alter the resulting infrared spectrum. Laboratory FTIR analysis of Apollo fine (60%) causes a 'flattening' of the spectrum, with reduced reflectance in the Reststrahlen Band region (RB) as much as 30% in comparison to samples that are dominated by a high proportion of crystalline material. Apollo 15401,147 is an immature regolith with a high proportion of volcanic glass pyroclastic beads [2]. The high mafic mineral content results in a systematic shift in the Christiansen Feature (CF - the point of lowest reflectance) to longer wavelength: 8.6 μm. The glass beads dominate the spectrum, displaying a broad peak around the main Si-O stretch band (at 10.8 μm). As such, individual mineral components of this sample cannot be resolved from the average spectrum alone. Apollo 67481,96 is a sub-mature regolith composed dominantly of anorthite plagioclase [2]. The CF position of the average spectrum is shifted to shorter wavelengths (8.2 μm) due to the higher proportion of felsic minerals. Its average spectrum is dominated by anorthite reflectance bands at 8.7, 9.1, 9.8, and 10.8 μm. The average reflectance is greater than the other samples due to a lower proportion of glassy material. In each soil, the smallest fractions (0-25 and 25-63 μm) have CF positions 0.1-0.4 μm higher than the larger grain sizes. Also, the bulk-sample spectra mostly closely resemble the 0-25 μm sieved size fraction spectrum, indicating that this size fraction of each sample dominates the bulk spectrum regardless of other physical properties. This has implications for surface analyses of other Solar System bodies where some mineral phases or components
Ross-Innes, Caryn S; Chettouh, Hamza; Achilleos, Achilleas; Galeano-Dalmau, Nuria; Debiram-Beecham, Irene; MacRae, Shona; Fessas, Petros; Walker, Elaine; Varghese, Sibu; Evan, Theodore; Lao-Sirieix, Pierre S; O'Donovan, Maria; Malhotra, Shalini; Novelli, Marco; Disep, Babett; Kaye, Phillip V; Lovat, Laurence B; Haidry, Rehan; Griffin, Michael; Ragunath, Krish; Bhandari, Pradeep; Haycock, Adam; Morris, Danielle; Attwood, Stephen; Dhar, Anjan; Rees, Colin; Rutter, Matt D; Ostler, Richard; Aigret, Benoit; Sasieni, Peter D; Fitzgerald, Rebecca C
2017-01-01
Barrett's oesophagus predisposes to adenocarcinoma. However, most patients with Barrett's oesophagus will not progress and endoscopic surveillance is invasive, expensive, and fraught by issues of sampling bias and the subjective assessment of dysplasia. We investigated whether a non-endoscopic device, the Cytosponge, could be coupled with clinical and molecular biomarkers to identify a group of patients with low risk of progression suitable for non-endoscopic follow-up. In this multicentre cohort study (BEST2), patients with Barrett's oesophagus underwent the Cytosponge test before their surveillance endoscopy. We collected clinical and demographic data and tested Cytosponge samples for a molecular biomarker panel including three protein biomarkers (P53, c-Myc, and Aurora kinase A), two methylation markers (MYOD1 and RUNX3), glandular atypia, and TP53 mutation status. We used a multivariable logistic regression model to compute the conditional probability of dysplasia status. We selected a simple model with high classification accuracy and applied it to an independent validation cohort. The BEST2 study is registered with ISRCTN, number 12730505. The discovery cohort consisted of 468 patients with Barrett's oesophagus and intestinal metaplasia. Of these, 376 had no dysplasia and 22 had high-grade dysplasia or intramucosal adenocarcinoma. In the discovery cohort, a model with high classification accuracy consisted of glandular atypia, P53 abnormality, and Aurora kinase A positivity, and the interaction of age, waist-to-hip ratio, and length of the Barrett's oesophagus segment. 162 (35%) of 468 of patients fell into the low-risk category and the probability of being a true non-dysplastic patient was 100% (99% CI 96-100) and the probability of having high-grade dysplasia or intramucosal adenocarcinoma was 0% (0-4). 238 (51%) of participants were classified as of moderate risk; the probability of having high-grade dysplasia was 14% (9-21). 58 (12%) of participants were
MetSizeR: selecting the optimal sample size for metabolomic studies using an analysis based approach
2013-01-01
Background Determining sample sizes for metabolomic experiments is important but due to the complexity of these experiments, there are currently no standard methods for sample size estimation in metabolomics. Since pilot studies are rarely done in metabolomics, currently existing sample size estimation approaches which rely on pilot data can not be applied. Results In this article, an analysis based approach called MetSizeR is developed to estimate sample size for metabolomic experiments even when experimental pilot data are not available. The key motivation for MetSizeR is that it considers the type of analysis the researcher intends to use for data analysis when estimating sample size. MetSizeR uses information about the data analysis technique and prior expert knowledge of the metabolomic experiment to simulate pilot data from a statistical model. Permutation based techniques are then applied to the simulated pilot data to estimate the required sample size. Conclusions The MetSizeR methodology, and a publicly available software package which implements the approach, are illustrated through real metabolomic applications. Sample size estimates, informed by the intended statistical analysis technique, and the associated uncertainty are provided. PMID:24261687
Comparing Server Energy Use and Efficiency Using Small Sample Sizes
Coles, Henry C.; Qin, Yong; Price, Phillip N.
2014-11-01
This report documents a demonstration that compared the energy consumption and efficiency of a limited sample size of server-type IT equipment from different manufacturers by measuring power at the server power supply power cords. The results are specific to the equipment and methods used. However, it is hoped that those responsible for IT equipment selection can used the methods described to choose models that optimize energy use efficiency. The demonstration was conducted in a data center at Lawrence Berkeley National Laboratory in Berkeley, California. It was performed with five servers of similar mechanical and electronic specifications; three from Intel and one each from Dell and Supermicro. Server IT equipment is constructed using commodity components, server manufacturer-designed assemblies, and control systems. Server compute efficiency is constrained by the commodity component specifications and integration requirements. The design freedom, outside of the commodity component constraints, provides room for the manufacturer to offer a product with competitive efficiency that meets market needs at a compelling price. A goal of the demonstration was to compare and quantify the server efficiency for three different brands. The efficiency is defined as the average compute rate (computations per unit of time) divided by the average energy consumption rate. The research team used an industry standard benchmark software package to provide a repeatable software load to obtain the compute rate and provide a variety of power consumption levels. Energy use when the servers were in an idle state (not providing computing work) were also measured. At high server compute loads, all brands, using the same key components (processors and memory), had similar results; therefore, from these results, it could not be concluded that one brand is more efficient than the other brands. The test results show that the power consumption variability caused by the key components as a
Eldridge, Sandra M; Ashby, Deborah; Kerry, Sally
2006-10-01
Cluster randomized trials are increasingly popular. In many of these trials, cluster sizes are unequal. This can affect trial power, but standard sample size formulae for these trials ignore this. Previous studies addressing this issue have mostly focused on continuous outcomes or methods that are sometimes difficult to use in practice. We show how a simple formula can be used to judge the possible effect of unequal cluster sizes for various types of analyses and both continuous and binary outcomes. We explore the practical estimation of the coefficient of variation of cluster size required in this formula and demonstrate the formula's performance for a hypothetical but typical trial randomizing UK general practices. The simple formula provides a good estimate of sample size requirements for trials analysed using cluster-level analyses weighting by cluster size and a conservative estimate for other types of analyses. For trials randomizing UK general practices the coefficient of variation of cluster size depends on variation in practice list size, variation in incidence or prevalence of the medical condition under examination, and practice and patient recruitment strategies, and for many trials is expected to be approximately 0.65. Individual-level analyses can be noticeably more efficient than some cluster-level analyses in this context. When the coefficient of variation is <0.23, the effect of adjustment for variable cluster size on sample size is negligible. Most trials randomizing UK general practices and many other cluster randomized trials should account for variable cluster size in their sample size calculations.
Size variation in samples of fossil and recent murid teeth
Freudenthal, M.; Martín Suárez, E.
1990-01-01
The variability coefficient proposed by Freudenthal & Cuenca Bescós (1984) for samples of fossil cricetid teeth, is calculated for about 200 samples of fossil and recent murid teeth. The results are discussed, and compared with those obtained for the Cricetidae.
Size variation in samples of fossil and recent murid teeth
Freudenthal, M.; Martín Suárez, E.
1990-01-01
The variability coefficient proposed by Freudenthal & Cuenca Bescós (1984) for samples of fossil cricetid teeth, is calculated for about 200 samples of fossil and recent murid teeth. The results are discussed, and compared with those obtained for the Cricetidae.
Sample Size Determination for Regression Models Using Monte Carlo Methods in R
Beaujean, A. Alexander
2014-01-01
A common question asked by researchers using regression models is, What sample size is needed for my study? While there are formulae to estimate sample sizes, their assumptions are often not met in the collected data. A more realistic approach to sample size determination requires more information such as the model of interest, strength of the…
The choice of sample size for mortality forecasting : A Bayesian learning approach
Li, Hong; De Waegenaere, Anja; Melenberg, Bertrand
2015-01-01
Forecasted mortality rates using mortality models proposed in the recent literature are sensitive to the sample size. In this paper we propose a method based on Bayesian learning to determine model-specific posterior distributions of the sample sizes. In particular, the sample size is included as an
2006-01-01
An inlet stratification device (5) for a circuit circulating a fluid through a tank (1 ) and for providing and maintaining stratification of the fluid in the tank (1 ). The stratification de- vice (5) is arranged vertically in the tank (1) and comprises an inlet pipe (6) being at least partially...... formed of a flexible porous material and having an inlet (19) and outlets formed of the pores of the porous material. The stratification device (5) further comprises at least one outer pipe (7) surrounding the inlet pipe (6) in spaced relationship thereto and being at least partially formed of a porous...
Muntaner, C; Parsons, P E
1996-01-01
Most studies of inequalities and access to health care have used income as the sole indicator of social stratification. Despite the significance of social theory in health insurance research, there are no empirical studies comparing the ability of different models of social stratification to predict health insurance coverage. The aim of this study is to provide a comparative analysis using a variety of theory-driven indicators of social stratification and assess the relative strength of the association between these indicators and private health insurance. Data were collected in a 1993 telephone interview of a random digit dialing sample of the white population in the Baltimore Metropolitan Statistical Area. Indicators of social stratification included employment status, full-time work, education, occupation, industry, household income, firm size, and three types of assets: ownership, organizational, and skill/credential. The association between social stratification and private health insurance was strongest for those having higher household incomes, having attained at least a bachelor's degree, and working in a firm with more than 50 employees, followed by being an owner or manager, and by being employed. The addition of education and firm size improved the prediction of the household income model. The authors conclude that studies of inequalities in health insurance coverage can benefit from the inclusion of theory-driven indicators of social stratification such as human capital, labor market segmentation, and control over productive assets.
Bill, Anthony; Henderson, Sally; Penman, John
2010-01-01
Two test items that examined high school students' beliefs of sample size for large populations using the context of opinion polls conducted prior to national and state elections were developed. A trial of the two items with 21 male and 33 female Year 9 students examined their naive understanding of sample size: over half of students chose a…
Utility of Inferential Norming with Smaller Sample Sizes
Zhu, Jianjun; Chen, Hsin-Yi
2011-01-01
We examined the utility of inferential norming using small samples drawn from the larger "Wechsler Intelligence Scales for Children-Fourth Edition" (WISC-IV) standardization data set. The quality of the norms was estimated with multiple indexes such as polynomial curve fit, percentage of cases receiving the same score, average absolute…
Influence of macroinvertebrate sample size on bioassessment of streams
Vlek, H.E.; Sporka, F.; Krno, I.
2006-01-01
In order to standardise biological assessment of surface waters in Europe, a standardised method for sampling, sorting and identification of benthic macroinvertebrates in running waters was developed during the AQEM project. The AQEM method has proved to be relatively time-consuming. Hence, this stu
Modeling Multimodal Stratification
Boeriis, Morten
2017-01-01
This article discusses one of the core axioms of social semiotic theory, namely stratification, in the light of developments in multimodality in recent years. The discussion takes a point of departure in the approaches to stratification taken by Hjelmslev, Halliday, and Kress and van Leeuwen...
Multiscale sampling of plant diversity: Effects of minimum mapping unit size
Stohlgren, T.J.; Chong, G.W.; Kalkhan, M.A.; Schell, L.D.
1997-01-01
Only a small portion of any landscape can be sampled for vascular plant diversity because of constraints of cost (salaries, travel time between sites, etc.). Often, the investigator decides to reduce the cost of creating a vegetation map by increasing the minimum mapping unit (MMU), and/or by reducing the number of vegetation classes to be considered. Questions arise about what information is sacrificed when map resolution is decreased. We compared plant diversity patterns from vegetation maps made with 100-ha, 50-ha, 2-ha, and 0.02-ha MMUs in a 754-ha study area in Rocky Mountain National Park, Colorado, United States, using four 0.025-ha and 21 0.1-ha multiscale vegetation plots. We developed and tested species-log(area) curves, correcting the curves for within-vegetation type heterogeneity with Jaccard's coefficients. Total species richness in the study area was estimated from vegetation maps at each resolution (MMU), based on the corrected species-area curves, total area of the vegetation type, and species overlap among vegetation types. With the 0.02-ha MMU, six vegetation types were recovered, resulting in an estimated 552 species (95% CI = 520-583 species) in the 754-ha study area (330 plant species were observed in the 25 plots). With the 2-ha MMU, five vegetation types were recognized, resulting in an estimated 473 species for the study area. With the 50-ha MMU, 439 plant species were estimated for the four vegetation types recognized in the study area. With the 100-ha MMU, only three vegetation types were recognized, resulting in an estimated 341 plant species for the study area. Locally rare species and keystone ecosystems (areas of high or unique plant diversity) were missed at the 2-ha, 50-ha, and 100-ha scales. To evaluate the effects of minimum mapping unit size requires: (1) an initial stratification of homogeneous, heterogeneous, and rare habitat types; and (2) an evaluation of within-type and between-type heterogeneity generated by environmental
Sample size reduction in groundwater surveys via sparse data assimilation
Hussain, Z.
2013-04-01
In this paper, we focus on sparse signal recovery methods for data assimilation in groundwater models. The objective of this work is to exploit the commonly understood spatial sparsity in hydrodynamic models and thereby reduce the number of measurements to image a dynamic groundwater profile. To achieve this we employ a Bayesian compressive sensing framework that lets us adaptively select the next measurement to reduce the estimation error. An extension to the Bayesian compressive sensing framework is also proposed which incorporates the additional model information to estimate system states from even lesser measurements. Instead of using cumulative imaging-like measurements, such as those used in standard compressive sensing, we use sparse binary matrices. This choice of measurements can be interpreted as randomly sampling only a small subset of dug wells at each time step, instead of sampling the entire grid. Therefore, this framework offers groundwater surveyors a significant reduction in surveying effort without compromising the quality of the survey. © 2013 IEEE.
Finite-sample-size effects on convection in mushy layers
Zhong, Jin-Qiang; Wells, Andrew J; Wettlaufer, John S
2012-01-01
We report theoretical and experimental investigations of the flow instability responsible for the mushy-layer mode of convection and the formation of chimneys, drainage channels devoid of solid, during steady-state solidification of aqueous ammonium chloride. Under certain growth conditions a state of steady mushy-layer growth with no flow is unstable to the onset of convection, resulting in the formation of chimneys. We present regime diagrams to quantify the state of the flow as a function of the initial liquid concentration, the porous-medium Rayleigh number, and the sample width. For a given liquid concentration, increasing both the porous-medium Rayleigh number and the sample width caused the system to change from a stable state of no flow to a different state with the formation of chimneys. Decreasing the concentration ratio destabilized the system and promoted the formation of chimneys. As the initial liquid concentration increased, onset of convection and formation of chimneys occurred at larger value...
Calculating sample sizes for cluster randomized trials: we can keep it simple and efficient !
van Breukelen, Gerard J.P.; Candel, Math J.J.M.
2012-01-01
Objective: Simple guidelines for efficient sample sizes in cluster randomized trials with unknown intraclass correlation and varying cluster sizes. Methods: A simple equation is given for the optimal number of clusters and sample size per cluster. Here, optimal means maximizing power for a given
Progressive prediction method for failure data with small sample size
WANG Zhi-hua; FU Hui-min; LIU Cheng-rui
2011-01-01
The small sample prediction problem which commonly exists in reliability analysis was discussed with the progressive prediction method in this paper.The modeling and estimation procedure,as well as the forecast and confidence limits formula of the progressive auto regressive（PAR） method were discussed in great detail.PAR model not only inherits the simple linear features of auto regressive（AR） model,but also has applicability for nonlinear systems.An application was illustrated for predicting the future fatigue failure for Tantalum electrolytic capacitors.Forecasting results of PAR model were compared with auto regressive moving average（ARMA） model,and it can be seen that the PAR method can be considered good and shows a promise for future applications.
A Note on Sample Size and Solution Propriety for Confirmatory Factor Analytic Models
Jackson, Dennis L.; Voth, Jennifer; Frey, Marc P.
2013-01-01
Determining an appropriate sample size for use in latent variable modeling techniques has presented ongoing challenges to researchers. In particular, small sample sizes are known to present concerns over sampling error for the variances and covariances on which model estimation is based, as well as for fit indexes and convergence failures. The…
Evaluation of design flood estimates with respect to sample size
Kobierska, Florian; Engeland, Kolbjorn
2016-04-01
Estimation of design floods forms the basis for hazard management related to flood risk and is a legal obligation when building infrastructure such as dams, bridges and roads close to water bodies. Flood inundation maps used for land use planning are also produced based on design flood estimates. In Norway, the current guidelines for design flood estimates give recommendations on which data, probability distribution, and method to use dependent on length of the local record. If less than 30 years of local data is available, an index flood approach is recommended where the local observations are used for estimating the index flood and regional data are used for estimating the growth curve. For 30-50 years of data, a 2 parameter distribution is recommended, and for more than 50 years of data, a 3 parameter distribution should be used. Many countries have national guidelines for flood frequency estimation, and recommended distributions include the log Pearson II, generalized logistic and generalized extreme value distributions. For estimating distribution parameters, ordinary and linear moments, maximum likelihood and Bayesian methods are used. The aim of this study is to r-evaluate the guidelines for local flood frequency estimation. In particular, we wanted to answer the following questions: (i) Which distribution gives the best fit to the data? (ii) Which estimation method provides the best fit to the data? (iii) Does the answer to (i) and (ii) depend on local data availability? To answer these questions we set up a test bench for local flood frequency analysis using data based cross-validation methods. The criteria were based on indices describing stability and reliability of design flood estimates. Stability is used as a criterion since design flood estimates should not excessively depend on the data sample. The reliability indices describe to which degree design flood predictions can be trusted.
RNAseqPS: A Web Tool for Estimating Sample Size and Power for RNAseq Experiment
Yan Guo; Shilin Zhao; Chung-I Li; Quanhu Sheng; Yu Shyr
2014-01-01
Sample size and power determination is the first step in the experimental design of a successful study. Sample size and power calculation is required for applications for National Institutes of Health (NIH) funding. Sample size and power calculation is well established for traditional biological studies such as mouse model, genome wide association study (GWAS), and microarray studies. Recent developments in high-throughput sequencing technology have allowed RNAseq to replace microarray as the...
Judgement post-stratification for designed experiments.
Du, Juan; MacEachern, Steven N
2008-06-01
In many scientific studies, information that is not easily translated into covariates is ignored in the analysis. However, this type of information may significantly improve inference. In this research, we apply the idea of judgment post-stratification to utilize such information. Specifically, we consider experiments that are conducted under a completely randomized design. Sets of experimental units are formed, and the units in a set are ranked. Estimation is performed conditional on the sets and ranks. We propose a new estimator for a treatment contrast. We improve the new estimator by Rao-Blackwellization. Asymptotic distribution theory and corresponding inferential procedures for both estimators are developed. Simulation studies quantify the superiority of the new estimators and show their desirable properties for small and moderate sample sizes. The impact of the new techniques is illustrated with data from a clinical trial.
Sampling bee communities using pan traps: alternative methods increase sample size
Monitoring of the status of bee populations and inventories of bee faunas require systematic sampling. Efficiency and ease of implementation has encouraged the use of pan traps to sample bees. Efforts to find an optimal standardized sampling method for pan traps have focused on pan trap color. Th...
Distribution of the two-sample t-test statistic following blinded sample size re-estimation.
Lu, Kaifeng
2016-05-01
We consider the blinded sample size re-estimation based on the simple one-sample variance estimator at an interim analysis. We characterize the exact distribution of the standard two-sample t-test statistic at the final analysis. We describe a simulation algorithm for the evaluation of the probability of rejecting the null hypothesis at given treatment effect. We compare the blinded sample size re-estimation method with two unblinded methods with respect to the empirical type I error, the empirical power, and the empirical distribution of the standard deviation estimator and final sample size. We characterize the type I error inflation across the range of standardized non-inferiority margin for non-inferiority trials, and derive the adjusted significance level to ensure type I error control for given sample size of the internal pilot study. We show that the adjusted significance level increases as the sample size of the internal pilot study increases. Copyright © 2016 John Wiley & Sons, Ltd.
CT dose survey in adults: what sample size for what precision?
Taylor, Stephen [Hopital Ambroise Pare, Department of Radiology, Mons (Belgium); Muylem, Alain van [Hopital Erasme, Department of Pneumology, Brussels (Belgium); Howarth, Nigel [Clinique des Grangettes, Department of Radiology, Chene-Bougeries (Switzerland); Gevenois, Pierre Alain [Hopital Erasme, Department of Radiology, Brussels (Belgium); Tack, Denis [EpiCURA, Clinique Louis Caty, Department of Radiology, Baudour (Belgium)
2017-01-15
To determine variability of volume computed tomographic dose index (CTDIvol) and dose-length product (DLP) data, and propose a minimum sample size to achieve an expected precision. CTDIvol and DLP values of 19,875 consecutive CT acquisitions of abdomen (7268), thorax (3805), lumbar spine (3161), cervical spine (1515) and head (4106) were collected in two centers. Their variabilities were investigated according to sample size (10 to 1000 acquisitions) and patient body weight categories (no weight selection, 67-73 kg and 60-80 kg). The 95 % confidence interval in percentage of their median (CI95/med) value was calculated for increasing sample sizes. We deduced the sample size that set a 95 % CI lower than 10 % of the median (CI95/med ≤ 10 %). Sample size ensuring CI95/med ≤ 10 %, ranged from 15 to 900 depending on the body region and the dose descriptor considered. In sample sizes recommended by regulatory authorities (i.e., from 10-20 patients), mean CTDIvol and DLP of one sample ranged from 0.50 to 2.00 times its actual value extracted from 2000 samples. The sampling error in CTDIvol and DLP means is high in dose surveys based on small samples of patients. Sample size should be increased at least tenfold to decrease this variability. (orig.)
Lazary, Judit; Dome, Peter; Faludi, Gabor
2011-03-01
Increasing amount of genetic data on nicotine dependence (ND) is available in the literature, sometimes extremely large population size is reported but the study design is not always consequent. Phenotypic measures can vary from a simple 6-item self-rating scale to breath CO or serum cotinine level test but in genetic investigations this is not sophisticated; moreover the population stratification is also usually ignored. In contrast, possibly because of the strict traditions of pharmacological investigations, pharmacogenomic studies on smoking cessation therapy use more reliable phenotypic measures with high quality design consequently involving fewer participants. In spite of the heavy epidemiological data on smoking in Hungary, genetic background of heavy smoking is still not studied in this population. In this review we sum up the most important, replicated results but we also provide some critical remarks about the methodological shortcomings of these studies. Keeping in mind the value of large scale population ND association studies we would also like to emphasize that the clinical implementation of studies with larger samples but with weaker methodology and statistical analyses is limited. Similar to many other psychiatric disorders, ND is a multifactorial condition, therefore the measure of genetic effects requires a more complex study design.
Manju, Md Abu; Candel, Math J J M; Berger, Martijn P F
2014-07-10
In this paper, the optimal sample sizes at the cluster and person levels for each of two treatment arms are obtained for cluster randomized trials where the cost-effectiveness of treatments on a continuous scale is studied. The optimal sample sizes maximize the efficiency or power for a given budget or minimize the budget for a given efficiency or power. Optimal sample sizes require information on the intra-cluster correlations (ICCs) for effects and costs, the correlations between costs and effects at individual and cluster levels, the ratio of the variance of effects translated into costs to the variance of the costs (the variance ratio), sampling and measuring costs, and the budget. When planning, a study information on the model parameters usually is not available. To overcome this local optimality problem, the current paper also presents maximin sample sizes. The maximin sample sizes turn out to be rather robust against misspecifying the correlation between costs and effects at the cluster and individual levels but may lose much efficiency when misspecifying the variance ratio. The robustness of the maximin sample sizes against misspecifying the ICCs depends on the variance ratio. The maximin sample sizes are robust under misspecification of the ICC for costs for realistic values of the variance ratio greater than one but not robust under misspecification of the ICC for effects. Finally, we show how to calculate optimal or maximin sample sizes that yield sufficient power for a test on the cost-effectiveness of an intervention.
Variance Estimation, Design Effects, and Sample Size Calculations for Respondent-Driven Sampling
Salganik, Matthew J
2006-01-01
.... A recently developed statistical approach called respondent-driven sampling improves our ability to study hidden populations by allowing researchers to make unbiased estimates of the prevalence...
Implications of sampling design and sample size for national carbon accounting systems
Michael Köhl; Andrew Lister; Charles T. Scott; Thomas Baldauf; Daniel. Plugge
2011-01-01
Countries willing to adopt a REDD regime need to establish a national Measurement, Reporting and Verification (MRV) system that provides information on forest carbon stocks and carbon stock changes. Due to the extensive areas covered by forests the information is generally obtained by sample based surveys. Most operational sampling approaches utilize a combination of...
Implications of sampling design and sample size for national carbon accounting systems.
Köhl, Michael; Lister, Andrew; Scott, Charles T; Baldauf, Thomas; Plugge, Daniel
2011-11-08
Countries willing to adopt a REDD regime need to establish a national Measurement, Reporting and Verification (MRV) system that provides information on forest carbon stocks and carbon stock changes. Due to the extensive areas covered by forests the information is generally obtained by sample based surveys. Most operational sampling approaches utilize a combination of earth-observation data and in-situ field assessments as data sources. We compared the cost-efficiency of four different sampling design alternatives (simple random sampling, regression estimators, stratified sampling, 2-phase sampling with regression estimators) that have been proposed in the scope of REDD. Three of the design alternatives provide for a combination of in-situ and earth-observation data. Under different settings of remote sensing coverage, cost per field plot, cost of remote sensing imagery, correlation between attributes quantified in remote sensing and field data, as well as population variability and the percent standard error over total survey cost was calculated. The cost-efficiency of forest carbon stock assessments is driven by the sampling design chosen. Our results indicate that the cost of remote sensing imagery is decisive for the cost-efficiency of a sampling design. The variability of the sample population impairs cost-efficiency, but does not reverse the pattern of cost-efficiency of the individual design alternatives. Our results clearly indicate that it is important to consider cost-efficiency in the development of forest carbon stock assessments and the selection of remote sensing techniques. The development of MRV-systems for REDD need to be based on a sound optimization process that compares different data sources and sampling designs with respect to their cost-efficiency. This helps to reduce the uncertainties related with the quantification of carbon stocks and to increase the financial benefits from adopting a REDD regime.
1988-09-01
5 Sample The samples taken from each population will not be random samples . They will be nonprobability , purposive samples . More specifically, they...section will justify why statistical techniques based on the assumption of a random sample , will be used. First, this is the only possible method of...w lu 88 12 21 029 AFIT/GSM/LSM/88S-22 DEVELOPING CRITERIA FOR SAMPLE SIZES IN JET ENGINE ANALYTICAL COMPONENT INSPECTIONS AND THE ASSOCIATED
A Comparative Study of Power and Sample Size Calculations for Multivariate General Linear Models
Shieh, Gwowen
2003-01-01
Repeated measures and longitudinal studies arise often in social and behavioral science research. During the planning stage of such studies, the calculations of sample size are of particular interest to the investigators and should be an integral part of the research projects. In this article, we consider the power and sample size calculations for…
Optimal adaptive group sequential design with flexible timing of sample size determination.
Cui, Lu; Zhang, Lanju; Yang, Bo
2017-04-26
Flexible sample size designs, including group sequential and sample size re-estimation designs, have been used as alternatives to fixed sample size designs to achieve more robust statistical power and better trial efficiency. In this work, a new representation of sample size re-estimation design suggested by Cui et al. [5,6] is introduced as an adaptive group sequential design with flexible timing of sample size determination. This generalized adaptive group sequential design allows one time sample size determination either before the start of or in the mid-course of a clinical study. The new approach leads to possible design optimization on an expanded space of design parameters. Its equivalence to sample size re-estimation design proposed by Cui et al. provides further insight on re-estimation design and helps to address common confusions and misunderstanding. Issues in designing flexible sample size trial, including design objective, performance evaluation and implementation are touched upon with an example to illustrate. Copyright © 2017. Published by Elsevier Inc.
The attention-weighted sample-size model of visual short-term memory
Smith, Philip L.; Lilburn, Simon D.; Corbett, Elaine A.
2016-01-01
exceeded that predicted by the sample-size model for both simultaneously and sequentially presented stimuli. Instead, the set-size effect and the serial position curves with sequential presentation were predicted by an attention-weighted version of the sample-size model, which assumes that one of the items......We investigated the capacity of visual short-term memory (VSTM) in a phase discrimination task that required judgments about the configural relations between pairs of black and white features. Sewell et al. (2014) previously showed that VSTM capacity in an orientation discrimination task was well...... described by a sample-size model, which views VSTM as a resource comprised of a finite number of noisy stimulus samples. The model predicts the invariance of ∑i(di ′)2, the sum of squared sensitivities across items, for displays of different sizes. For phase discrimination, the set-size effect significantly...
Thermomagnetic behavior of magnetic susceptibility – heating rate and sample size effects
Diana eJordanova
2016-01-01
Full Text Available Thermomagnetic analysis of magnetic susceptibility k(T was carried out for a number of natural powder materials from soils, baked clay and anthropogenic dust samples using fast (11oC/min and slow (6.5oC/min heating rates available in the furnace of Kappabridge KLY2 (Agico. Based on the additional data for mineralogy, grain size and magnetic properties of the studied samples, behaviour of k(T cycles and the observed differences in the curves for fast and slow heating rate are interpreted in terms of mineralogical transformations and Curie temperatures (Tc. The effect of different sample size is also explored, using large volume and small volume of powder material. It is found that soil samples show enhanced information on mineralogical transformations and appearance of new strongly magnetic phases when using fast heating rate and large sample size. This approach moves the transformation at higher temperature, but enhances the amplitude of the signal of newly created phase. Large sample size gives prevalence of the local micro- environment, created by evolving gases, released during transformations. The example from archeological brick reveals the effect of different sample sizes on the observed Curie temperatures on heating and cooling curves, when the magnetic carrier is substituted magnetite (Mn0.2Fe2.70O4. Large sample size leads to bigger differences in Tcs on heating and cooling, while small sample size results in similar Tcs for both heating rates.
A C Bouman
Full Text Available Non-inferiority trials are performed when the main therapeutic effect of the new therapy is expected to be not unacceptably worse than that of the standard therapy, and the new therapy is expected to have advantages over the standard therapy in costs or other (health consequences. These advantages however are not included in the classic frequentist approach of sample size calculation for non-inferiority trials. In contrast, the decision theory approach of sample size calculation does include these factors. The objective of this study is to compare the conceptual and practical aspects of the frequentist approach and decision theory approach of sample size calculation for non-inferiority trials, thereby demonstrating that the decision theory approach is more appropriate for sample size calculation of non-inferiority trials.The frequentist approach and decision theory approach of sample size calculation for non-inferiority trials are compared and applied to a case of a non-inferiority trial on individually tailored duration of elastic compression stocking therapy compared to two years elastic compression stocking therapy for the prevention of post thrombotic syndrome after deep vein thrombosis.The two approaches differ substantially in conceptual background, analytical approach, and input requirements. The sample size calculated according to the frequentist approach yielded 788 patients, using a power of 80% and a one-sided significance level of 5%. The decision theory approach indicated that the optimal sample size was 500 patients, with a net value of €92 million.This study demonstrates and explains the differences between the classic frequentist approach and the decision theory approach of sample size calculation for non-inferiority trials. We argue that the decision theory approach of sample size estimation is most suitable for sample size calculation of non-inferiority trials.
Are sample sizes clear and justified in RCTs published in dental journals?
Despina Koletsi
Full Text Available Sample size calculations are advocated by the CONSORT group to justify sample sizes in randomized controlled trials (RCTs. The aim of this study was primarily to evaluate the reporting of sample size calculations, to establish the accuracy of these calculations in dental RCTs and to explore potential predictors associated with adequate reporting. Electronic searching was undertaken in eight leading specific and general dental journals. Replication of sample size calculations was undertaken where possible. Assumed variances or odds for control and intervention groups were also compared against those observed. The relationship between parameters including journal type, number of authors, trial design, involvement of methodologist, single-/multi-center study and region and year of publication, and the accuracy of sample size reporting was assessed using univariable and multivariable logistic regression. Of 413 RCTs identified, sufficient information to allow replication of sample size calculations was provided in only 121 studies (29.3%. Recalculations demonstrated an overall median overestimation of sample size of 15.2% after provisions for losses to follow-up. There was evidence that journal, methodologist involvement (OR = 1.97, CI: 1.10, 3.53, multi-center settings (OR = 1.86, CI: 1.01, 3.43 and time since publication (OR = 1.24, CI: 1.12, 1.38 were significant predictors of adequate description of sample size assumptions. Among journals JCP had the highest odds of adequately reporting sufficient data to permit sample size recalculation, followed by AJODO and JDR, with 61% (OR = 0.39, CI: 0.19, 0.80 and 66% (OR = 0.34, CI: 0.15, 0.75 lower odds, respectively. Both assumed variances and odds were found to underestimate the observed values. Presentation of sample size calculations in the dental literature is suboptimal; incorrect assumptions may have a bearing on the power of RCTs.
Are sample sizes clear and justified in RCTs published in dental journals?
Koletsi, Despina; Fleming, Padhraig S; Seehra, Jadbinder; Bagos, Pantelis G; Pandis, Nikolaos
2014-01-01
Sample size calculations are advocated by the CONSORT group to justify sample sizes in randomized controlled trials (RCTs). The aim of this study was primarily to evaluate the reporting of sample size calculations, to establish the accuracy of these calculations in dental RCTs and to explore potential predictors associated with adequate reporting. Electronic searching was undertaken in eight leading specific and general dental journals. Replication of sample size calculations was undertaken where possible. Assumed variances or odds for control and intervention groups were also compared against those observed. The relationship between parameters including journal type, number of authors, trial design, involvement of methodologist, single-/multi-center study and region and year of publication, and the accuracy of sample size reporting was assessed using univariable and multivariable logistic regression. Of 413 RCTs identified, sufficient information to allow replication of sample size calculations was provided in only 121 studies (29.3%). Recalculations demonstrated an overall median overestimation of sample size of 15.2% after provisions for losses to follow-up. There was evidence that journal, methodologist involvement (OR = 1.97, CI: 1.10, 3.53), multi-center settings (OR = 1.86, CI: 1.01, 3.43) and time since publication (OR = 1.24, CI: 1.12, 1.38) were significant predictors of adequate description of sample size assumptions. Among journals JCP had the highest odds of adequately reporting sufficient data to permit sample size recalculation, followed by AJODO and JDR, with 61% (OR = 0.39, CI: 0.19, 0.80) and 66% (OR = 0.34, CI: 0.15, 0.75) lower odds, respectively. Both assumed variances and odds were found to underestimate the observed values. Presentation of sample size calculations in the dental literature is suboptimal; incorrect assumptions may have a bearing on the power of RCTs.
Bouman, A C; ten Cate-Hoek, A J; Ramaekers, B L T; Joore, M A
2015-01-01
Non-inferiority trials are performed when the main therapeutic effect of the new therapy is expected to be not unacceptably worse than that of the standard therapy, and the new therapy is expected to have advantages over the standard therapy in costs or other (health) consequences. These advantages however are not included in the classic frequentist approach of sample size calculation for non-inferiority trials. In contrast, the decision theory approach of sample size calculation does include these factors. The objective of this study is to compare the conceptual and practical aspects of the frequentist approach and decision theory approach of sample size calculation for non-inferiority trials, thereby demonstrating that the decision theory approach is more appropriate for sample size calculation of non-inferiority trials. The frequentist approach and decision theory approach of sample size calculation for non-inferiority trials are compared and applied to a case of a non-inferiority trial on individually tailored duration of elastic compression stocking therapy compared to two years elastic compression stocking therapy for the prevention of post thrombotic syndrome after deep vein thrombosis. The two approaches differ substantially in conceptual background, analytical approach, and input requirements. The sample size calculated according to the frequentist approach yielded 788 patients, using a power of 80% and a one-sided significance level of 5%. The decision theory approach indicated that the optimal sample size was 500 patients, with a net value of €92 million. This study demonstrates and explains the differences between the classic frequentist approach and the decision theory approach of sample size calculation for non-inferiority trials. We argue that the decision theory approach of sample size estimation is most suitable for sample size calculation of non-inferiority trials.
Voss, Sebastian; Zimmermann, Beate; Zimmermann, Alexander
2016-09-01
In the last decades, an increasing number of studies analyzed spatial patterns in throughfall by means of variograms. The estimation of the variogram from sample data requires an appropriate sampling scheme: most importantly, a large sample and a layout of sampling locations that often has to serve both variogram estimation and geostatistical prediction. While some recommendations on these aspects exist, they focus on Gaussian data and high ratios of the variogram range to the extent of the study area. However, many hydrological data, and throughfall data in particular, do not follow a Gaussian distribution. In this study, we examined the effect of extent, sample size, sampling design, and calculation method on variogram estimation of throughfall data. For our investigation, we first generated non-Gaussian random fields based on throughfall data with large outliers. Subsequently, we sampled the fields with three extents (plots with edge lengths of 25 m, 50 m, and 100 m), four common sampling designs (two grid-based layouts, transect and random sampling) and five sample sizes (50, 100, 150, 200, 400). We then estimated the variogram parameters by method-of-moments (non-robust and robust estimators) and residual maximum likelihood. Our key findings are threefold. First, the choice of the extent has a substantial influence on the estimation of the variogram. A comparatively small ratio of the extent to the correlation length is beneficial for variogram estimation. Second, a combination of a minimum sample size of 150, a design that ensures the sampling of small distances and variogram estimation by residual maximum likelihood offers a good compromise between accuracy and efficiency. Third, studies relying on method-of-moments based variogram estimation may have to employ at least 200 sampling points for reliable variogram estimates. These suggested sample sizes exceed the number recommended by studies dealing with Gaussian data by up to 100 %. Given that most previous
Optimal and maximin sample sizes for multicentre cost-effectiveness trials.
Manju, Md Abu; Candel, Math J J M; Berger, Martijn P F
2015-10-01
This paper deals with the optimal sample sizes for a multicentre trial in which the cost-effectiveness of two treatments in terms of net monetary benefit is studied. A bivariate random-effects model, with the treatment-by-centre interaction effect being random and the main effect of centres fixed or random, is assumed to describe both costs and effects. The optimal sample sizes concern the number of centres and the number of individuals per centre in each of the treatment conditions. These numbers maximize the efficiency or power for given research costs or minimize the research costs at a desired level of efficiency or power. Information on model parameters and sampling costs are required to calculate these optimal sample sizes. In case of limited information on relevant model parameters, sample size formulas are derived for so-called maximin sample sizes which guarantee a power level at the lowest study costs. Four different maximin sample sizes are derived based on the signs of the lower bounds of two model parameters, with one case being worst compared to others. We numerically evaluate the efficiency of the worst case instead of using others. Finally, an expression is derived for calculating optimal and maximin sample sizes that yield sufficient power to test the cost-effectiveness of two treatments. © The Author(s) 2015.
Ismet DOGAN
2015-10-01
Full Text Available Objective: Choosing the most efficient statistical test is one of the essential problems of statistics. Asymptotic relative efficiency is a notion which enables to implement in large samples the quantitative comparison of two different tests used for testing of the same statistical hypothesis. The notion of the asymptotic efficiency of tests is more complicated than that of asymptotic efficiency of estimates. This paper discusses the effect of sample size on expected values and variances of non-parametric tests for independent two samples and determines the most effective test for different sample sizes using Fraser efficiency value. Material and Methods: Since calculating the power value in comparison of the tests is not practical most of the time, using the asymptotic relative efficiency value is favorable. Asymptotic relative efficiency is an indispensable technique for comparing and ordering statistical test in large samples. It is especially useful in nonparametric statistics where there exist numerous heuristic tests such as the linear rank tests. In this study, the sample size is determined as 2 ≤ n ≤ 50. Results: In both balanced and unbalanced cases, it is found that, as the sample size increases expected values and variances of all the tests discussed in this paper increase as well. Additionally, considering the Fraser efficiency, Mann-Whitney U test is found as the most efficient test among the non-parametric tests that are used in comparison of independent two samples regardless of their sizes. Conclusion: According to Fraser efficiency, Mann-Whitney U test is found as the most efficient test.
Blinded sample size reestimation in non-inferiority trials with binary endpoints.
Friede, Tim; Mitchell, Charles; Müller-Velten, Günther
2007-12-01
Sample size calculations in the planning of clinical trials depend on good estimates of the model parameters involved. When the estimates of these parameters have a high degree of uncertainty attached to them, it is advantageous to reestimate the sample size after an internal pilot study. For non-inferiority trials with binary outcome we compare the performance of Type I error rate and power between fixed-size designs and designs with sample size reestimation. The latter design shows itself to be effective in correcting sample size and power of the tests when misspecification of nuisance parameters occurs with the former design. (c) 2007 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim
The PowerAtlas: a power and sample size atlas for microarray experimental design and research
Wang Jelai
2006-02-01
Full Text Available Abstract Background Microarrays permit biologists to simultaneously measure the mRNA abundance of thousands of genes. An important issue facing investigators planning microarray experiments is how to estimate the sample size required for good statistical power. What is the projected sample size or number of replicate chips needed to address the multiple hypotheses with acceptable accuracy? Statistical methods exist for calculating power based upon a single hypothesis, using estimates of the variability in data from pilot studies. There is, however, a need for methods to estimate power and/or required sample sizes in situations where multiple hypotheses are being tested, such as in microarray experiments. In addition, investigators frequently do not have pilot data to estimate the sample sizes required for microarray studies. Results To address this challenge, we have developed a Microrarray PowerAtlas 1. The atlas enables estimation of statistical power by allowing investigators to appropriately plan studies by building upon previous studies that have similar experimental characteristics. Currently, there are sample sizes and power estimates based on 632 experiments from Gene Expression Omnibus (GEO. The PowerAtlas also permits investigators to upload their own pilot data and derive power and sample size estimates from these data. This resource will be updated regularly with new datasets from GEO and other databases such as The Nottingham Arabidopsis Stock Center (NASC. Conclusion This resource provides a valuable tool for investigators who are planning efficient microarray studies and estimating required sample sizes.
Wei Lin Teoh
Full Text Available Designs of the double sampling (DS X chart are traditionally based on the average run length (ARL criterion. However, the shape of the run length distribution changes with the process mean shifts, ranging from highly skewed when the process is in-control to almost symmetric when the mean shift is large. Therefore, we show that the ARL is a complicated performance measure and that the median run length (MRL is a more meaningful measure to depend on. This is because the MRL provides an intuitive and a fair representation of the central tendency, especially for the rightly skewed run length distribution. Since the DS X chart can effectively reduce the sample size without reducing the statistical efficiency, this paper proposes two optimal designs of the MRL-based DS X chart, for minimizing (i the in-control average sample size (ASS and (ii both the in-control and out-of-control ASSs. Comparisons with the optimal MRL-based EWMA X and Shewhart X charts demonstrate the superiority of the proposed optimal MRL-based DS X chart, as the latter requires a smaller sample size on the average while maintaining the same detection speed as the two former charts. An example involving the added potassium sorbate in a yoghurt manufacturing process is used to illustrate the effectiveness of the proposed MRL-based DS X chart in reducing the sample size needed.
Vallejo, Adriana; Muniesa, Ana; Ferreira, Chelo; de Blas, Ignacio
2013-10-01
Nowadays the formula to calculate the sample size for estimate a proportion (as prevalence) is based on the Normal distribution, however it would be based on a Binomial distribution which confidence interval was possible to be calculated using the Wilson Score method. By comparing the two formulae (Normal and Binomial distributions), the variation of the amplitude of the confidence intervals is relevant in the tails and the center of the curves. In order to calculate the needed sample size we have simulated an iterative sampling procedure, which shows an underestimation of the sample size for values of prevalence closed to 0 or 1, and also an overestimation for values closed to 0.5. Attending to these results we proposed an algorithm based on Wilson Score method that provides similar values for the sample size than empirically obtained by simulation.
Light propagation in tissues: effect of finite size of tissue sample
Melnik, Ivan S.; Dets, Sergiy M.; Rusina, Tatyana V.
1995-12-01
Laser beam propagation inside tissues with different lateral dimensions has been considered. Scattering and anisotropic properties of tissue critically determine spatial fluence distribution and predict sizes of tissue specimens when deviations of this distribution can be neglected. Along the axis of incident beam the fluence rate weakly depends on sample size whereas its relative increase (more than 20%) towards the lateral boundaries. The finite sizes were considered to be substantial only for samples with sizes comparable with the diameter of the laser beam. Interstitial irradiance patterns simulated by Monte Carlo method were compared with direct measurements in human brain specimens.
Sample size adjustment designs with time-to-event outcomes: A caution.
Freidlin, Boris; Korn, Edward L
2017-08-01
Sample size adjustment designs, which allow increasing the study sample size based on interim analysis of outcome data from a randomized clinical trial, have been increasingly promoted in the biostatistical literature. Although it is recognized that group sequential designs can be at least as efficient as sample size adjustment designs, many authors argue that a key advantage of these designs is their flexibility; interim sample size adjustment decisions can incorporate information and business interests external to the trial. Recently, Chen et al. (Clinical Trials 2015) considered sample size adjustment applications in the time-to-event setting using a design (CDL) that limits adjustments to situations where the interim results are promising. The authors demonstrated that while CDL provides little gain in unconditional power (versus fixed-sample-size designs), there is a considerable increase in conditional power for trials in which the sample size is adjusted. In time-to-event settings, sample size adjustment allows an increase in the number of events required for the final analysis. This can be achieved by either (a) following the original study population until the additional events are observed thus focusing on the tail of the survival curves or (b) enrolling a potentially large number of additional patients thus focusing on the early differences in survival curves. We use the CDL approach to investigate performance of sample size adjustment designs in time-to-event trials. Through simulations, we demonstrate that when the magnitude of the true treatment effect changes over time, interim information on the shape of the survival curves can be used to enrich the final analysis with events from the time period with the strongest treatment effect. In particular, interested parties have the ability to make the end-of-trial treatment effect larger (on average) based on decisions using interim outcome data. Furthermore, in "clinical null" cases where there is no
A NONPARAMETRIC PROCEDURE OF THE SAMPLE SIZE DETERMINATION FOR SURVIVAL RATE TEST
无
2000-01-01
Objective This paper proposes a nonparametric procedure of the sample size determination for survival rate test. Methods Using the classical asymptotic normal procedure yields the required homogenetic effective sample size and using the inverse operation with the prespecified value of the survival function of censoring times yields the required sample size. Results It is matched with the rate test for censored data, does not involve survival distributions, and reduces to its classical counterpart when there is no censoring. The observed power of the test coincides with the prescribed power under usual clinical conditions. Conclusion It can be used for planning survival studies of chronic diseases.
Bice, K.; Clement, S. C.
1981-01-01
X-ray diffraction and spectroscopy were used to investigate the mineralogical and chemical properties of the Calvert, Ball Old Mine, Ball Martin, and Jordan Sediments. The particle size distribution and index of refraction of each sample were determined. The samples are composed primarily of quartz, kaolinite, and illite. The clay minerals are most abundant in the finer particle size fractions. The chemical properties of the four samples are similar. The Calvert sample is most notably different in that it contains a relatively high amount of iron. The dominant particle size fraction in each sample is silt, with lesser amounts of clay and sand. The indices of refraction of the sediments are the same with the exception of the Calvert sample which has a slightly higher value.
Honório Kanegae Junior
2006-06-01
Full Text Available The stands stratification for successive forest inventory is usually based on stands cadastral information, such as theage, the species, the spacing, and the management regime, among others. The size of the sample is usually conditioned by thevariability of the forest and by the required precision. Thus, the control of the variation through the efficient stratification has stronginfluence on sample precision and size. This study evaluated: the stratification propitiated by two spatial interpolators, the statisticianone represented by the krigage and the deterministic one represented by the inverse of the square of the distance; evaluated theinterpolators in relation to simple random sampling and the traditional stratification based on cadastral data, in the reduction of thevariance of the average and sampling error; and defined the optimal number of strata when spatial interpolators are used. For thegeneration of the strata, it was studied 4 different dendrometric variables: volume, basal area, dominant height and site index in 2different ages: 2.5 years and 3.5 years. It was concluded that the krigage of the volume per hectare obtained at 3.5 years of age reducedin 47% the stand average variance and in 32% the inventory sampling error, when compared to the simple random sampling. Thevolume interpolator IDW, at 3.5 years of age, reduced in 74% the stand average variance and in 48% the inventory sampling error.The less efficient stratificator was the one based on age, species and spacing. In spite of the IDW method having presented highefficiency, it doesn t guarantee that the efficiency be maintained, if a new sampling is accomplished in the same projects, contrarily tothe geostatistic krigage. In forest stands that don t present spatial dependence, the IDW method can be used with great efficiency in thetraditional stratification. The less efficient stratification method is the one based on the control of age, species and spacing (STR
Frictional behaviour of sandstone: A sample-size dependent triaxial investigation
Roshan, Hamid; Masoumi, Hossein; Regenauer-Lieb, Klaus
2017-01-01
Frictional behaviour of rocks from the initial stage of loading to final shear displacement along the formed shear plane has been widely investigated in the past. However the effect of sample size on such frictional behaviour has not attracted much attention. This is mainly related to the limitations in rock testing facilities as well as the complex mechanisms involved in sample-size dependent frictional behaviour of rocks. In this study, a suite of advanced triaxial experiments was performed on Gosford sandstone samples at different sizes and confining pressures. The post-peak response of the rock along the formed shear plane has been captured for the analysis with particular interest in sample-size dependency. Several important phenomena have been observed from the results of this study: a) the rate of transition from brittleness to ductility in rock is sample-size dependent where the relatively smaller samples showed faster transition toward ductility at any confining pressure; b) the sample size influences the angle of formed shear band and c) the friction coefficient of the formed shear plane is sample-size dependent where the relatively smaller sample exhibits lower friction coefficient compared to larger samples. We interpret our results in terms of a thermodynamics approach in which the frictional properties for finite deformation are viewed as encompassing a multitude of ephemeral slipping surfaces prior to the formation of the through going fracture. The final fracture itself is seen as a result of the self-organisation of a sufficiently large ensemble of micro-slip surfaces and therefore consistent in terms of the theory of thermodynamics. This assumption vindicates the use of classical rock mechanics experiments to constrain failure of pressure sensitive rocks and the future imaging of these micro-slips opens an exciting path for research in rock failure mechanisms.
McCain, J.D.; Dawes, S.S.; Farthing, W.E.
1986-05-01
The report is Attachment No. 2 to the Final Report of ARB Contract A3-092-32 and provides a tutorial on the use of Cascade (Series) Cyclones to obtain size-fractionated particulate samples from industrial flue gases at stationary sources. The instrumentation and procedures described are designed to protect the purity of the collected samples so that post-test chemical analysis may be performed for organic and inorganic compounds, including instrumental analysis for trace elements. The instrumentation described collects bulk quantities for each of six size fractions over the range 10 to 0.4 micrometer diameter. The report describes the operating principles, calibration, and empirical modeling of small cyclone performance. It also discusses the preliminary calculations, operation, sample retrieval, and data analysis associated with the use of cyclones to obtain size-segregated samples and to measure particle-size distributions.
Sample size determination for logistic regression on a logit-normal distribution.
Kim, Seongho; Heath, Elisabeth; Heilbrun, Lance
2017-06-01
Although the sample size for simple logistic regression can be readily determined using currently available methods, the sample size calculation for multiple logistic regression requires some additional information, such as the coefficient of determination ([Formula: see text]) of a covariate of interest with other covariates, which is often unavailable in practice. The response variable of logistic regression follows a logit-normal distribution which can be generated from a logistic transformation of a normal distribution. Using this property of logistic regression, we propose new methods of determining the sample size for simple and multiple logistic regressions using a normal transformation of outcome measures. Simulation studies and a motivating example show several advantages of the proposed methods over the existing methods: (i) no need for [Formula: see text] for multiple logistic regression, (ii) available interim or group-sequential designs, and (iii) much smaller required sample size.
Monotonicity in the Sample Size of the Length of Classical Confidence Intervals
Kagan, Abram M
2012-01-01
It is proved that the average length of standard confidence intervals for parameters of gamma and normal distributions monotonically decrease with the sample size. The proofs are based on fine properties of the classical gamma function.
Moore, R. P.; Shah, B. V.
The average design effects for statistics estimated from the base-year National Longitudinal Study data are presented. Attempts to partition the effects into those due to stratification, clustering, and unequal weighting are discussed. The expected increases in subpopulation sample sizes due to oversampling are calculated and compared with the…
RNAseqPS: A Web Tool for Estimating Sample Size and Power for RNAseq Experiment.
Guo, Yan; Zhao, Shilin; Li, Chung-I; Sheng, Quanhu; Shyr, Yu
2014-01-01
Sample size and power determination is the first step in the experimental design of a successful study. Sample size and power calculation is required for applications for National Institutes of Health (NIH) funding. Sample size and power calculation is well established for traditional biological studies such as mouse model, genome wide association study (GWAS), and microarray studies. Recent developments in high-throughput sequencing technology have allowed RNAseq to replace microarray as the technology of choice for high-throughput gene expression profiling. However, the sample size and power analysis of RNAseq technology is an underdeveloped area. Here, we present RNAseqPS, an advanced online RNAseq power and sample size calculation tool based on the Poisson and negative binomial distributions. RNAseqPS was built using the Shiny package in R. It provides an interactive graphical user interface that allows the users to easily conduct sample size and power analysis for RNAseq experimental design. RNAseqPS can be accessed directly at http://cqs.mc.vanderbilt.edu/shiny/RNAseqPS/.
Sample size choices for XRCT scanning of highly unsaturated soil mixtures
Smith Jonathan C.
2016-01-01
Full Text Available Highly unsaturated soil mixtures (clay, sand and gravel are used as building materials in many parts of the world, and there is increasing interest in understanding their mechanical and hydraulic behaviour. In the laboratory, x-ray computed tomography (XRCT is becoming more widely used to investigate the microstructures of soils, however a crucial issue for such investigations is the choice of sample size, especially concerning the scanning of soil mixtures where there will be a range of particle and void sizes. In this paper we present a discussion (centred around a new set of XRCT scans on sample sizing for scanning of samples comprising soil mixtures, where a balance has to be made between realistic representation of the soil components and the desire for high resolution scanning, We also comment on the appropriateness of differing sample sizes in comparison to sample sizes used for other geotechnical testing. Void size distributions for the samples are presented and from these some hypotheses are made as to the roles of inter- and intra-aggregate voids in the mechanical behaviour of highly unsaturated soils.
Guo, Jiin-Huarng; Chen, Hubert J; Luh, Wei-Ming
2011-11-01
The allocation of sufficient participants into different experimental groups for various research purposes under given constraints is an important practical problem faced by researchers. We address the problem of sample size determination between two independent groups for unequal and/or unknown variances when both the power and the differential cost are taken into consideration. We apply the well-known Welch approximate test to derive various sample size allocation ratios by minimizing the total cost or, equivalently, maximizing the statistical power. Two types of hypotheses including superiority/non-inferiority and equivalence of two means are each considered in the process of sample size planning. A simulation study is carried out and the proposed method is validated in terms of Type I error rate and statistical power. As a result, the simulation study reveals that the proposed sample size formulas are very satisfactory under various variances and sample size allocation ratios. Finally, a flowchart, tables, and figures of several sample size allocations are presented for practical reference.
A margin based approach to determining sample sizes via tolerance bounds.
Newcomer, Justin T.; Freeland, Katherine Elizabeth
2013-09-01
This paper proposes a tolerance bound approach for determining sample sizes. With this new methodology we begin to think of sample size in the context of uncertainty exceeding margin. As the sample size decreases the uncertainty in the estimate of margin increases. This can be problematic when the margin is small and only a few units are available for testing. In this case there may be a true underlying positive margin to requirements but the uncertainty may be too large to conclude we have sufficient margin to those requirements with a high level of statistical confidence. Therefore, we provide a methodology for choosing a sample size large enough such that an estimated QMU uncertainty based on the tolerance bound approach will be smaller than the estimated margin (assuming there is positive margin). This ensures that the estimated tolerance bound will be within performance requirements and the tolerance ratio will be greater than one, supporting a conclusion that we have sufficient margin to the performance requirements. In addition, this paper explores the relationship between margin, uncertainty, and sample size and provides an approach and recommendations for quantifying risk when sample sizes are limited.
Zhao, Jinsong; Chen, Boyu
2016-03-01
Species sensitivity distribution (SSD) is a widely used model that extrapolates the ecological risk to ecosystem levels from the ecotoxicity of a chemical to individual organisms. However, model choice and sample size significantly affect the development of the SSD model and the estimation of hazardous concentrations at the 5th centile (HC5). To interpret their effects, the SSD model for chlorpyrifos, a widely used organophosphate pesticide, to aquatic organisms is presented with emphases on model choice and sample size. Three subsets of median effective concentration (EC50) with different sample sizes were obtained from ECOTOX and used to build SSD models based on parametric distribution (normal, logistic, and triangle distribution) and nonparametric bootstrap. The SSD models based on the triangle distribution are superior to the normal and logistic distributions according to several goodness-of-fit techniques. Among all parametric SSD models, the one with the largest sample size based on the triangle distribution gives the most strict HC5 with 0.141μmolL(-1). The HC5 derived from the nonparametric bootstrap is 0.159μmol L(-1). The minimum sample size required to build a stable SSD model is 11 based on parametric distribution and 23 based on nonparametric bootstrap. The study suggests that model choice and sample size are important sources of uncertainty for application of the SSD model. Copyright © 2015 Elsevier Inc. All rights reserved.
Shrinkage anisotropy characteristics from soil structure and initial sample/layer size
Chertkov, V Y
2014-01-01
The objective of this work is a physical prediction of such soil shrinkage anisotropy characteristics as variation with drying of (i) different sample/layer sizes and (ii) the shrinkage geometry factor. With that, a new presentation of the shrinkage anisotropy concept is suggested through the sample/layer size ratios. The work objective is reached in two steps. First, the relations are derived between the indicated soil shrinkage anisotropy characteristics and three different shrinkage curves of a soil relating to: small samples (without cracking at shrinkage), sufficiently large samples (with internal cracking), and layers of similar thickness. Then, the results of a recent work with respect to the physical prediction of the three shrinkage curves are used. These results connect the shrinkage curves with the initial sample size/layer thickness as well as characteristics of soil texture and structure (both inter- and intra-aggregate) as physical parameters. The parameters determining the reference shrinkage c...
A normative inference approach for optimal sample sizes in decisions from experience.
Ostwald, Dirk; Starke, Ludger; Hertwig, Ralph
2015-01-01
"Decisions from experience" (DFE) refers to a body of work that emerged in research on behavioral decision making over the last decade. One of the major experimental paradigms employed to study experience-based choice is the "sampling paradigm," which serves as a model of decision making under limited knowledge about the statistical structure of the world. In this paradigm respondents are presented with two payoff distributions, which, in contrast to standard approaches in behavioral economics, are specified not in terms of explicit outcome-probability information, but by the opportunity to sample outcomes from each distribution without economic consequences. Participants are encouraged to explore the distributions until they feel confident enough to decide from which they would prefer to draw from in a final trial involving real monetary payoffs. One commonly employed measure to characterize the behavior of participants in the sampling paradigm is the sample size, that is, the number of outcome draws which participants choose to obtain from each distribution prior to terminating sampling. A natural question that arises in this context concerns the "optimal" sample size, which could be used as a normative benchmark to evaluate human sampling behavior in DFE. In this theoretical study, we relate the DFE sampling paradigm to the classical statistical decision theoretic literature and, under a probabilistic inference assumption, evaluate optimal sample sizes for DFE. In our treatment we go beyond analytically established results by showing how the classical statistical decision theoretic framework can be used to derive optimal sample sizes under arbitrary, but numerically evaluable, constraints. Finally, we critically evaluate the value of deriving optimal sample sizes under this framework as testable predictions for the experimental study of sampling behavior in DFE.
A normative inference approach for optimal sample sizes in decisions from experience
Dirk eOstwald
2015-09-01
Full Text Available Decisions from experience (DFE refers to a body of work that emerged in research on behavioral decision making over the last decade. One of the major experimental paradigms employed to study experienced-based choice is the sampling paradigm, which serves as a model of decision making under limited knowledge about the statistical structure of the world. In this paradigm respondents are presented with two payoff distributions, which, in contrast to standard approaches in behavioral economics, are specified not in terms of explicit outcome-probability information, but by the opportunity to sample outcomes from each distribution without economic consequences. Participants are encouraged to explore the distributions until they feel confident enough to decide from which they would prefer to draw from in a final trial involving real monetary payoffs. One commonly employed measure to characterize the behavior of participants in the sampling paradigm is the sample size, that is, the number of outcome draws which participants choose to obtain from each distribution prior to terminating sampling. A natural question that arises in this context concerns the optimal sample size, which could be used as a normative benchmark to evaluate human sampling behavior in DFE. In this theoretical manuscript, we relate the DFE sampling paradigm to the classical statistical decision theoretic literature and, under a probabilistic inference assumption, evaluate optimal sample sizes for decisions from experience. In our treatment we go beyond analytically established results by showing how the classical statistical decision theoretic framework can be used to derive optimal sample sizes under arbitrary, but numerically evaluable, constraints. Finally, we critically evaluate the value of deriving optimal sample sizes under this framework as testable predictions for the experimental study of sampling behavior in DFE.
Comparison of population-based association study methods correcting for population stratification.
Feng Zhang
Full Text Available Population stratification can cause spurious associations in population-based association studies. Several statistical methods have been proposed to reduce the impact of population stratification on population-based association studies. We simulated a set of stratified populations based on the real haplotype data from the HapMap ENCODE project, and compared the relative power, type I error rates, accuracy and positive prediction value of four prevailing population-based association study methods: traditional case-control tests, structured association (SA, genomic control (GC and principal components analysis (PCA under various population stratification levels. Additionally, we evaluated the effects of sample sizes and frequencies of disease susceptible allele on the performance of the four analytical methods in the presence of population stratification. We found that the performance of PCA was very stable under various scenarios. Our comparison results suggest that SA and PCA have comparable performance, if sufficient ancestral informative markers are used in SA analysis. GC appeared to be strongly conservative in significantly stratified populations. It may be better to apply GC in the stratified populations with low stratification level. Our study intends to provide a practical guideline for researchers to select proper study methods and make appropriate inference of the results in population-based association studies.
Granström, Sara; Pipper, Christian Bressen; Møgelvang, Rasmus; Sogaard, Peter; Willesen, Jakob Lundgren; Koch, Jørgen
2012-12-01
The aims of this study were to compare the effect of sample volume (SV) size settings and sampling method on measurement variability and peak systolic (s'), and early (e') and late (a') diastolic longitudinal myocardial velocities using color tissue Doppler imaging (cTDI) in cats. Twenty cats with normal echocardiograms and 20 cats with hypertrophic cardiomyopathy. We quantified and compared empirical variance and average absolute values of s', e' and a' for three cardiac cycles using eight different SV settings (length 1,2,3 and 5 mm; width 1 and 2 mm) and three methods of sampling (end-diastolic sampling with manual tracking of the SV, end-systolic sampling without tracking, and random-frame sampling without tracking). No significant difference in empirical variance could be demonstrated between most of the tested SVs. However, the two settings with a length of 1 mm resulted in a significantly higher variance compared with all settings where the SV length exceeded 2 mm (p sampling method on the variability of measurements (p = 0.003) and manual tracking obtained the lowest variance. No difference in average values of s', e' or a' could be found between any of the SV settings or sampling methods. Within the tested range of SV settings, an SV length of 1 mm resulted in higher measurement variability compared with an SV length of 3 and 5 mm, and should therefore be avoided. Manual tracking of the sample volume is recommended. Copyright © 2012 Elsevier B.V. All rights reserved.
Blinded sample size re-estimation in three-arm trials with 'gold standard' design.
Mütze, Tobias; Friede, Tim
2017-10-15
In this article, we study blinded sample size re-estimation in the 'gold standard' design with internal pilot study for normally distributed outcomes. The 'gold standard' design is a three-arm clinical trial design that includes an active and a placebo control in addition to an experimental treatment. We focus on the absolute margin approach to hypothesis testing in three-arm trials at which the non-inferiority of the experimental treatment and the assay sensitivity are assessed by pairwise comparisons. We compare several blinded sample size re-estimation procedures in a simulation study assessing operating characteristics including power and type I error. We find that sample size re-estimation based on the popular one-sample variance estimator results in overpowered trials. Moreover, sample size re-estimation based on unbiased variance estimators such as the Xing-Ganju variance estimator results in underpowered trials, as it is expected because an overestimation of the variance and thus the sample size is in general required for the re-estimation procedure to eventually meet the target power. To overcome this problem, we propose an inflation factor for the sample size re-estimation with the Xing-Ganju variance estimator and show that this approach results in adequately powered trials. Because of favorable features of the Xing-Ganju variance estimator such as unbiasedness and a distribution independent of the group means, the inflation factor does not depend on the nuisance parameter and, therefore, can be calculated prior to a trial. Moreover, we prove that the sample size re-estimation based on the Xing-Ganju variance estimator does not bias the effect estimate. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.
SMALL SAMPLE SIZE IN 2X2 CROSS OVER DESIGNS: CONDITIONS OF DETERMINATION
B SOLEYMANI
2001-09-01
Full Text Available Introduction. Determination of small sample size in some clinical trials is a matter of importance. In cross-over studies which are one types of clinical trials, the matter is more significant. In this article, the conditions in which determination of small sample size in cross-over studies are possible were considered, and the effect of deviation from normality on the matter has been shown. Methods. The present study has been done on such 2x2 cross-over studies that variable of interest is quantitative one and is measurable by ratio or interval scale. The method of consideration is based on use of variable and sample mean"s distributions, central limit theorem, method of sample size determination in two groups, and cumulant or moment generating function. Results. In normal variables or transferable to normal variables, there is no restricting factors other than significant level and power of the test for determination of sample size, but in the case of non-normal variables, it should be determined such large that guarantee the normality of sample mean"s distribution. Discussion. In such cross over studies that because of existence of theoretical base, few samples can be computed, one should not do it without taking applied worth of results into consideration. While determining sample size, in addition to variance, it is necessary to consider distribution of variable, particularly through its skewness and kurtosis coefficients. the more deviation from normality, the more need of samples. Since in medical studies most of the continuous variables are closed to normal distribution, a few number of samples often seems to be adequate for convergence of sample mean to normal distribution.
Age differences in body size stereotyping in a sample of preschool girls.
Harriger, Jennifer A
2015-01-01
Researchers have demonstrated that societal concerns about dieting and body size have led to an increase in negative attitudes toward obese people and that girls as young as 3 years old endorse similar body size stereotypes as have been previously found with adults. Few studies, however, have examined age differences in their participants. A sample of 102 girls (3-5-years-old) completed measures of body size stereotyping. Results indicate that while body-size stereotyping is present by age 3, pro-thin beliefs may develop prior to anti-fat beliefs. Implications and future directions for research with preschool children are discussed.
Norm Block Sample Sizes: A Review of 17 Individually Administered Intelligence Tests
Norfolk, Philip A.; Farmer, Ryan L.; Floyd, Randy G.; Woods, Isaac L.; Hawkins, Haley K.; Irby, Sarah M.
2015-01-01
The representativeness, recency, and size of norm samples strongly influence the accuracy of inferences drawn from their scores. Inadequate norm samples may lead to inflated or deflated scores for individuals and poorer prediction of developmental and academic outcomes. The purpose of this study was to apply Kranzler and Floyd's method for…
Page sample size in web accessibility testing: how many pages is enough?
Velleman, Eric; Geest, van der Thea
2013-01-01
Various countries and organizations use a different sampling approach and sample size of web pages in accessibility conformance tests. We are conducting a systematic analysis to determine how many pages is enough for testing whether a website is compliant with standard accessibility guidelines. This
Finch, W. Holmes; Finch, Maria E. Hernandez
2016-01-01
Researchers and data analysts are sometimes faced with the problem of very small samples, where the number of variables approaches or exceeds the overall sample size; i.e. high dimensional data. In such cases, standard statistical models such as regression or analysis of variance cannot be used, either because the resulting parameter estimates…
Stefanović Milena
2013-01-01
Full Text Available In studies of population variability, particular attention has to be paid to the selection of a representative sample. The aim of this study was to assess the size of the new representative sample on the basis of the variability of chemical content of the initial sample on the example of a whitebark pine population. Statistical analysis included the content of 19 characteristics (terpene hydrocarbons and their derivates of the initial sample of 10 elements (trees. It was determined that the new sample should contain 20 trees so that the mean value calculated from it represents a basic set with a probability higher than 95 %. Determination of the lower limit of the representative sample size that guarantees a satisfactory reliability of generalization proved to be very important in order to achieve cost efficiency of the research. [Projekat Ministarstva nauke Republike Srbije, br. OI-173011, br. TR-37002 i br. III-43007
Ifoulis, A A; Savopoulou-Soultani, M
2006-10-01
The purpose of this research was to quantify the spatial pattern and develop a sampling program for larvae of Lobesia botrana Denis and Schiffermüller (Lepidoptera: Tortricidae), an important vineyard pest in northern Greece. Taylor's power law and Iwao's patchiness regression were used to model the relationship between the mean and the variance of larval counts. Analysis of covariance was carried out, separately for infestation and injury, with combined second and third generation data, for vine and half-vine sample units. Common regression coefficients were estimated to permit use of the sampling plan over a wide range of conditions. Optimum sample sizes for infestation and injury, at three levels of precision, were developed. An investigation of a multistage sampling plan with a nested analysis of variance showed that if the goal of sampling is focusing on larval infestation, three grape clusters should be sampled in a half-vine; if the goal of sampling is focusing on injury, then two grape clusters per half-vine are recommended.
Constrained statistical inference: sample-size tables for ANOVA and regression.
Vanbrabant, Leonard; Van De Schoot, Rens; Rosseel, Yves
2014-01-01
Researchers in the social and behavioral sciences often have clear expectations about the order/direction of the parameters in their statistical model. For example, a researcher might expect that regression coefficient β1 is larger than β2 and β3. The corresponding hypothesis is H: β1 > {β2, β3} and this is known as an (order) constrained hypothesis. A major advantage of testing such a hypothesis is that power can be gained and inherently a smaller sample size is needed. This article discusses this gain in sample size reduction, when an increasing number of constraints is included into the hypothesis. The main goal is to present sample-size tables for constrained hypotheses. A sample-size table contains the necessary sample-size at a pre-specified power (say, 0.80) for an increasing number of constraints. To obtain sample-size tables, two Monte Carlo simulations were performed, one for ANOVA and one for multiple regression. Three results are salient. First, in an ANOVA the needed sample-size decreases with 30-50% when complete ordering of the parameters is taken into account. Second, small deviations from the imposed order have only a minor impact on the power. Third, at the maximum number of constraints, the linear regression results are comparable with the ANOVA results. However, in the case of fewer constraints, ordering the parameters (e.g., β1 > β2) results in a higher power than assigning a positive or a negative sign to the parameters (e.g., β1 > 0).
n4Studies: Sample Size Calculation for an Epidemiological Study on a Smart Device
Chetta Ngamjarus
2016-05-01
Full Text Available Objective: This study was to develop a sample size application (called “n4Studies” for free use on iPhone and Android devices and to compare sample size functions between n4Studies with other applications and software. Methods: Objective-C programming language was used to create the application for the iPhone OS (operating system while javaScript, jquery mobile, PhoneGap and jstat were used to develop it for Android phones. Other sample size applications were searched from the Apple app and Google play stores. The applications’ characteristics and sample size functions were collected. Spearman’s rank correlation was used to investigate the relationship between number of sample size functions and price. Results: “n4Studies” provides several functions for sample size and power calculations for various epidemiological study designs. It can be downloaded from the Apple App and Google play store. Comparing n4Studies with other applications, it covers several more types of epidemiological study designs, gives similar results for estimation of infinite/finite population mean and infinite/finite proportion from GRANMO, for comparing two independent means from BioStats, for comparing two independent proportions from EpiCal application. When using the same parameters, n4Studies gives similar results to STATA, epicalc package in R, PS, G*Power, and OpenEpi. Conclusion: “n4Studies” can be an alternative tool for calculating the sample size. It may be useful to students, lecturers and researchers in conducting their research projects.
Sample size for collecting germplasms – a polyploid model with mixed mating system
R L Sapra; Prem Narain; S V S Chauhan; S K Lal; B B Singh
2003-03-01
The present paper discusses a general expression for determining the minimum sample size (plants) for a given number of seeds or vice versa for capturing multiple allelic diversity. The model considers sampling from a large 2 k-ploid population under a broad range of mating systems. Numerous expressions/results developed for germplasm collection/regeneration for diploid populations by earlier workers can be directly deduced from our general expression by assigning appropriate values of the corresponding parameters. A seed factor which influences the plant sample size has also been isolated to aid the collectors in selecting the appropriate combination of number of plants and seeds per plant. When genotypic multiplicity of seeds is taken into consideration, a sample size of even less than 172 plants can conserve diversity of 20 alleles from 50,000 polymorphic loci with a very large probability of conservation (0.9999) in most of the cases.
Lachin, John M
2006-10-15
Various methods have been described for re-estimating the final sample size in a clinical trial based on an interim assessment of the treatment effect. Many re-weight the observations after re-sizing so as to control the pursuant inflation in the type I error probability alpha. Lan and Trost (Estimation of parameters and sample size re-estimation. Proceedings of the American Statistical Association Biopharmaceutical Section 1997; 48-51) proposed a simple procedure based on conditional power calculated under the current trend in the data (CPT). The study is terminated for futility if CPT or = CU, or re-sized by a factor m to yield CPT = CU if CL stopping for futility can balance the inflation due to sample size re-estimation, thus permitting any form of final analysis with no re-weighting. Herein the statistical properties of this approach are described including an evaluation of the probabilities of stopping for futility or re-sizing, the distribution of the re-sizing factor m, and the unconditional type I and II error probabilities alpha and beta. Since futility stopping does not allow a type I error but commits a type II error, then as the probability of stopping for futility increases, alpha decreases and beta increases. An iterative procedure is described for choice of the critical test value and the futility stopping boundary so as to ensure that specified alpha and beta are obtained. However, inflation in beta is controlled by reducing the probability of futility stopping, that in turn dramatically increases the possible re-sizing factor m. The procedure is also generalized to limit the maximum sample size inflation factor, such as at m max = 4. However, doing so then allows for a non-trivial fraction of studies to be re-sized at this level that still have low conditional power. These properties also apply to other methods for sample size re-estimation with a provision for stopping for futility. Sample size re-estimation procedures should be used with caution
2014-01-01
) with an inlet passage way (16). The upper end of the inlet pipe (6) is connected with a top cap (9). The top cap (9) and the bottom cap (10) are mutually connected by means of a wire (8) and the top cap (9) is configured as a floating device providing a buoyancy force larger than the downwardly directed force......An inlet stratification (5) is adapted to be arranged vertically in a tank (1) during operation. The stratification device (5) comprises an inlet pipe (6) formed of a flexible porous material and having a lower and upper end. The lower end of the inlet pipe (6) is connected to a bottom cap (10...
Power and sample size calculations for Mendelian randomization studies using one genetic instrument.
Freeman, Guy; Cowling, Benjamin J; Schooling, C Mary
2013-08-01
Mendelian randomization, which is instrumental variable analysis using genetic variants as instruments, is an increasingly popular method of making causal inferences from observational studies. In order to design efficient Mendelian randomization studies, it is essential to calculate the sample sizes required. We present formulas for calculating the power of a Mendelian randomization study using one genetic instrument to detect an effect of a given size, and the minimum sample size required to detect effects for given levels of significance and power, using asymptotic statistical theory. We apply the formulas to some example data and compare the results with those from simulation methods. Power and sample size calculations using these formulas should be more straightforward to carry out than simulation approaches. These formulas make explicit that the sample size needed for Mendelian randomization study is inversely proportional to the square of the correlation between the genetic instrument and the exposure and proportional to the residual variance of the outcome after removing the effect of the exposure, as well as inversely proportional to the square of the effect size.
National Center for Education Statistics (DHEW), Washington, DC.
A complex two-stage sample selection process was used in designing the National Longitudinal Study of the High School Class of 1972. The first-stage sampling frame used in the selection of schools was stratified by the following seven variables: public vs. private control, geographic region, grade 12 enrollment, proximity to institutions of higher…
Sample size re-estimation in paired comparative diagnostic accuracy studies with a binary response.
McCray, Gareth P J; Titman, Andrew C; Ghaneh, Paula; Lancaster, Gillian A
2017-07-14
The sample size required to power a study to a nominal level in a paired comparative diagnostic accuracy study, i.e. studies in which the diagnostic accuracy of two testing procedures is compared relative to a gold standard, depends on the conditional dependence between the two tests - the lower the dependence the greater the sample size required. A priori, we usually do not know the dependence between the two tests and thus cannot determine the exact sample size required. One option is to use the implied sample size for the maximal negative dependence, giving the largest possible sample size. However, this is potentially wasteful of resources and unnecessarily burdensome on study participants as the study is likely to be overpowered. A more accurate estimate of the sample size can be determined at a planned interim analysis point where the sample size is re-estimated. This paper discusses a sample size estimation and re-estimation method based on the maximum likelihood estimates, under an implied multinomial model, of the observed values of conditional dependence between the two tests and, if required, prevalence, at a planned interim. The method is illustrated by comparing the accuracy of two procedures for the detection of pancreatic cancer, one procedure using the standard battery of tests, and the other using the standard battery with the addition of a PET/CT scan all relative to the gold standard of a cell biopsy. Simulation of the proposed method illustrates its robustness under various conditions. The results show that the type I error rate of the overall experiment is stable using our suggested method and that the type II error rate is close to or above nominal. Furthermore, the instances in which the type II error rate is above nominal are in the situations where the lowest sample size is required, meaning a lower impact on the actual number of participants recruited. We recommend multinomial model maximum likelihood estimation of the conditional
Son, Dae-Soon; Lee, DongHyuk; Lee, Kyusang; Jung, Sin-Ho; Ahn, Taejin; Lee, Eunjin; Sohn, Insuk; Chung, Jongsuk; Park, Woongyang; Huh, Nam; Lee, Jae Won
2015-02-01
An empirical method of sample size determination for building prediction models was proposed recently. Permutation method which is used in this procedure is a commonly used method to address the problem of overfitting during cross-validation while evaluating the performance of prediction models constructed from microarray data. But major drawback of such methods which include bootstrapping and full permutations is prohibitively high cost of computation required for calculating the sample size. In this paper, we propose that a single representative null distribution can be used instead of a full permutation by using both simulated and real data sets. During simulation, we have used a dataset with zero effect size and confirmed that the empirical type I error approaches to 0.05. Hence this method can be confidently applied to reduce overfitting problem during cross-validation. We have observed that pilot data set generated by random sampling from real data could be successfully used for sample size determination. We present our results using an experiment that was repeated for 300 times while producing results comparable to that of full permutation method. Since we eliminate full permutation, sample size estimation time is not a function of pilot data size. In our experiment we have observed that this process takes around 30min. With the increasing number of clinical studies, developing efficient sample size determination methods for building prediction models is critical. But empirical methods using bootstrap and permutation usually involve high computing costs. In this study, we propose a method that can reduce required computing time drastically by using representative null distribution of permutations. We use data from pilot experiments to apply this method for designing clinical studies efficiently for high throughput data.
A simple nomogram for sample size for estimating sensitivity and specificity of medical tests
Malhotra Rajeev
2010-01-01
Full Text Available Sensitivity and specificity measure inherent validity of a diagnostic test against a gold standard. Researchers develop new diagnostic methods to reduce the cost, risk, invasiveness, and time. Adequate sample size is a must to precisely estimate the validity of a diagnostic test. In practice, researchers generally decide about the sample size arbitrarily either at their convenience, or from the previous literature. We have devised a simple nomogram that yields statistically valid sample size for anticipated sensitivity or anticipated specificity. MS Excel version 2007 was used to derive the values required to plot the nomogram using varying absolute precision, known prevalence of disease, and 95% confidence level using the formula already available in the literature. The nomogram plot was obtained by suitably arranging the lines and distances to conform to this formula. This nomogram could be easily used to determine the sample size for estimating the sensitivity or specificity of a diagnostic test with required precision and 95% confidence level. Sample size at 90% and 99% confidence level, respectively, can also be obtained by just multiplying 0.70 and 1.75 with the number obtained for the 95% confidence level. A nomogram instantly provides the required number of subjects by just moving the ruler and can be repeatedly used without redoing the calculations. This can also be applied for reverse calculations. This nomogram is not applicable for testing of the hypothesis set-up and is applicable only when both diagnostic test and gold standard results have a dichotomous category.
Demonstration of Multi- and Single-Reader Sample Size Program for Diagnostic Studies software.
Hillis, Stephen L; Schartz, Kevin M
2015-02-01
The recently released software Multi- and Single-Reader Sample Size Sample Size Program for Diagnostic Studies, written by Kevin Schartz and Stephen Hillis, performs sample size computations for diagnostic reader-performance studies. The program computes the sample size needed to detect a specified difference in a reader performance measure between two modalities, when using the analysis methods initially proposed by Dorfman, Berbaum, and Metz (DBM) and Obuchowski and Rockette (OR), and later unified and improved by Hillis and colleagues. A commonly used reader performance measure is the area under the receiver-operating-characteristic curve. The program can be used with typical common reader-performance measures which can be estimated parametrically or nonparametrically. The program has an easy-to-use step-by-step intuitive interface that walks the user through the entry of the needed information. Features of the software include the following: (1) choice of several study designs; (2) choice of inputs obtained from either OR or DBM analyses; (3) choice of three different inference situations: both readers and cases random, readers fixed and cases random, and readers random and cases fixed; (4) choice of two types of hypotheses: equivalence or noninferiority; (6) choice of two output formats: power for specified case and reader sample sizes, or a listing of case-reader combinations that provide a specified power; (7) choice of single or multi-reader analyses; and (8) functionality in Windows, Mac OS, and Linux.
Tsai, Chen-An; Huang, Chih-Yang; Liu, Jen-Pei
2014-08-30
The approval of generic drugs requires the evidence of average bioequivalence (ABE) on both the area under the concentration-time curve and the peak concentration Cmax . The bioequivalence (BE) hypothesis can be decomposed into the non-inferiority (NI) and non-superiority (NS) hypothesis. Most of regulatory agencies employ the two one-sided tests (TOST) procedure to test ABE between two formulations. As it is based on the intersection-union principle, the TOST procedure is conservative in terms of the type I error rate. However, the type II error rate is the sum of the type II error rates with respect to each null hypothesis of NI and NS hypotheses. When the difference in population means between two treatments is not 0, no close-form solution for the sample size for the BE hypothesis is available. Current methods provide the sample sizes with either insufficient power or unnecessarily excessive power. We suggest an approximate method for sample size determination, which can also provide the type II rate for each of NI and NS hypotheses. In addition, the proposed method is flexible to allow extension from one pharmacokinetic (PK) response to determination of the sample size required for multiple PK responses. We report the results of a numerical study. An R code is provided to calculate the sample size for BE testing based on the proposed methods.
Sleeth, Darrah K
2013-05-01
In 2010, the American Conference of Governmental Industrial Hygienists (ACGIH) formally changed its Threshold Limit Value (TLV) for beryllium from a 'total' particulate sample to an inhalable particulate sample. This change may have important implications for workplace air sampling of beryllium. A history of particle size-selective sampling methods, with a special focus on beryllium, will be provided. The current state of the science on inhalable sampling will also be presented, including a look to the future at what new methods or technology may be on the horizon. This includes new sampling criteria focused on particle deposition in the lung, proposed changes to the existing inhalable convention, as well as how the issues facing beryllium sampling may help drive other changes in sampling technology.
Information-based sample size re-estimation in group sequential design for longitudinal trials.
Zhou, Jing; Adewale, Adeniyi; Shentu, Yue; Liu, Jiajun; Anderson, Keaven
2014-09-28
Group sequential design has become more popular in clinical trials because it allows for trials to stop early for futility or efficacy to save time and resources. However, this approach is less well-known for longitudinal analysis. We have observed repeated cases of studies with longitudinal data where there is an interest in early stopping for a lack of treatment effect or in adapting sample size to correct for inappropriate variance assumptions. We propose an information-based group sequential design as a method to deal with both of these issues. Updating the sample size at each interim analysis makes it possible to maintain the target power while controlling the type I error rate. We will illustrate our strategy with examples and simulations and compare the results with those obtained using fixed design and group sequential design without sample size re-estimation.
A comparison of methods for sample size estimation for non-inferiority studies with binary outcomes.
Julious, Steven A; Owen, Roger J
2011-12-01
Non-inferiority trials are motivated in the context of clinical research where a proven active treatment exists and placebo-controlled trials are no longer acceptable for ethical reasons. Instead, active-controlled trials are conducted where a treatment is compared to an established treatment with the objective of demonstrating that it is non-inferior to this treatment. We review and compare the methodologies for calculating sample sizes and suggest appropriate methods to use. We demonstrate how the simplest method of using the anticipated response is predominantly consistent with simulations. In the context of trials with binary outcomes with expected high proportions of positive responses, we show how the sample size is quite sensitive to assumptions about the control response. We recommend when designing such a study that sensitivity analyses be performed with respect to the underlying assumptions and that the Bayesian methods described in this article be adopted to assess sample size.
Species-genetic diversity correlations in habitat fragmentation can be biased by small sample sizes.
Nazareno, Alison G; Jump, Alistair S
2012-06-01
Predicted parallel impacts of habitat fragmentation on genes and species lie at the core of conservation biology, yet tests of this rule are rare. In a recent article in Ecology Letters, Struebig et al. (2011) report that declining genetic diversity accompanies declining species diversity in tropical forest fragments. However, this study estimates diversity in many populations through extrapolation from very small sample sizes. Using the data of this recent work, we show that results estimated from the smallest sample sizes drive the species-genetic diversity correlation (SGDC), owing to a false-positive association between habitat fragmentation and loss of genetic diversity. Small sample sizes are a persistent problem in habitat fragmentation studies, the results of which often do not fit simple theoretical models. It is essential, therefore, that data assessing the proposed SGDC are sufficient in order that conclusions be robust.
Threshold-dependent sample sizes for selenium assessment with stream fish tissue
Hitt, Nathaniel P.; Smith, David R.
2015-01-01
Natural resource managers are developing assessments of selenium (Se) contamination in freshwater ecosystems based on fish tissue concentrations. We evaluated the effects of sample size (i.e., number of fish per site) on the probability of correctly detecting mean whole-body Se values above a range of potential management thresholds. We modeled Se concentrations as gamma distributions with shape and scale parameters fitting an empirical mean-to-variance relationship in data from southwestern West Virginia, USA (63 collections, 382 individuals). We used parametric bootstrapping techniques to calculate statistical power as the probability of detecting true mean concentrations up to 3 mg Se/kg above management thresholds ranging from 4 to 8 mg Se/kg. Sample sizes required to achieve 80% power varied as a function of management thresholds and Type I error tolerance (α). Higher thresholds required more samples than lower thresholds because populations were more heterogeneous at higher mean Se levels. For instance, to assess a management threshold of 4 mg Se/kg, a sample of eight fish could detect an increase of approximately 1 mg Se/kg with 80% power (given α = 0.05), but this sample size would be unable to detect such an increase from a management threshold of 8 mg Se/kg with more than a coin-flip probability. Increasing α decreased sample size requirements to detect above-threshold mean Se concentrations with 80% power. For instance, at an α-level of 0.05, an 8-fish sample could detect an increase of approximately 2 units above a threshold of 8 mg Se/kg with 80% power, but when α was relaxed to 0.2, this sample size was more sensitive to increasing mean Se concentrations, allowing detection of an increase of approximately 1.2 units with equivalent power. Combining individuals into 2- and 4-fish composite samples for laboratory analysis did not decrease power because the reduced number of laboratory samples was compensated for by increased
Sample size requirements for indirect association studies of gene-environment interactions (G x E).
Hein, Rebecca; Beckmann, Lars; Chang-Claude, Jenny
2008-04-01
Association studies accounting for gene-environment interactions (G x E) may be useful for detecting genetic effects. Although current technology enables very dense marker spacing in genetic association studies, the true disease variants may not be genotyped. Thus, causal genes are searched for by indirect association using genetic markers in linkage disequilibrium (LD) with the true disease variants. Sample sizes needed to detect G x E effects in indirect case-control association studies depend on the true genetic main effects, disease allele frequencies, whether marker and disease allele frequencies match, LD between loci, main effects and prevalence of environmental exposures, and the magnitude of interactions. We explored variables influencing sample sizes needed to detect G x E, compared these sample sizes with those required to detect genetic marginal effects, and provide an algorithm for power and sample size estimations. Required sample sizes may be heavily inflated if LD between marker and disease loci decreases. More than 10,000 case-control pairs may be required to detect G x E. However, given weak true genetic main effects, moderate prevalence of environmental exposures, as well as strong interactions, G x E effects may be detected with smaller sample sizes than those needed for the detection of genetic marginal effects. Moreover, in this scenario, rare disease variants may only be detectable when G x E is included in the analyses. Thus, the analysis of G x E appears to be an attractive option for the detection of weak genetic main effects of rare variants that may not be detectable in the analysis of genetic marginal effects only.
Estimation of grain size in asphalt samples using digital image analysis
Källén, Hanna; Heyden, Anders; Lindh, Per
2014-09-01
Asphalt is made of a mixture of stones of different sizes and a binder called bitumen, the size distribution of the stones is determined by the recipe of the asphalt. One quality check of asphalt is to see if the real size distribution of asphalt samples is consistent with the recipe. This is usually done by first extracting the binder using methylenchloride and the sieving the stones and see how much that pass every sieve size. Methylenchloride is highly toxic and it is desirable to find the size distribution in some other way. In this paper we find the size distribution by slicing up the asphalt sample and using image analysis techniques to analyze the cross-sections. First the stones are segmented from the background, bitumen, and then rectangles are fit to the detected stones. We then estimate the sizes of the stones by using the width of the rectangle. The result is compared with both the recipe for the asphalt and with the result from the standard analysis method, and our method shows good correlation with those.
Hoyle, Rick H; Gottfredson, Nisha C
2015-10-01
When the goal of prevention research is to capture in statistical models some measure of the dynamic complexity in structures and processes implicated in problem behavior and its prevention, approaches such as multilevel modeling (MLM) and structural equation modeling (SEM) are indicated. Yet the assumptions that must be satisfied if these approaches are to be used responsibly raise concerns regarding their use in prevention research involving smaller samples. In this article, we discuss in nontechnical terms the role of sample size in MLM and SEM and present findings from the latest simulation work on the performance of each approach at sample sizes typical of prevention research. For each statistical approach, we draw from extant simulation studies to establish lower bounds for sample size (e.g., MLM can be applied with as few as ten groups comprising ten members with normally distributed data, restricted maximum likelihood estimation, and a focus on fixed effects; sample sizes as small as N = 50 can produce reliable SEM results with normally distributed data and at least three reliable indicators per factor) and suggest strategies for making the best use of the modeling approach when N is near the lower bound.
Chung, Ren-Hua; Schmidt, Michael A; Morris, Richard W; Martin, Eden R
2010-11-01
The recent successes of GWAS based on large sample sizes motivate combining independent datasets to obtain larger sample sizes and thereby increase statistical power. Analysis methods that can accommodate different study designs, such as family-based and case-control designs, are of general interest. However, population stratification can cause spurious association for population-based association analyses. For family-based association analysis that infers missing parental genotypes based on the allele frequencies estimated in the entire sample, the parental mating-type probabilities may not be correctly estimated in the presence of population stratification. Therefore, any approach to combining family and case-control data should also properly account for population stratification. Although several methods have been proposed to accommodate family-based and case-control data, all have restrictions. Most of them require sampling a homogeneous population, which may not be a reasonable assumption for data from a large consortium. One of the methods, FamCC, can account for population stratification and uses nuclear families with arbitrary number of siblings but requires parental genotype data, which are often unavailable for late-onset diseases. We extended the family-based test, Association in the Presence of Linkage (APL), to combine family and case-control data (CAPL). CAPL can accommodate case-control data and families with multiple affected siblings and missing parents in the presence of population stratification. We used simulations to demonstrate that CAPL is a valid test either in a homogeneous population or in the presence of population stratification. We also showed that CAPL can have more power than other methods that combine family and case-control data.
A simulation study of sample size for multilevel logistic regression models
Moineddin Rahim
2007-07-01
Full Text Available Abstract Background Many studies conducted in health and social sciences collect individual level data as outcome measures. Usually, such data have a hierarchical structure, with patients clustered within physicians, and physicians clustered within practices. Large survey data, including national surveys, have a hierarchical or clustered structure; respondents are naturally clustered in geographical units (e.g., health regions and may be grouped into smaller units. Outcomes of interest in many fields not only reflect continuous measures, but also binary outcomes such as depression, presence or absence of a disease, and self-reported general health. In the framework of multilevel studies an important problem is calculating an adequate sample size that generates unbiased and accurate estimates. Methods In this paper simulation studies are used to assess the effect of varying sample size at both the individual and group level on the accuracy of the estimates of the parameters and variance components of multilevel logistic regression models. In addition, the influence of prevalence of the outcome and the intra-class correlation coefficient (ICC is examined. Results The results show that the estimates of the fixed effect parameters are unbiased for 100 groups with group size of 50 or higher. The estimates of the variance covariance components are slightly biased even with 100 groups and group size of 50. The biases for both fixed and random effects are severe for group size of 5. The standard errors for fixed effect parameters are unbiased while for variance covariance components are underestimated. Results suggest that low prevalent events require larger sample sizes with at least a minimum of 100 groups and 50 individuals per group. Conclusion We recommend using a minimum group size of 50 with at least 50 groups to produce valid estimates for multi-level logistic regression models. Group size should be adjusted under conditions where the prevalence
Sheehan, Sara; Harris, Kelley; Song, Yun S
2013-07-01
Throughout history, the population size of modern humans has varied considerably due to changes in environment, culture, and technology. More accurate estimates of population size changes, and when they occurred, should provide a clearer picture of human colonization history and help remove confounding effects from natural selection inference. Demography influences the pattern of genetic variation in a population, and thus genomic data of multiple individuals sampled from one or more present-day populations contain valuable information about the past demographic history. Recently, Li and Durbin developed a coalescent-based hidden Markov model, called the pairwise sequentially Markovian coalescent (PSMC), for a pair of chromosomes (or one diploid individual) to estimate past population sizes. This is an efficient, useful approach, but its accuracy in the very recent past is hampered by the fact that, because of the small sample size, only few coalescence events occur in that period. Multiple genomes from the same population contain more information about the recent past, but are also more computationally challenging to study jointly in a coalescent framework. Here, we present a new coalescent-based method that can efficiently infer population size changes from multiple genomes, providing access to a new store of information about the recent past. Our work generalizes the recently developed sequentially Markov conditional sampling distribution framework, which provides an accurate approximation of the probability of observing a newly sampled haplotype given a set of previously sampled haplotypes. Simulation results demonstrate that we can accurately reconstruct the true population histories, with a significant improvement over the PSMC in the recent past. We apply our method, called diCal, to the genomes of multiple human individuals of European and African ancestry to obtain a detailed population size change history during recent times.
Wejnert, Cyprian; Pham, Huong; Krishna, Nevin; Le, Binh; DiNenno, Elizabeth
2012-05-01
Respondent-driven sampling (RDS) has become increasingly popular for sampling hidden populations, including injecting drug users (IDU). However, RDS data are unique and require specialized analysis techniques, many of which remain underdeveloped. RDS sample size estimation requires knowing design effect (DE), which can only be calculated post hoc. Few studies have analyzed RDS DE using real world empirical data. We analyze estimated DE from 43 samples of IDU collected using a standardized protocol. We find the previous recommendation that sample size be at least doubled, consistent with DE = 2, underestimates true DE and recommend researchers use DE = 4 as an alternate estimate when calculating sample size. A formula for calculating sample size for RDS studies among IDU is presented. Researchers faced with limited resources may wish to accept slightly higher standard errors to keep sample size requirements low. Our results highlight dangers of ignoring sampling design in analysis.
Correction of population stratification in large multi-ethnic association studies.
David Serre
Full Text Available BACKGROUND: The vast majority of genetic risk factors for complex diseases have, taken individually, a small effect on the end phenotype. Population-based association studies therefore need very large sample sizes to detect significant differences between affected and non-affected individuals. Including thousands of affected individuals in a study requires recruitment in numerous centers, possibly from different geographic regions. Unfortunately such a recruitment strategy is likely to complicate the study design and to generate concerns regarding population stratification. METHODOLOGY/PRINCIPAL FINDINGS: We analyzed 9,751 individuals representing three main ethnic groups - Europeans, Arabs and South Asians - that had been enrolled from 154 centers involving 52 countries for a global case/control study of acute myocardial infarction. All individuals were genotyped at 103 candidate genes using 1,536 SNPs selected with a tagging strategy that captures most of the genetic diversity in different populations. We show that relying solely on self-reported ethnicity is not sufficient to exclude population stratification and we present additional methods to identify and correct for stratification. CONCLUSIONS/SIGNIFICANCE: Our results highlight the importance of carefully addressing population stratification and of carefully "cleaning" the sample prior to analyses to obtain stronger signals of association and to avoid spurious results.
Umesh P. Agarwal; Sally A. Ralph; Carlos Baez; Richard S. Reiner; Steve P. Verrill
2017-01-01
Although X-ray diffraction (XRD) has been the most widely used technique to investigate crystallinity index (CrI) and crystallite size (L200) of cellulose materials, there are not many studies that have taken into account the role of sample moisture on these measurements. The present investigation focuses on a variety of celluloses and cellulose...
Power and Sample Size Calculations for Logistic Regression Tests for Differential Item Functioning
Li, Zhushan
2014-01-01
Logistic regression is a popular method for detecting uniform and nonuniform differential item functioning (DIF) effects. Theoretical formulas for the power and sample size calculations are derived for likelihood ratio tests and Wald tests based on the asymptotic distribution of the maximum likelihood estimators for the logistic regression model.…
Size Distributions and Characterization of Native and Ground Samples for Toxicology Studies
McKay, David S.; Cooper, Bonnie L.; Taylor, Larry A.
2010-01-01
This slide presentation shows charts and graphs that review the particle size distribution and characterization of natural and ground samples for toxicology studies. There are graphs which show the volume distribution versus the number distribution for natural occurring dust, jet mill ground dust, and ball mill ground dust.
Sample Size Calculation for Estimating or Testing a Nonzero Squared Multiple Correlation Coefficient
Krishnamoorthy, K.; Xia, Yanping
2008-01-01
The problems of hypothesis testing and interval estimation of the squared multiple correlation coefficient of a multivariate normal distribution are considered. It is shown that available one-sided tests are uniformly most powerful, and the one-sided confidence intervals are uniformly most accurate. An exact method of calculating sample size to…
Sample Size Calculation for Estimating or Testing a Nonzero Squared Multiple Correlation Coefficient
Krishnamoorthy, K.; Xia, Yanping
2008-01-01
The problems of hypothesis testing and interval estimation of the squared multiple correlation coefficient of a multivariate normal distribution are considered. It is shown that available one-sided tests are uniformly most powerful, and the one-sided confidence intervals are uniformly most accurate. An exact method of calculating sample size to…
Analysis of variograms with various sample sizes from a multispectral image
Variogram plays a crucial role in remote sensing application and geostatistics. It is very important to estimate variogram reliably from sufficient data. In this study, the analysis of variograms with various sample sizes of remotely sensed data was conducted. A 100x100-pixel subset was chosen from ...
Fan, Xitao; Wang, Lin; Thompson, Bruce
1999-01-01
A Monte Carlo simulation study investigated the effects on 10 structural equation modeling fit indexes of sample size, estimation method, and model specification. Some fit indexes did not appear to be comparable, and it was apparent that estimation method strongly influenced almost all fit indexes examined, especially for misspecified models. (SLD)
Sideridis, Georgios; Simos, Panagiotis; Papanicolaou, Andrew; Fletcher, Jack
2014-01-01
The present study assessed the impact of sample size on the power and fit of structural equation modeling applied to functional brain connectivity hypotheses. The data consisted of time-constrained minimum norm estimates of regional brain activity during performance of a reading task obtained with magnetoencephalography. Power analysis was first…
B-graph sampling to estimate the size of a hidden population
Spreen, M.; Bogaerts, S.
2015-01-01
Link-tracing designs are often used to estimate the size of hidden populations by utilizing the relational links between their members. A major problem in studies of hidden populations is the lack of a convenient sampling frame. The most frequently applied design in studies of hidden populations is
Got Power? A Systematic Review of Sample Size Adequacy in Health Professions Education Research
Cook, David A.; Hatala, Rose
2015-01-01
Many education research studies employ small samples, which in turn lowers statistical power. We re-analyzed the results of a meta-analysis of simulation-based education to determine study power across a range of effect sizes, and the smallest effect that could be plausibly excluded. We systematically searched multiple databases through May 2011,…
A preliminary model to avoid the overestimation of sample size in bioequivalence studies.
Ramírez, E; Abraira, V; Guerra, P; Borobia, A M; Duque, B; López, J L; Mosquera, B; Lubomirov, R; Carcas, A J; Frías, J
2013-02-01
Often the only available data in literature for sample size estimations in bioequivalence studies is intersubject variability, which tends to result in overestimation of sample size. In this paper, we proposed a preliminary model of intrasubject variability based on intersubject variability for Cmax and AUC data from randomized, crossovers, bioequivalence (BE) studies. From 93 Cmax and 121 AUC data from test-reference comparisons that fulfilled BE criteria, we calculated intersubject variability for the reference formulation and intrasubject variability from ANOVA. Lineal and exponential models (y=a(1-e-bx)) were fitted weighted by the inverse of the variance, to predict the intrasubject variability based on intersubject variability. To validate the model we calculated the coefficient of cross-validation of data from 30 new BE studies. The models fit very well (R2=0.997 and 0.990 for Cmax and AUC respectively) and the cross-validation correlation were 0.847 for Cmax and 0.572 for AUC. A preliminary model analyses allow us to estimate the intrasubject variability based on intersubject variability for sample size calculation purposes in BE studies. This approximation provides an opportunity for sample size reduction avoiding unnecessary exposure of healthy volunteers. Further modelling studies are desirable to confirm these results especially suggestions of the higher intersubject variability range.
On the repeated measures designs and sample sizes for randomized controlled trials.
Tango, Toshiro
2016-04-01
For the analysis of longitudinal or repeated measures data, generalized linear mixed-effects models provide a flexible and powerful tool to deal with heterogeneity among subject response profiles. However, the typical statistical design adopted in usual randomized controlled trials is an analysis of covariance type analysis using a pre-defined pair of "pre-post" data, in which pre-(baseline) data are used as a covariate for adjustment together with other covariates. Then, the major design issue is to calculate the sample size or the number of subjects allocated to each treatment group. In this paper, we propose a new repeated measures design and sample size calculations combined with generalized linear mixed-effects models that depend not only on the number of subjects but on the number of repeated measures before and after randomization per subject used for the analysis. The main advantages of the proposed design combined with the generalized linear mixed-effects models are (1) it can easily handle missing data by applying the likelihood-based ignorable analyses under the missing at random assumption and (2) it may lead to a reduction in sample size, compared with the simple pre-post design. The proposed designs and the sample size calculations are illustrated with real data arising from randomized controlled trials.
The Influence of Virtual Sample Size on Confidence and Causal-Strength Judgments
Liljeholm, Mimi; Cheng, Patricia W.
2009-01-01
The authors investigated whether confidence in causal judgments varies with virtual sample size--the frequency of cases in which the outcome is (a) absent before the introduction of a generative cause or (b) present before the introduction of a preventive cause. Participants were asked to evaluate the influence of various candidate causes on an…
Power and Sample Size Calculations for Logistic Regression Tests for Differential Item Functioning
Li, Zhushan
2014-01-01
Logistic regression is a popular method for detecting uniform and nonuniform differential item functioning (DIF) effects. Theoretical formulas for the power and sample size calculations are derived for likelihood ratio tests and Wald tests based on the asymptotic distribution of the maximum likelihood estimators for the logistic regression model.…
Required sample size for monitoring stand dynamics in strict forest reserves: a case study
Diego Van Den Meersschaut; Bart De Cuyper; Kris Vandekerkhove; Noel Lust
2000-01-01
Stand dynamics in European strict forest reserves are commonly monitored using inventory densities of 5 to 15 percent of the total surface. The assumption that these densities guarantee a representative image of certain parameters is critically analyzed in a case study for the parameters basal area and stem number. The required sample sizes for different accuracy and...
Kelley, Ken
2007-11-01
The accuracy in parameter estimation approach to sample size planning is developed for the coefficient of variation, where the goal of the method is to obtain an accurate parameter estimate by achieving a sufficiently narrow confidence interval. The first method allows researchers to plan sample size so that the expected width of the confidence interval for the population coefficient of variation is sufficiently narrow. A modification allows a desired degree of assurance to be incorporated into the method, so that the obtained confidence interval will be sufficiently narrow with some specified probability (e.g., 85% assurance that the 95 confidence interval width will be no wider than to units). Tables of necessary sample size are provided for a variety of scenarios that may help researchers planning a study where the coefficient of variation is of interest plan an appropriate sample size in order to have a sufficiently narrow confidence interval, optionally with somespecified assurance of the confidence interval being sufficiently narrow. Freely available computer routines have been developed that allow researchers to easily implement all of the methods discussed in the article.
European environmental stratifications and typologies
Hazeu, G.W,; Metzger, M.J.; Mücher, C.A.
2011-01-01
A range of new spatial datasets classifying the European environment has been constructed over the last few years. These datasets share the common objective of dividing European environmental gradients into convenient units, within which objects and variables of interest have relatively homogeneous...... characteristics. The stratifications and typologies can be used as a basis for up-scaling, for stratified random sampling of ecological resources, for the representative selection of sites for studies across the continent and for the provision of frameworks for modeling exercises and reporting at the European...... their limitations and challenges. As such, they provide a sound basis for describing the factors affecting the robustness of such datasets. The latter is especially relevant, since there is likely to be further interest in European environmental assessment. In addition, advances in data availability and analysis...
Gutenberg-Richter b-value maximum likelihood estimation and sample size
Nava, F. A.; Márquez-Ramírez, V. H.; Zúñiga, F. R.; Ávila-Barrientos, L.; Quinteros, C. B.
2017-01-01
The Aki-Utsu maximum likelihood method is widely used for estimation of the Gutenberg-Richter b-value, but not all authors are conscious of the method's limitations and implicit requirements. The Aki/Utsu method requires a representative estimate of the population mean magnitude; a requirement seldom satisfied in b-value studies, particularly in those that use data from small geographic and/or time windows, such as b-mapping and b-vs-time studies. Monte Carlo simulation methods are used to determine how large a sample is necessary to achieve representativity, particularly for rounded magnitudes. The size of a representative sample weakly depends on the actual b-value. It is shown that, for commonly used precisions, small samples give meaningless estimations of b. Our results give estimates on the probabilities of getting correct estimates of b for a given desired precision for samples of different sizes. We submit that all published studies reporting b-value estimations should include information about the size of the samples used.
Factors Influencing Sample Size for Internal Audit Evidence Collection in the Public Sector in Kenya
Kamau Charles Guandaru
2017-01-01
Full Text Available The internal audit department has a role of providing objective assurance and consulting services designed to add value and improve an organization’s operations. In performing this role the internal auditors are required to provide an auditor’s opinion which is supported by sufficient and reliable audit evidence. Since auditors are not in a position to examine 100% of the records and transactions, they are required to sample a few and make conclusions on the basis of the sample selected. The literature suggests several factors which affects the sample size for audit purposes of the internal auditors in the public sector in Kenya. This research collected data from 32 public sector internal auditors. The research carried out simple regression and correlation analysis on the data collected so as to test hypotheses and make conclusions on the factors affecting the sample size for audit purposes of the internal auditors in the public sector in Kenya. The study found out that that materiality of audit issue, type of information available, source of information, degree of risk of misstatement and auditor skills and independence are some of the factors influencing the sample size determination for the purposes of internal audit evidence collection in public sector in Kenya.
Size selective isocyanate aerosols personal air sampling using porous plastic foams
Cong Khanh Huynh; Trinh Vu Duc, E-mail: chuynh@hospvd.c [Institut Universitaire Romand de Sante au Travail (IST), 21 rue du Bugnon - CH-1011 Lausanne (Switzerland)
2009-02-01
As part of a European project (SMT4-CT96-2137), various European institutions specialized in occupational hygiene (BGIA, HSL, IOM, INRS, IST, Ambiente e Lavoro) have established a program of scientific collaboration to develop one or more prototypes of European personal samplers for the collection of simultaneous three dust fractions: inhalable, thoracic and respirable. These samplers based on existing sampling heads (IOM, GSP and cassettes) use Polyurethane Plastic Foam (PUF) according to their porosity to support sampling and separator size of the particles. In this study, the authors present an original application of size selective personal air sampling using chemical impregnated PUF to perform isocyanate aerosols capturing and derivatizing in industrial spray-painting shops.
Andreasen, Jo Bønding; Pistor-Riebold, Thea Unger; Knudsen, Ingrid Hell;
2014-01-01
count remained stable using a 3.6 mL tube during the entire observation period of 120 min (p=0.74), but decreased significantly after 60 min when using tubes smaller than 3.6 mL (pblood sampling tubes. Therefore, 1.8 mL tubes should...... be preferred for RoTEM® analyses in order to minimise the volume of blood drawn. With regard to platelet aggregation analysed by impedance aggregometry tubes of different size cannot be used interchangeably. If platelet count is determined later than 10 min after blood sampling using tubes containing citrate......Background: To minimise the volume of blood used for diagnostic procedures, especially in children, we investigated whether the size of sample tubes affected whole blood coagulation analyses. Methods: We included 20 healthy individuals for rotational thromboelastometry (RoTEM®) analyses...
{sup 10}Be measurements at MALT using reduced-size samples of bulk sediments
Horiuchi, Kazuho, E-mail: kh@cc.hirosaki-u.ac.jp [Graduate School of Science and Technology, Hirosaki University, 3, Bunkyo-chou, Hirosaki, Aomori 036-8561 (Japan); Oniyanagi, Itsumi [Graduate School of Science and Technology, Hirosaki University, 3, Bunkyo-chou, Hirosaki, Aomori 036-8561 (Japan); Wasada, Hiroshi [Institute of Geology and Paleontology, Graduate school of Science, Tohoku University, 6-3, Aramaki Aza-Aoba, Aoba-ku, Sendai 980-8578 (Japan); Matsuzaki, Hiroyuki [MALT, School of Engineering, University of Tokyo, 2-11-16, Yayoi, Bunkyo-ku, Tokyo 113-0032 (Japan)
2013-01-15
In order to establish {sup 10}Be measurements on reduced-size (1-10 mg) samples of bulk sediments, we investigated four different pretreatment designs using lacustrine and marginal-sea sediments and the AMS system of the Micro Analysis Laboratory, Tandem accelerator (MALT) at University of Tokyo. The {sup 10}Be concentrations obtained from the samples of 1-10 mg agreed within a precision of 3-5% with the values previously determined using corresponding ordinary-size ({approx}200 mg) samples and the same AMS system. This fact demonstrates reliable determinations of {sup 10}Be with milligram levels of recent bulk sediments at MALT. On the other hand, a clear decline of the BeO{sup -} beam with tens of micrograms of {sup 9}Be carrier suggests that the combination of ten milligrams of sediments and a few hundred micrograms of the {sup 9}Be carrier is more convenient at this stage.
Thayse Aparecida Dourado
2012-08-01
Full Text Available The dispersion of the samples in soil particle-size analysis is a fundamental step, which is commonly achieved with a combination of chemical agents and mechanical agitation. The purpose of this study was to evaluate the efficiency of a low-speed reciprocal shaker for the mechanical dispersion of soil samples of different textural classes. The particle size of 61 soil samples was analyzed in four replications, using the pipette method to determine the clay fraction and sieving to determine coarse, fine and total sand fractions. The silt content was obtained by difference. To evaluate the performance, the results of the reciprocal shaker (RSh were compared with data of the same soil samples available in reports of the Proficiency testing for Soil Analysis Laboratories of the Agronomic Institute of Campinas (Prolab/IAC. The accuracy was analyzed based on the maximum and minimum values defining the confidence intervals for the particle-size fractions of each soil sample. Graphical indicators were also used for data comparison, based on dispersion and linear adjustment. The descriptive statistics indicated predominantly low variability in more than 90 % of the results for sand, medium-textured and clay samples, and for 68 % of the results for heavy clay samples, indicating satisfactory repeatability of measurements with the RSh. Medium variability was frequently associated with silt, followed by the fine sand fraction. The sensitivity analyses indicated an accuracy of 100 % for the three main separates (total sand, silt and clay, in all 52 samples of the textural classes heavy clay, clay and medium. For the nine sand soil samples, the average accuracy was 85.2 %; highest deviations were observed for the silt fraction. In relation to the linear adjustments, the correlation coefficients of 0.93 (silt or > 0.93 (total sand and clay, as well as the differences between the angular coefficients and the unit < 0.16, indicated a high correlation between the
Hawkins, K A; Tulsky, D S
2001-11-01
Since memory performance expectations may be IQ-based, unidirectional base rate data for IQ-Memory Score discrepancies are provided in the WAIS-III/WMS-III Technical Manual. The utility of these data partially rests on the assumption that discrepancy base rates do not vary across ability levels. FSIQ stratified base rate data generated from the standardization sample, however, demonstrate substantial variability across the IQ spectrum. A superiority of memory score over FSIQ is typical at lower IQ levels, whereas the converse is true at higher IQ levels. These data indicate that the use of IQ-memory score unstratified "simple difference" tables could lead to erroneous conclusions for clients with low or high IQ. IQ stratified standardization base rate data are provided as a complement to the "predicted difference" method detailed in the Technical Manual.
Sample size calculation for microarray experiments with blocked one-way design
Jung Sin-Ho
2009-05-01
Full Text Available Abstract Background One of the main objectives of microarray analysis is to identify differentially expressed genes for different types of cells or treatments. Many statistical methods have been proposed to assess the treatment effects in microarray experiments. Results In this paper, we consider discovery of the genes that are differentially expressed among K (> 2 treatments when each set of K arrays consists of a block. In this case, the array data among K treatments tend to be correlated because of block effect. We propose to use the blocked one-way ANOVA F-statistic to test if each gene is differentially expressed among K treatments. The marginal p-values are calculated using a permutation method accounting for the block effect, adjusting for the multiplicity of the testing procedure by controlling the false discovery rate (FDR. We propose a sample size calculation method for microarray experiments with a blocked one-way design. With FDR level and effect sizes of genes specified, our formula provides a sample size for a given number of true discoveries. Conclusion The calculated sample size is shown via simulations to provide an accurate number of true discoveries while controlling the FDR at the desired level.
Stratification of zooplankton in the northwestern Indian Ocean
Paulinose, V.T.; Gopalakrishnan, T.C.; Nair, K.K.C.; Aravindakshan, P.N.
Study on stratification of zooplankton in the north western Indian Ocean was carried out with special reference to its relative abundance and distribution. Samples were collected using multiple plankton net, during first cruise of ORV Sagar Kanya...
Bayesian sample size calculation for estimation of the difference between two binomial proportions.
Pezeshk, Hamid; Nematollahi, Nader; Maroufy, Vahed; Marriott, Paul; Gittins, John
2013-12-01
In this study, we discuss a decision theoretic or fully Bayesian approach to the sample size question in clinical trials with binary responses. Data are assumed to come from two binomial distributions. A Dirichlet distribution is assumed to describe prior knowledge of the two success probabilities p1 and p2. The parameter of interest is p = p1 - p2. The optimal size of the trial is obtained by maximising the expected net benefit function. The methodology presented in this article extends previous work by the assumption of dependent prior distributions for p1 and p2.
Profit based phase II sample size determination when adaptation by design is adopted
Martini, D.
2014-01-01
Background. Adaptation by design consists in conservatively estimating the phase III sample size on the basis of phase II data, and can be applied in almost all therapeutic areas; it is based on the assumption that the effect size of the drug is the same in phase II and phase III trials, that is a very common scenario assumed in product development. Adaptation by design reduces the probability on underpowered experiments and can improve the overall success probability of phase II and III tria...
Dong, Nianbo; Maynard, Rebecca
2013-01-01
This paper and the accompanying tool are intended to complement existing supports for conducting power analysis tools by offering a tool based on the framework of Minimum Detectable Effect Sizes (MDES) formulae that can be used in determining sample size requirements and in estimating minimum detectable effect sizes for a range of individual- and…
Resampling: An improvement of importance sampling in varying population size models.
Merle, C; Leblois, R; Rousset, F; Pudlo, P
2017-04-01
Sequential importance sampling algorithms have been defined to estimate likelihoods in models of ancestral population processes. However, these algorithms are based on features of the models with constant population size, and become inefficient when the population size varies in time, making likelihood-based inferences difficult in many demographic situations. In this work, we modify a previous sequential importance sampling algorithm to improve the efficiency of the likelihood estimation. Our procedure is still based on features of the model with constant size, but uses a resampling technique with a new resampling probability distribution depending on the pairwise composite likelihood. We tested our algorithm, called sequential importance sampling with resampling (SISR) on simulated data sets under different demographic cases. In most cases, we divided the computational cost by two for the same accuracy of inference, in some cases even by one hundred. This study provides the first assessment of the impact of such resampling techniques on parameter inference using sequential importance sampling, and extends the range of situations where likelihood inferences can be easily performed.
Atkins, T J; Duck, F A; Tooley, M A [Department of Medical Physics and Bioengineering, Royal United Hospital, Combe Park, Bath BA1 3NG (United Kingdom); Humphrey, V F, E-mail: timothy.atkins@nhs.net [Institute of Sound and Vibration Research, University of Southampton, Southampton SO17 1BJ (United Kingdom)
2011-02-01
The response of two coaxially aligned weakly focused ultrasonic transducers, typical of those employed for measuring the attenuation of small samples using the immersion method, has been investigated. The effects of the sample size on transmission measurements have been analyzed by integrating the sound pressure distribution functions of the radiator and receiver over different limits to determine the size of the region that contributes to the system response. The results enable the errors introduced into measurements of attenuation to be estimated as a function of sample size. A theoretical expression has been used to examine how the transducer separation affects the receiver output. The calculations are compared with an experimental study of the axial response of three unpaired transducers in water. The separation of each transducer pair giving the maximum response was determined, and compared with the field characteristics of the individual transducers. The optimum transducer separation, for accurate estimation of sample properties, was found to fall between the sum of the focal distances and the sum of the geometric focal lengths as this reduced diffraction errors.
The role of the upper sample size limit in two-stage bioequivalence designs.
Karalis, Vangelis
2013-11-01
Two-stage designs (TSDs) are currently recommended by the regulatory authorities for bioequivalence (BE) assessment. The TSDs presented until now rely on an assumed geometric mean ratio (GMR) value of the BE metric in stage I in order to avoid inflation of type I error. In contrast, this work proposes a more realistic TSD design where sample re-estimation relies not only on the variability of stage I, but also on the observed GMR. In these cases, an upper sample size limit (UL) is introduced in order to prevent inflation of type I error. The aim of this study is to unveil the impact of UL on two TSD bioequivalence approaches which are based entirely on the interim results. Monte Carlo simulations were used to investigate several different scenarios of UL levels, within-subject variability, different starting number of subjects, and GMR. The use of UL leads to no inflation of type I error. As UL values increase, the % probability of declaring BE becomes higher. The starting sample size and the variability of the study affect type I error. Increased UL levels result in higher total sample sizes of the TSD which are more pronounced for highly variable drugs.
Effect of sample size on the fluid flow through a single fractured granitoid
Kunal Kumar Singh; Devendra Narain Singh; Ranjith Pathegama Gamage
2016-01-01
Most of deep geological engineered structures, such as rock caverns, nuclear waste disposal repositories, metro rail tunnels, multi-layer underground parking, are constructed within hard crystalline rocks because of their high quality and low matrix permeability. In such rocks, fluid flows mainly through fractures. Quantification of fractures along with the behavior of the fluid flow through them, at different scales, becomes quite important. Earlier studies have revealed the influence of sample size on the confining stressepermeability relationship and it has been demonstrated that permeability of the fractured rock mass decreases with an increase in sample size. However, most of the researchers have employed numerical simulations to model fluid flow through the fracture/fracture network, or laboratory investigations on intact rock samples with diameter ranging between 38 mm and 45 cm and the diameter-to-length ratio of 1:2 using different experimental methods. Also, the confining stress, s3, has been considered to be less than 30 MPa and the effect of fracture roughness has been ignored. In the present study, an extension of the previous studies on “laboratory simulation of flow through single fractured granite” was conducted, in which consistent fluid flow experiments were performed on cy-lindrical samples of granitoids of two different sizes (38 mm and 54 mm in diameters), containing a“rough walled single fracture”. These experiments were performed under varied confining pressure (s3 ¼ 5e40 MPa), fluid pressure (fp ? 25 MPa), and fracture roughness. The results indicate that a nonlinear relationship exists between the discharge, Q, and the effective confining pressure, sef ., and Q decreases with an increase in sef .. Also, the effects of sample size and fracture roughness do not persist when sef . ? 20 MPa. It is expected that such a study will be quite useful in correlating and extrapolating the laboratory scale investigations to in-situ scale and
A simple method for estimating genetic diversity in large populations from finite sample sizes
Rajora Om P
2009-12-01
Full Text Available Abstract Background Sample size is one of the critical factors affecting the accuracy of the estimation of population genetic diversity parameters. Small sample sizes often lead to significant errors in determining the allelic richness, which is one of the most important and commonly used estimators of genetic diversity in populations. Correct estimation of allelic richness in natural populations is challenging since they often do not conform to model assumptions. Here, we introduce a simple and robust approach to estimate the genetic diversity in large natural populations based on the empirical data for finite sample sizes. Results We developed a non-linear regression model to infer genetic diversity estimates in large natural populations from finite sample sizes. The allelic richness values predicted by our model were in good agreement with those observed in the simulated data sets and the true allelic richness observed in the source populations. The model has been validated using simulated population genetic data sets with different evolutionary scenarios implied in the simulated populations, as well as large microsatellite and allozyme experimental data sets for four conifer species with contrasting patterns of inherent genetic diversity and mating systems. Our model was a better predictor for allelic richness in natural populations than the widely-used Ewens sampling formula, coalescent approach, and rarefaction algorithm. Conclusions Our regression model was capable of accurately estimating allelic richness in natural populations regardless of the species and marker system. This regression modeling approach is free from assumptions and can be widely used for population genetic and conservation applications.
Limitations of mRNA amplification from small-size cell samples
Myklebost Ola
2005-10-01
Full Text Available Abstract Background Global mRNA amplification has become a widely used approach to obtain gene expression profiles from limited material. An important concern is the reliable reflection of the starting material in the results obtained. This is especially important with extremely low quantities of input RNA where stochastic effects due to template dilution may be present. This aspect remains under-documented in the literature, as quantitative measures of data reliability are most often lacking. To address this issue, we examined the sensitivity levels of each transcript in 3 different cell sample sizes. ANOVA analysis was used to estimate the overall effects of reduced input RNA in our experimental design. In order to estimate the validity of decreasing sample sizes, we examined the sensitivity levels of each transcript by applying a novel model-based method, TransCount. Results From expression data, TransCount provided estimates of absolute transcript concentrations in each examined sample. The results from TransCount were used to calculate the Pearson correlation coefficient between transcript concentrations for different sample sizes. The correlations were clearly transcript copy number dependent. A critical level was observed where stochastic fluctuations became significant. The analysis allowed us to pinpoint the gene specific number of transcript templates that defined the limit of reliability with respect to number of cells from that particular source. In the sample amplifying from 1000 cells, transcripts expressed with at least 121 transcripts/cell were statistically reliable and for 250 cells, the limit was 1806 transcripts/cell. Above these thresholds, correlation between our data sets was at acceptable values for reliable interpretation. Conclusion These results imply that the reliability of any amplification experiment must be validated empirically to justify that any gene exists in sufficient quantity in the input material. This
Sample size for estimating the mean concentration of organisms in ballast water.
Costa, Eliardo G; Lopes, Rubens M; Singer, Julio M
2016-09-15
We consider the computation of sample sizes for estimating the mean concentration of organisms in ballast water. Given the possible heterogeneity of their distribution in the tank, we adopt a negative binomial model to obtain confidence intervals for the mean concentration. We show that the results obtained by Chen and Chen (2012) in a different set-up hold for the proposed model and use them to develop algorithms to compute sample sizes both in cases where the mean concentration is known to lie in some bounded interval or where there is no information about its range. We also construct simple diagrams that may be easily employed to decide for compliance with the D-2 regulation of the International Maritime Organization (IMO). Copyright © 2016 Elsevier Ltd. All rights reserved.
Shao, Quanxi; Wang, You-Gan
2009-09-01
Power calculation and sample size determination are critical in designing environmental monitoring programs. The traditional approach based on comparing the mean values may become statistically inappropriate and even invalid when substantial proportions of the response values are below the detection limits or censored because strong distributional assumptions have to be made on the censored observations when implementing the traditional procedures. In this paper, we propose a quantile methodology that is robust to outliers and can also handle data with a substantial proportion of below-detection-limit observations without the need of imputing the censored values. As a demonstration, we applied the methods to a nutrient monitoring project, which is a part of the Perth Long-Term Ocean Outlet Monitoring Program. In this example, the sample size required by our quantile methodology is, in fact, smaller than that by the traditional t-test, illustrating the merit of our method.
Hauschke, D; Steinijans, W V; Diletti, E; Schall, R; Luus, H G; Elze, M; Blume, H
1994-07-01
Bioequivalence studies are generally performed as crossover studies and, therefore, information on the intrasubject coefficient of variation is needed for sample size planning. Unfortunately, this information is usually not presented in publications on bioequivalence studies, and only the pooled inter- and intrasubject coefficient of variation for either test or reference formulation is reported. Thus, the essential information for sample size planning of future studies is not made available to other researchers. In order to overcome such shortcomings, the presentation of results from bioequivalence studies should routinely include the intrasubject coefficient of variation. For the relevant coefficients of variation, theoretical background together with modes of calculation and presentation are given in this communication with particular emphasis on the multiplicative model.
Sample sizing of biological materials analyzed by energy dispersion X-ray fluorescence
Paiva, Jose D.S.; Franca, Elvis J.; Magalhaes, Marcelo R.L.; Almeida, Marcio E.S.; Hazin, Clovis A., E-mail: dan-paiva@hotmail.com, E-mail: ejfranca@cnen.gov.br, E-mail: marcelo_rlm@hotmail.com, E-mail: maensoal@yahoo.com.br, E-mail: chazin@cnen.gov.b [Centro Regional de Ciencias Nucleares do Nordeste (CRCN-NE/CNEN-PE), Recife, PE (Brazil)
2013-07-01
Analytical portions used in chemical analyses are usually less than 1g. Errors resulting from the sampling are barely evaluated, since this type of study is a time-consuming procedure, with high costs for the chemical analysis of large number of samples. The energy dispersion X-ray fluorescence - EDXRF is a non-destructive and fast analytical technique with the possibility of determining several chemical elements. Therefore, the aim of this study was to provide information on the minimum analytical portion for quantification of chemical elements in biological matrices using EDXRF. Three species were sampled in mangroves from the Pernambuco, Brazil. Tree leaves were washed with distilled water, oven-dried at 60 deg C and milled until 0.5 mm particle size. Ten test-portions of approximately 500 mg for each species were transferred to vials sealed with polypropylene film. The quality of the analytical procedure was evaluated from the reference materials IAEA V10 Hay Powder, SRM 2976 Apple Leaves. After energy calibration, all samples were analyzed under vacuum for 100 seconds for each group of chemical elements. The voltage used was 15 kV and 50 kV for chemical elements of atomic number lower than 22 and the others, respectively. For the best analytical conditions, EDXRF was capable of estimating the sample size uncertainty for further determination of chemical elements in leaves. (author)
A Complete Sample of Megaparsec Size Double Radio Sources from SUMSS
Saripalli, L; Subramanian, R; Boyce, E
2005-01-01
We present a complete sample of megaparsec-size double radio sources compiled from the Sydney University Molonglo Sky Survey (SUMSS). Almost complete redshift information has been obtained for the sample. The sample has the following defining criteria: Galactic latitude |b| > 12.5 deg, declination 5 arcmin. All the sources have projected linear size larger than 0.7 Mpc (assuming H_o = 71 km/s/Mpc). The sample is chosen from a region of the sky covering 2100 square degrees. In this paper, we present 843-MHz radio images of the extended radio morphologies made using the Molonglo Observatory Synthesis Telescope (MOST), higher resolution radio observations of any compact radio structures using the Australia Telescope Compact Array (ATCA), and low resolution optical spectra of the host galaxies from the 2.3-m Australian National University (ANU) telescope at Siding Spring Observatory. The sample presented here is the first in the southern hemisphere and significantly enhances the database of known giant radio sou...
Vanessa Colombo-Corbi; Maria José Dellamano-Oliveira; Armando Augusto Henriques Vieira
2011-01-01
Glycolytic activities of eight enzymes in size-fractionated water samples from a eutrophic tropical reservoir are presented in this study, including enzymes assayed for the first time in a freshwater environment. Among these enzymes, rhamnosidase, arabinosidase and fucosidase presented high activity in the free-living fraction, while glucosidase, mannosidase and galactosidase exhibited high activity in the attached fraction. The low activity registered for rhamnosidase, arabinosidase and fuco...
Comparing spectral densities of stationary time series with unequal sample sizes
Hildebrandt, Thimo; Preuß, Philip
2012-01-01
This paper deals with the comparison of several stationary processes with unequal sample sizes. We provide a detailed theoretical framework on the testing problem for equality of spectral densities in the bivariate case, after which the generalization of our approach to the m dimensional case and to other statistical applications (like testing for zero correlation or clustering of time series data with different length) is straightforward. We prove asymptotic normality of an appropriately sta...
A contemporary decennial global Landsat sample of changing agricultural field sizes
White, Emma; Roy, David
2014-05-01
Agriculture has caused significant human induced Land Cover Land Use (LCLU) change, with dramatic cropland expansion in the last century and significant increases in productivity over the past few decades. Satellite data have been used for agricultural applications including cropland distribution mapping, crop condition monitoring, crop production assessment and yield prediction. Satellite based agricultural applications are less reliable when the sensor spatial resolution is small relative to the field size. However, to date, studies of agricultural field size distributions and their change have been limited, even though this information is needed to inform the design of agricultural satellite monitoring systems. Moreover, the size of agricultural fields is a fundamental description of rural landscapes and provides an insight into the drivers of rural LCLU change. In many parts of the world field sizes may have increased. Increasing field sizes cause a subsequent decrease in the number of fields and therefore decreased landscape spatial complexity with impacts on biodiversity, habitat, soil erosion, plant-pollinator interactions, and impacts on the diffusion of herbicides, pesticides, disease pathogens, and pests. The Landsat series of satellites provide the longest record of global land observations, with 30m observations available since 1982. Landsat data are used to examine contemporary field size changes in a period (1980 to 2010) when significant global agricultural changes have occurred. A multi-scale sampling approach is used to locate global hotspots of field size change by examination of a recent global agricultural yield map and literature review. Nine hotspots are selected where significant field size change is apparent and where change has been driven by technological advancements (Argentina and U.S.), abrupt societal changes (Albania and Zimbabwe), government land use and agricultural policy changes (China, Malaysia, Brazil), and/or constrained by
Jha, Anjani K.
Particulate materials are routinely handled in large quantities by industries such as, agriculture, electronic, ceramic, chemical, cosmetic, fertilizer, food, nutraceutical, pharmaceutical, power, and powder metallurgy. These industries encounter segregation due to the difference in physical and mechanical properties of particulates. The general goal of this research was to study percolation segregation in multi-size and multi-component particulate mixtures, especially measurement, sampling, and modeling. A second generation primary segregation shear cell (PSSC-II), an industrial vibrator, a true cubical triaxial tester, and two samplers (triers) were used as primary test apparatuses for quantifying segregation and flowability; furthermore, to understand and propose strategies to mitigate segregation in particulates. Toward this end, percolation segregation in binary, ternary, and quaternary size mixtures for two particulate types: urea (spherical) and potash (angular) were studied. Three coarse size ranges 3,350-4,000 mum (mean size = 3,675 mum), 2,800-3,350 mum (3,075 mum), and 2,360-2,800 mum (2,580 mum) and three fines size ranges 2,000-2,360 mum (2,180 mum), 1,700-2,000 mum (1,850 mum), and 1,400-1,700 mum (1,550 mum) for angular-shaped and spherical-shaped were selected for tests. Since the fines size 1,550 mum of urea was not available in sufficient quantity; therefore, it was not included in tests. Percolation segregation in fertilizer bags was tested also at two vibration frequencies of 5 Hz and 7Hz. The segregation and flowability of binary mixtures of urea under three equilibrium relative humidities (40%, 50%, and 60%) were also tested. Furthermore, solid fertilizer sampling was performed to compare samples obtained from triers of opening widths 12.7 mm and 19.1 mm and to determine size segregation in blend fertilizers. Based on experimental results, the normalized segregation rate (NSR) of binary mixtures was dependent on size ratio, mixing ratio
Shengyu eJiang
2016-02-01
Full Text Available Likert types of rating scales in which a respondent chooses a response from an ordered set of response options are used to measure a wide variety of psychological, educational, and medical outcome variables. The most appropriate item response theory model for analyzing and scoring these instruments when they provide scores on multiple scales is the multidimensional graded response model (MGRM. A simulation study was conducted to investigate the variables that might affect item parameter recovery for the MGRM. Data were generated based on different sample sizes, test lengths, and scale intercorrelations. Parameter estimates were obtained through the flexiMIRT software. The quality of parameter recovery was assessed by the correlation between true and estimated parameters as well as bias and root- mean-square-error. Results indicated that for the vast majority of cases studied a sample size of N = 500 provided accurate parameter estimates, except for tests with 240 items when 1,000 examinees were necessary to obtain accurate parameter estimates. Increasing sample size beyond N = 1,000 did not increase the accuracy of MGRM parameter estimates.
A Web-based Simulator for Sample Size and Power Estimation in Animal Carcinogenicity Studies
Hojin Moon
2002-12-01
Full Text Available A Web-based statistical tool for sample size and power estimation in animal carcinogenicity studies is presented in this paper. It can be used to provide a design with sufficient power for detecting a dose-related trend in the occurrence of a tumor of interest when competing risks are present. The tumors of interest typically are occult tumors for which the time to tumor onset is not directly observable. It is applicable to rodent tumorigenicity assays that have either a single terminal sacrifice or multiple (interval sacrifices. The design is achieved by varying sample size per group, number of sacrifices, number of sacrificed animals at each interval, if any, and scheduled time points for sacrifice. Monte Carlo simulation is carried out in this tool to simulate experiments of rodent bioassays because no closed-form solution is available. It takes design parameters for sample size and power estimation as inputs through the World Wide Web. The core program is written in C and executed in the background. It communicates with the Web front end via a Component Object Model interface passing an Extensible Markup Language string. The proposed statistical tool is illustrated with an animal study in lung cancer prevention research.
Estimating the Size of a Large Network and its Communities from a Random Sample
Chen, Lin; Crawford, Forrest W
2016-01-01
Most real-world networks are too large to be measured or studied directly and there is substantial interest in estimating global network properties from smaller sub-samples. One of the most important global properties is the number of vertices/nodes in the network. Estimating the number of vertices in a large network is a major challenge in computer science, epidemiology, demography, and intelligence analysis. In this paper we consider a population random graph G = (V;E) from the stochastic block model (SBM) with K communities/blocks. A sample is obtained by randomly choosing a subset W and letting G(W) be the induced subgraph in G of the vertices in W. In addition to G(W), we observe the total degree of each sampled vertex and its block membership. Given this partial information, we propose an efficient PopULation Size Estimation algorithm, called PULSE, that correctly estimates the size of the whole population as well as the size of each community. To support our theoretical analysis, we perform an exhausti...
SAMPLE SIZE DETERMINATION IN NON-RADOMIZED SURVIVAL STUDIES WITH NON-CENSORED AND CENSORED DATA
S FAGHIHZADEH
2003-06-01
Full Text Available Introduction: In survival analysis, determination of sufficient sample size to achieve suitable statistical power is important .In both parametric and non-parametric methods of classic statistics, randomn selection of samples is a basic condition. practically, in most clinical trials and health surveys randomn allocation is impossible. Fixed - effect multiple linear regression analysis covers this need and this feature could be extended to survival regression analysis. This paper is the result of sample size determination in non-randomnized surval analysis with censored and non -censored data. Methods: In non-randomnized survival studies, linear regression with fixed -effect variable could be used. In fact such a regression is conditional expectation of dependent variable, conditioned on independent variable. Likelihood fuction with exponential hazard constructed by considering binary variable for allocation of each subject to one of two comparing groups, stating the variance of coefficient of fixed - effect independent variable by determination coefficient , sample size determination formulas are obtained with both censored and non-cencored data. So estimation of sample size is not based on the relation of a single independent variable but it could be attain the required power for a test adjusted for effect of the other explanatory covariates. Since the asymptotic distribution of the likelihood estimator of parameter is normal, we obtained the variance of the regression coefficient estimator formula then by stating the variance of regression coefficient of fixed-effect variable, by determination coefficient we derived formulas for determination of sample size in both censored and non-censored data. Results: In no-randomnized survival analysis ,to compare hazard rates of two groups without censored data, we obtained an estimation of determination coefficient ,risk ratio and proportion of membership to each group and their variances from
Williams Test Required Sample Size For Determining The Minimum Effective Dose
Mustafa Agah TEKINDAL
2016-04-01
Full Text Available Objective: The biological activity of a substance may be explored through a series of experiments on increased or decreased doses of such substance. One of the purposes in studies of this sort is the determination of minimum effective dose. Use of appropriate sample size has an indisputable effect on the reliability of the decisions made in studies made for this purpose. This study attempts to provide a summary of sample sizes, in different scenarios, needed by researchers during the use of Williams test by taking into consideration the number of groups in dose-response studies as well as minimal clinically significant difference, standard deviation, and the test’s power through asymptotic power analyses. Material and Methods: When Type I error was taken as 0.05, scenarios were determined in different sample sizes for each group (5 to 100 with an increase of 5 at a time and different numbers of groups (from 3 to 10, with an increase of 1 at a time. Minimal clinically significant difference refers to the difference between the control group and the experimental group. In this instance, when the control group is zero and takes a specific average value, it refers to the difference from the experimental group. In the resent study, such differences are defined from 1 to 10 with an increase of 1 at a time. For the test’s power would change when the standard deviation changed, the relevant value was changed in all scenarios from 1 to 10 with an increase of 1 at a time to explore the test’s power. Dose-response distributions are skew. In the present study, data were derived from the Poisson distribution with λ= 1 parameter that was determined in accordance with dose-response curves. Results: When changes occurring in the determined scenarios are considered, it can be said, in general, that the significant difference must be set between 1 and 3; and standard deviation must be set between 1 and 2. Conclusion: It is certain that change in the number
Paper coatings with multi-scale roughness evaluated at different sampling sizes
Samyn, Pieter, E-mail: Pieter.Samyn@UGent.be [Ghent University - Department of Textiles, Technologiepark 907, B-9052 Zwijnaarde (Belgium); Van Erps, Juergen; Thienpont, Hugo [Vrije Universiteit Brussels - Department of Applied Physics and Photonics, Pleinlaan 2, B-1050 Brussels (Belgium); Schoukens, Gustaaf [Ghent University - Department of Textiles, Technologiepark 907, B-9052 Zwijnaarde (Belgium)
2011-04-15
Papers have a complex hierarchical structure and the end-user functionalities such as hydrophobicity are controlled by a finishing layer. The application of an organic nanoparticle coating and drying of the aqueous dispersion results in an unique surface morphology with microscale domains that are internally patterned with nanoparticles. Better understanding of the multi-scale surface roughness patterns is obtained by monitoring the topography with non-contact profilometry (NCP) and atomic force microscopy (AFM) at different sampling areas ranging from 2000 {mu}m x 2000 {mu}m to 0.5 {mu}m x 0.5 {mu}m. The statistical roughness parameters are uniquely related to each other over the different measuring techniques and sampling sizes, as they are purely statistically determined. However, they cannot be directly extrapolated over the different sampling areas as they represent transitions at the nano-, micro-to-nano and microscale level. Therefore, the spatial roughness parameters including the correlation length and the specific frequency bandwidth should be taken into account for each measurement, which both allow for direct correlation of roughness data at different sampling sizes.
Lechner, Isabel; Barboza, Perry; Collins, William; Fritz, Julia; Günther, Detlef; Hattendorf, Bodo; Hummel, Jürgen; Südekum, Karl-Heinz; Clauss, Marcus
2010-02-01
Ruminant species differ in the degree that their rumen contents are stratified but are similar insofar that only very fine particles are passed from the forestomach to the lower digestive tract. We investigated the passage kinetics of fluid and particle markers (2, 10 and 20 mm) in fistulated cattle (Bos primigenius f. taurus), muskoxen (Ovibos moschatus), reindeer (Rangifer tarandus) and moose (Alces alces) on different diets. The distribution of dry matter in the rumen and the viscosity of rumen fluids suggested that the rumen contents were more stratified in muskoxen than moose. Correspondingly, as in previous studies, the species differed in the ratio of mean retention times of small particles to fluids in the reticulorumen, which was highest in cattle (2.03) and muskoxen (1.97-1.98), intermediate in reindeer (1.70) and lowest in moose (0.98-1.29). However, the ratio of large to small particle retention did not differ between the species, indicating similarity in the efficiency of the particle sorting mechanism. Passage kinetics of the two largest particle classes did not differ, indicating that particle retention is not a continuous function of particle size but rather threshold-dependent. Overall, the results suggest that fluid flow through the forestomach differs between ruminant species. A lower relative fluid passage, such as in moose, might limit species to a browse-based dietary niche, whereas a higher relative fluid passage broadens the dietary niche options and facilitates the inclusion of, or specialization on, grass. The function of fluid flow in the ruminant forestomach should be further investigated.
Subspace Leakage Analysis and Improved DOA Estimation With Small Sample Size
Shaghaghi, Mahdi; Vorobyov, Sergiy A.
2015-06-01
Classical methods of DOA estimation such as the MUSIC algorithm are based on estimating the signal and noise subspaces from the sample covariance matrix. For a small number of samples, such methods are exposed to performance breakdown, as the sample covariance matrix can largely deviate from the true covariance matrix. In this paper, the problem of DOA estimation performance breakdown is investigated. We consider the structure of the sample covariance matrix and the dynamics of the root-MUSIC algorithm. The performance breakdown in the threshold region is associated with the subspace leakage where some portion of the true signal subspace resides in the estimated noise subspace. In this paper, the subspace leakage is theoretically derived. We also propose a two-step method which improves the performance by modifying the sample covariance matrix such that the amount of the subspace leakage is reduced. Furthermore, we introduce a phenomenon named as root-swap which occurs in the root-MUSIC algorithm in the low sample size region and degrades the performance of the DOA estimation. A new method is then proposed to alleviate this problem. Numerical examples and simulation results are given for uncorrelated and correlated sources to illustrate the improvement achieved by the proposed methods. Moreover, the proposed algorithms are combined with the pseudo-noise resampling method to further improve the performance.
Rosenthal, Mariana; Anderson, Katey; Tengelsen, Leslie; Carter, Kris; Hahn, Christine; Ball, Christopher
2017-08-24
The Right Size Roadmap was developed by the Association of Public Health Laboratories and the Centers for Disease Control and Prevention to improve influenza virologic surveillance efficiency. Guidelines were provided to state health departments regarding representativeness and statistical estimates of specimen numbers needed for seasonal influenza situational awareness, rare or novel influenza virus detection, and rare or novel influenza virus investigation. The aim of this study was to compare Roadmap sampling recommendations with Idaho's influenza virologic surveillance to determine implementation feasibility. We calculated the proportion of medically attended influenza-like illness (MA-ILI) from Idaho's influenza-like illness surveillance among outpatients during October 2008 to May 2014, applied data to Roadmap-provided sample size calculators, and compared calculations with actual numbers of specimens tested for influenza by the Idaho Bureau of Laboratories (IBL). We assessed representativeness among patients' tested specimens to census estimates by age, sex, and health district residence. Among outpatients surveilled, Idaho's mean annual proportion of MA-ILI was 2.30% (20,834/905,818) during a 5-year period. Thus, according to Roadmap recommendations, Idaho needs to collect 128 specimens from MA-ILI patients/week for situational awareness, 1496 influenza-positive specimens/week for detection of a rare or novel influenza virus at 0.2% prevalence, and after detection, 478 specimens/week to confirm true prevalence is ≤2% of influenza-positive samples. The mean number of respiratory specimens Idaho tested for influenza/week, excluding the 2009-2010 influenza season, ranged from 6 to 24. Various influenza virus types and subtypes were collected and specimen submission sources were representative in terms of geographic distribution, patient age range and sex, and disease severity. Insufficient numbers of respiratory specimens are submitted to IBL for influenza
Reducing sample size in experiments with animals: historical controls and related strategies.
Kramer, Matthew; Font, Enrique
2017-02-01
Reducing the number of animal subjects used in biomedical experiments is desirable for ethical and practical reasons. Previous reviews of the benefits of reducing sample sizes have focused on improving experimental designs and methods of statistical analysis, but reducing the size of control groups has been considered rarely. We discuss how the number of current control animals can be reduced, without loss of statistical power, by incorporating information from historical controls, i.e. subjects used as controls in similar previous experiments. Using example data from published reports, we describe how to incorporate information from historical controls under a range of assumptions that might be made in biomedical experiments. Assuming more similarities between historical and current controls yields higher savings and allows the use of smaller current control groups. We conducted simulations, based on typical designs and sample sizes, to quantify how different assumptions about historical controls affect the power of statistical tests. We show that, under our simulation conditions, the number of current control subjects can be reduced by more than half by including historical controls in the analyses. In other experimental scenarios, control groups may be unnecessary. Paying attention to both the function and to the statistical requirements of control groups would result in reducing the total number of animals used in experiments, saving time, effort and money, and bringing research with animals within ethically acceptable bounds. © 2015 Cambridge Philosophical Society.
Effect of sample size on the fluid flow through a single fractured granitoid
Kunal Kumar Singh
2016-06-01
Full Text Available Most of deep geological engineered structures, such as rock caverns, nuclear waste disposal repositories, metro rail tunnels, multi-layer underground parking, are constructed within hard crystalline rocks because of their high quality and low matrix permeability. In such rocks, fluid flows mainly through fractures. Quantification of fractures along with the behavior of the fluid flow through them, at different scales, becomes quite important. Earlier studies have revealed the influence of sample size on the confining stress–permeability relationship and it has been demonstrated that permeability of the fractured rock mass decreases with an increase in sample size. However, most of the researchers have employed numerical simulations to model fluid flow through the fracture/fracture network, or laboratory investigations on intact rock samples with diameter ranging between 38 mm and 45 cm and the diameter-to-length ratio of 1:2 using different experimental methods. Also, the confining stress, σ3, has been considered to be less than 30 MPa and the effect of fracture roughness has been ignored. In the present study, an extension of the previous studies on “laboratory simulation of flow through single fractured granite” was conducted, in which consistent fluid flow experiments were performed on cylindrical samples of granitoids of two different sizes (38 mm and 54 mm in diameters, containing a “rough walled single fracture”. These experiments were performed under varied confining pressure (σ3 = 5–40 MPa, fluid pressure (fp ≤ 25 MPa, and fracture roughness. The results indicate that a nonlinear relationship exists between the discharge, Q, and the effective confining pressure, σeff., and Q decreases with an increase in σeff.. Also, the effects of sample size and fracture roughness do not persist when σeff. ≥ 20 MPa. It is expected that such a study will be quite useful in correlating and extrapolating the laboratory
Adjustable virtual pore-size filter for automated sample preparation using acoustic radiation force
Jung, B; Fisher, K; Ness, K; Rose, K; Mariella, R
2008-05-22
We present a rapid and robust size-based separation method for high throughput microfluidic devices using acoustic radiation force. We developed a finite element modeling tool to predict the two-dimensional acoustic radiation force field perpendicular to the flow direction in microfluidic devices. Here we compare the results from this model with experimental parametric studies including variations of the PZT driving frequencies and voltages as well as various particle sizes and compressidensities. These experimental parametric studies also provide insight into the development of an adjustable 'virtual' pore-size filter as well as optimal operating conditions for various microparticle sizes. We demonstrated the separation of Saccharomyces cerevisiae and MS2 bacteriophage using acoustic focusing. The acoustic radiation force did not affect the MS2 viruses, and their concentration profile remained unchanged. With optimized design of our microfluidic flow system we were able to achieve yields of > 90% for the MS2 with > 80% of the S. cerevisiae being removed in this continuous-flow sample preparation device.
Saccenti, Edoardo; Timmerman, Marieke E
2016-08-01
Sample size determination is a fundamental step in the design of experiments. Methods for sample size determination are abundant for univariate analysis methods, but scarce in the multivariate case. Omics data are multivariate in nature and are commonly investigated using multivariate statistical methods, such as principal component analysis (PCA) and partial least-squares discriminant analysis (PLS-DA). No simple approaches to sample size determination exist for PCA and PLS-DA. In this paper we will introduce important concepts and offer strategies for (minimally) required sample size estimation when planning experiments to be analyzed using PCA and/or PLS-DA.
Laczo, Roxanne M; Sackett, Paul R; Bobko, Philip; Cortina, José M
2005-07-01
The authors discuss potential confusion in conducting primary studies and meta-analyses on the basis of differences between groups. First, the authors show that a formula for the sampling error of the standardized mean difference (d) that is based on equal group sample sizes can produce substantially biased results if applied with markedly unequal group sizes. Second, the authors show that the same concerns are present when primary analyses or meta-analyses are conducted with point-biserial correlations, as the point-biserial correlation (r) is a transformation of d. Third, the authors examine the practice of correcting a point-biserial r for unequal sample sizes and note that such correction would also increase the sampling error of the corrected r. Correcting rs for unequal sample sizes, but using the standard formula for sampling error in uncorrected r, can result in bias. The authors offer a set of recommendations for conducting meta-analyses of group differences.
Distance software: design and analysis of distance sampling surveys for estimating population size.
Thomas, Len; Buckland, Stephen T; Rexstad, Eric A; Laake, Jeff L; Strindberg, Samantha; Hedley, Sharon L; Bishop, Jon Rb; Marques, Tiago A; Burnham, Kenneth P
2010-02-01
1.Distance sampling is a widely used technique for estimating the size or density of biological populations. Many distance sampling designs and most analyses use the software Distance.2.We briefly review distance sampling and its assumptions, outline the history, structure and capabilities of Distance, and provide hints on its use.3.Good survey design is a crucial prerequisite for obtaining reliable results. Distance has a survey design engine, with a built-in geographic information system, that allows properties of different proposed designs to be examined via simulation, and survey plans to be generated.4.A first step in analysis of distance sampling data is modelling the probability of detection. Distance contains three increasingly sophisticated analysis engines for this: conventional distance sampling, which models detection probability as a function of distance from the transect and assumes all objects at zero distance are detected; multiple-covariate distance sampling, which allows covariates in addition to distance; and mark-recapture distance sampling, which relaxes the assumption of certain detection at zero distance.5.All three engines allow estimation of density or abundance, stratified if required, with associated measures of precision calculated either analytically or via the bootstrap.6.Advanced analysis topics covered include the use of multipliers to allow analysis of indirect surveys (such as dung or nest surveys), the density surface modelling analysis engine for spatial and habitat modelling, and information about accessing the analysis engines directly from other software.7.Synthesis and applications. Distance sampling is a key method for producing abundance and density estimates in challenging field conditions. The theory underlying the methods continues to expand to cope with realistic estimation situations. In step with theoretical developments, state-of-the-art software that implements these methods is described that makes the methods
Sample-size calculations for multi-group comparison in population pharmacokinetic experiments.
Ogungbenro, Kayode; Aarons, Leon
2010-01-01
This paper describes an approach for calculating sample size for population pharmacokinetic experiments that involve hypothesis testing based on multi-group comparison detecting the difference in parameters between groups under mixed-effects modelling. This approach extends what has been described for generalized linear models and nonlinear population pharmacokinetic models that involve only binary covariates to more complex nonlinear population pharmacokinetic models. The structural nonlinear model is linearized around the random effects to obtain the marginal model and the hypothesis testing involving model parameters is based on Wald's test. This approach provides an efficient and fast method for calculating sample size for hypothesis testing in population pharmacokinetic models. The approach can also handle different design problems such as unequal allocation of subjects to groups and unbalanced sampling times between and within groups. The results obtained following application to a one compartment intravenous bolus dose model that involved three different hypotheses under different scenarios showed good agreement between the power obtained from NONMEM simulations and nominal power.
Efficient adaptive designs with mid-course sample size adjustment in clinical trials
Bartroff, Jay
2011-01-01
Adaptive designs have been proposed for clinical trials in which the nuisance parameters or alternative of interest are unknown or likely to be misspecified before the trial. Whereas most previous works on adaptive designs and mid-course sample size re-estimation have focused on two-stage or group sequential designs in the normal case, we consider here a new approach that involves at most three stages and is developed in the general framework of multiparameter exponential families. Not only does this approach maintain the prescribed type I error probability, but it also provides a simple but asymptotically efficient sequential test whose finite-sample performance, measured in terms of the expected sample size and power functions, is shown to be comparable to the optimal sequential design, determined by dynamic programming, in the simplified normal mean case with known variance and prespecified alternative, and superior to the existing two-stage designs and also to adaptive group sequential designs when the al...
Autoregressive Prediction with Rolling Mechanism for Time Series Forecasting with Small Sample Size
Zhihua Wang
2014-01-01
Full Text Available Reasonable prediction makes significant practical sense to stochastic and unstable time series analysis with small or limited sample size. Motivated by the rolling idea in grey theory and the practical relevance of very short-term forecasting or 1-step-ahead prediction, a novel autoregressive (AR prediction approach with rolling mechanism is proposed. In the modeling procedure, a new developed AR equation, which can be used to model nonstationary time series, is constructed in each prediction step. Meanwhile, the data window, for the next step ahead forecasting, rolls on by adding the most recent derived prediction result while deleting the first value of the former used sample data set. This rolling mechanism is an efficient technique for its advantages of improved forecasting accuracy, applicability in the case of limited and unstable data situations, and requirement of little computational effort. The general performance, influence of sample size, nonlinearity dynamic mechanism, and significance of the observed trends, as well as innovation variance, are illustrated and verified with Monte Carlo simulations. The proposed methodology is then applied to several practical data sets, including multiple building settlement sequences and two economic series.
Enhanced Z-LDA for Small Sample Size Training in Brain-Computer Interface Systems
Dongrui Gao
2015-01-01
Full Text Available Background. Usually the training set of online brain-computer interface (BCI experiment is small. For the small training set, it lacks enough information to deeply train the classifier, resulting in the poor classification performance during online testing. Methods. In this paper, on the basis of Z-LDA, we further calculate the classification probability of Z-LDA and then use it to select the reliable samples from the testing set to enlarge the training set, aiming to mine the additional information from testing set to adjust the biased classification boundary obtained from the small training set. The proposed approach is an extension of previous Z-LDA and is named enhanced Z-LDA (EZ-LDA. Results. We evaluated the classification performance of LDA, Z-LDA, and EZ-LDA on simulation and real BCI datasets with different sizes of training samples, and classification results showed EZ-LDA achieved the best classification performance. Conclusions. EZ-LDA is promising to deal with the small sample size training problem usually existing in online BCI system.
Stratification in Natural Water Bodies
Møller, Jacob Steen
2004-01-01
Density stratification of natural water bodies plays an important role for a number of civil engineering problems. The origin of stratification in natural water is discussed and the Black Sea, the Gulf of Katchch, and Maarmorilik Fiord in Greenland are described and used as examples. Stratification...... has a number of civil engineering implications. The lock exchange problem is used as a canonical example, and implications for water exchange and sedimentation is discussed by means of examples: Sedimentation in locks and estuaries, salt transport into fresh water reservoirs, water exchange...
Larson, Michael J; Carbine, Kaylie A
2017-01-01
There is increasing focus across scientific fields on adequate sample sizes to ensure non-biased and reproducible effects. Very few studies, however, report sample size calculations or even the information needed to accurately calculate sample sizes for grants and future research. We systematically reviewed 100 randomly selected clinical human electrophysiology studies from six high impact journals that frequently publish electroencephalography (EEG) and event-related potential (ERP) research to determine the proportion of studies that reported sample size calculations, as well as the proportion of studies reporting the necessary components to complete such calculations. Studies were coded by the two authors blinded to the other's results. Inter-rater reliability was 100% for the sample size calculations and kappa above 0.82 for all other variables. Zero of the 100 studies (0%) reported sample size calculations. 77% utilized repeated-measures designs, yet zero studies (0%) reported the necessary variances and correlations among repeated measures to accurately calculate future sample sizes. Most studies (93%) reported study statistical values (e.g., F or t values). Only 40% reported effect sizes, 56% reported mean values, and 47% reported indices of variance (e.g., standard deviations/standard errors). Absence of such information hinders accurate determination of sample sizes for study design, grant applications, and meta-analyses of research and whether studies were adequately powered to detect effects of interest. Increased focus on sample size calculations, utilization of registered reports, and presenting information detailing sample size calculations and statistics for future researchers are needed and will increase sample size-related scientific rigor in human electrophysiology research.
An In Situ Method for Sizing Insoluble Residues in Precipitation and Other Aqueous Samples.
Axson, Jessica L; Creamean, Jessie M; Bondy, Amy L; Capracotta, Sonja S; Warner, Katy Y; Ault, Andrew P
2015-01-01
Particles are frequently incorporated into clouds or precipitation, influencing climate by acting as cloud condensation or ice nuclei, taking up coatings during cloud processing, and removing species through wet deposition. Many of these particles, particularly ice nuclei, can remain suspended within cloud droplets/crystals as insoluble residues. While previous studies have measured the soluble or bulk mass of species within clouds and precipitation, no studies to date have determined the number concentration and size distribution of insoluble residues in precipitation or cloud water using in situ methods. Herein, for the first time we demonstrate that Nanoparticle Tracking Analysis (NTA) is a powerful in situ method for determining the total number concentration, number size distribution, and surface area distribution of insoluble residues in precipitation, both of rain and melted snow. The method uses 500 μL or less of liquid sample and does not require sample modification. Number concentrations for the insoluble residues in aqueous precipitation samples ranged from 2.0-3.0(±0.3)×10(8) particles cm(-3), while surface area ranged from 1.8(±0.7)-3.2(±1.0)×10(7) μm(2) cm(-3). Number size distributions peaked between 133-150 nm, with both single and multi-modal character, while surface area distributions peaked between 173-270 nm. Comparison with electron microscopy of particles up to 10 μm show that, by number, > 97% residues are <1 μm in diameter, the upper limit of the NTA. The range of concentration and distribution properties indicates that insoluble residue properties vary with ambient aerosol concentrations, cloud microphysics, and meteorological dynamics. NTA has great potential for studying the role that insoluble residues play in critical atmospheric processes.
A Bayesian cost-benefit approach to the determination of sample size in clinical trials.
Kikuchi, Takashi; Pezeshk, Hamid; Gittins, John
2008-01-15
Current practice for sample size computations in clinical trials is largely based on frequentist or classical methods. These methods have the drawback of requiring a point estimate of the variance of the treatment effect and are based on arbitrary settings of type I and II errors. They also do not directly address the question of achieving the best balance between the cost of the trial and the possible benefits from using the new treatment, and fail to consider the important fact that the number of users depends on the evidence for improvement compared with the current treatment. Our approach, Behavioural Bayes (or BeBay for short), assumes that the number of patients switching to the new medical treatment depends on the strength of the evidence that is provided by clinical trials, and takes a value between zero and the number of potential patients. The better a new treatment, the more the number of patients who want to switch to it and the more the benefit is obtained. We define the optimal sample size to be the sample size that maximizes the expected net benefit resulting from a clinical trial. Gittins and Pezeshk (Drug Inf. Control 2000; 34:355-363; The Statistician 2000; 49(2):177-187) used a simple form of benefit function and assumed paired comparisons between two medical treatments and that the variance of the treatment effect is known. We generalize this setting, by introducing a logistic benefit function, and by extending the more usual unpaired case, without assuming the variance to be known.
Terry, Leann; Kelley, Ken
2012-11-01
Composite measures play an important role in psychology and related disciplines. Composite measures almost always have error. Correspondingly, it is important to understand the reliability of the scores from any particular composite measure. However, the point estimates of the reliability of composite measures are fallible and thus all such point estimates should be accompanied by a confidence interval. When confidence intervals are wide, there is much uncertainty in the population value of the reliability coefficient. Given the importance of reporting confidence intervals for estimates of reliability, coupled with the undesirability of wide confidence intervals, we develop methods that allow researchers to plan sample size in order to obtain narrow confidence intervals for population reliability coefficients. We first discuss composite reliability coefficients and then provide a discussion on confidence interval formation for the corresponding population value. Using the accuracy in parameter estimation approach, we develop two methods to obtain accurate estimates of reliability by planning sample size. The first method provides a way to plan sample size so that the expected confidence interval width for the population reliability coefficient is sufficiently narrow. The second method ensures that the confidence interval width will be sufficiently narrow with some desired degree of assurance (e.g., 99% assurance that the 95% confidence interval for the population reliability coefficient will be less than W units wide). The effectiveness of our methods was verified with Monte Carlo simulation studies. We demonstrate how to easily implement the methods with easy-to-use and freely available software. ©2011 The British Psychological Society.
S. Shahid Shaukat; Toqeer Ahmed Rao; Moazzam A. Khan
2016-01-01
...) on the eigenvalues and eigenvectors resulting from principal component analysis (PCA). For each sample size, 100 bootstrap samples were drawn from environmental data matrix pertaining to water quality variables (p = 22...
Magnetic response and critical current properties of mesoscopic-size YBCO superconducting samples
Lisboa-Filho, P N [UNESP - Universidade Estadual Paulista, Grupo de Materiais Avancados, Departamento de Fisica, Bauru (Brazil); Deimling, C V; Ortiz, W A, E-mail: plisboa@fc.unesp.b [Grupo de Supercondutividade e Magnetismo, Departamento de Fisica, Universidade Federal de Sao Carlos, Sao Carlos (Brazil)
2010-01-15
In this contribution superconducting specimens of YBa{sub 2}Cu{sub 3}O{sub 7-{delta}} were synthesized by a modified polymeric precursor method, yielding a ceramic powder with particles of mesoscopic-size. Samples of this powder were then pressed into pellets and sintered under different conditions. The critical current density was analyzed by isothermal AC-susceptibility measurements as a function of the excitation field, as well as with isothermal DC-magnetization runs at different values of the applied field. Relevant features of the magnetic response could be associated to the microstructure of the specimens and, in particular, to the superconducting intra- and intergranular critical current properties.
Origin of sample size effect: Stochastic dislocation formation in crystalline metals at small scales
Huang, Guan-Rong; Huang, J. C.; Tsai, W. Y.
2016-12-01
In crystalline metals at small scales, the dislocation density will be increased by stochastic events of dislocation network, leading to a universal power law for various material structures. In this work, we develop a model obeyed by a probability distribution of dislocation density to describe the dislocation formation in terms of a chain reaction. The leading order terms of steady-state of probability distribution gives physical and quantitative insight to the scaling exponent n values in the power law of sample size effect. This approach is found to be consistent with experimental n values in a wide range.
Towards Stratification Learning through Homology Inference
Bendich, Paul; Wang, Bei
2010-01-01
A topological approach to stratification learning is developed for point cloud data drawn from a stratified space. Given such data, our objective is to infer which points belong to the same strata. First we define a multi-scale notion of a stratified space, giving a stratification for each radius level. We then use methods derived from kernel and cokernel persistent homology to cluster the data points into different strata, and we prove a result which guarantees the correctness of our clustering, given certain topological conditions; some geometric intuition for these topological conditions is also provided. Our correctness result is then given a probabilistic flavor: we give bounds on the minimum number of sample points required to infer, with probability, which points belong to the same strata. Finally, we give an explicit algorithm for the clustering, prove its correctness, and apply it to some simulated data.
Garamszegi, László Z; Møller, Anders P
2010-11-01
Comparative analyses aim to explain interspecific variation in phenotype among taxa. In this context, phylogenetic approaches are generally applied to control for similarity due to common descent, because such phylogenetic relationships can produce spurious similarity in phenotypes (known as phylogenetic inertia or bias). On the other hand, these analyses largely ignore potential biases due to within-species variation. Phylogenetic comparative studies inherently assume that species-specific means from intraspecific samples of modest sample size are biologically meaningful. However, within-species variation is often significant, because measurement errors, within- and between-individual variation, seasonal fluctuations, and differences among populations can all reduce the repeatability of a trait. Although simulations revealed that low repeatability can increase the type I error in a phylogenetic study, researchers only exercise great care in accounting for similarity in phenotype due to common phylogenetic descent, while problems posed by intraspecific variation are usually neglected. A meta-analysis of 194 comparative analyses all adjusting for similarity due to common phylogenetic descent revealed that only a few studies reported intraspecific repeatabilities, and hardly any considered or partially dealt with errors arising from intraspecific variation. This is intriguing, because the meta-analytic data suggest that the effect of heterogeneous sampling can be as important as phylogenetic bias, and thus they should be equally controlled in comparative studies. We provide recommendations about how to handle such effects of heterogeneous sampling.
Effect of sample aliquot size on the limit of detection and reproducibility of clinical assays.
Chen, Guorong; Kobayashi, Lori; Nazarenko, Irina
2007-11-01
Nucleic acid amplification technologies significantly improved the limit of detection (LOD) for diagnostic assays. The ability of these assays to amplify fewer than 10 target copies of DNA or RNA imposes new requirements on the preparation of clinical samples. We report a statistical method to determine how large of an aliquot is necessary to reproducibly provide a detectable number of cells. We determined the success probability (p) based on aliquot size and sample volume. The binomial distribution, based on p and the concentration of cells in sample, was used to calculate the probability of getting no target objects in an aliquot and to determine the minimum number of objects per aliquot necessary to generate a reproducible clinical assay. The described method was applied to find a minimum aliquot volume required for a set LOD, false-negative rate (FNR), and %CV. For example, to keep FNR FNRs are 47.2% and 44.9%. This probability model is a useful tool to predict the impact of aliquot volume on the LOD and reproducibility of clinical assays. Even for samples for which pathogens are homogeneously distributed, it is theoretically impossible to collect a single pathogen consistently if the concentration of pathogen is below a certain limit.
Bolton tooth size ratio among qatari population sample: An odontometric study
Hashim, Hayder A; AL-Sayed, Najah; AL-Hussain, Hashim
2017-01-01
Objectives: To establish the overall and anterior Bolton ratio among a sample of Qatari population and to investigate whether there is a difference between males and females, as well as to compare the result obtained by Bolton. Materials and Methods: The current study consisted of 100 orthodontic study participants (50 males and 50 females) with different malocclusions and age ranging between 15 and 20 years. An electronic digital caliper was used to measure the mesiodistal tooth width of all maxillary and mandibular permanent teeth except second and third molars. The Student's t-test was used to compare tooth-size ratios between males and females and between the results of the present study and Bolton's result. Results: The anterior and overall ratio in Qatari individuals were 78.6 ± 3.4 and 91.8 ± 3.1, respectively. The tooth size ratios were slightly greater in males than that in females, however, the differences were not statistically significant (P > 0.05). There were no significant differences in the overall ratio between Qatari individuals and Bolton's results (P > 0.05), whereas statistical significant differences were observed in anterior ratio (P = 0.007). Conclusions: Within the limitation of the limitations of the present study, definite conclusion was difficult to establish. Thus, a further study with a large sample in each malocclusion group is required. PMID:28197399
Exact calculation of power and sample size in bioequivalence studies using two one-sided tests.
Shen, Meiyu; Russek-Cohen, Estelle; Slud, Eric V
2015-01-01
The number of subjects in a pharmacokinetic two-period two-treatment crossover bioequivalence study is typically small, most often less than 60. The most common approach to testing for bioequivalence is the two one-sided tests procedure. No explicit mathematical formula for the power function in the context of the two one-sided tests procedure exists in the statistical literature, although the exact power based on Owen's special case of bivariate noncentral t-distribution has been tabulated and graphed. Several approximations have previously been published for the probability of rejection in the two one-sided tests procedure for crossover bioequivalence studies. These approximations and associated sample size formulas are reviewed in this article and compared for various parameter combinations with exact power formulas derived here, which are computed analytically as univariate integrals and which have been validated by Monte Carlo simulations. The exact formulas for power and sample size are shown to improve markedly in realistic parameter settings over the previous approximations.
Sudden cardiac death risk stratification.
Deyell, Marc W; Krahn, Andrew D; Goldberger, Jeffrey J
2015-06-01
Arrhythmic sudden cardiac death (SCD) may be caused by ventricular tachycardia/fibrillation or pulseless electric activity/asystole. Effective risk stratification to identify patients at risk of arrhythmic SCD is essential for targeting our healthcare and research resources to tackle this important public health issue. Although our understanding of SCD because of pulseless electric activity/asystole is growing, the overwhelming majority of research in risk stratification has focused on SCD-ventricular tachycardia/ventricular fibrillation. This review focuses on existing and novel risk stratification tools for SCD-ventricular tachycardia/ventricular fibrillation. For patients with left ventricular dysfunction or myocardial infarction, advances in imaging, measures of cardiac autonomic function, and measures of repolarization have shown considerable promise in refining risk. Yet the majority of SCD-ventricular tachycardia/ventricular fibrillation occurs in patients without known cardiac disease. Biomarkers and novel imaging techniques may provide further risk stratification in the general population beyond traditional risk stratification for coronary artery disease alone. Despite these advances, significant challenges in risk stratification remain that must be overcome before a meaningful impact on SCD can be realized.
Johnson, David R; Bachan, Lauren K
2013-08-01
In a recent article, Regan, Lakhanpal, and Anguiano (2012) highlighted the lack of evidence for different relationship outcomes between arranged and love-based marriages. Yet the sample size (n = 58) used in the study is insufficient for making such inferences. This reply discusses and demonstrates how small sample sizes reduce the utility of this research.
Sample size requirements and analysis of tag recoveries for paired releases of lake trout
Elrod, Joseph H.; Frank, Anthony
1990-01-01
A simple chi-square test can be used to analyze recoveries from a paired-release experiment to determine whether differential survival occurs between two groups of fish. The sample size required for analysis is a function of (1) the proportion of fish stocked, (2) the expected proportion at recovery, (3) the level of significance (a) at which the null hypothesis is tested, and (4) the power (1-I?) of the statistical test. Detection of a 20% change from a stocking ratio of 50:50 requires a sample of 172 (I?=0.10; 1-I?=0.80) to 459 (I?=0.01; 1-I?=0.95) fish. Pooling samples from replicate pairs is sometimes an appropriate way to increase statistical precision without increasing numbers stocked or sampling intensity. Summing over time is appropriate if catchability or survival of the two groups of fish does not change relative to each other through time. Twelve pairs of identical groups of yearling lake trout Salvelinus namaycush were marked with coded wire tags and stocked into Lake Ontario. Recoveries of fish at ages 2-8 showed differences of 1-14% from the initial stocking ratios. Mean tag recovery rates were 0.217%, 0.156%, 0.128%, 0.121%, 0.093%, 0.042%, and 0.016% for ages 2-8, respectively. At these rates, stocking 12,100-29,700 fish per group would yield samples of 172-459 fish at ages 2-8 combined.
Heo, Yongju; Park, Jiyeon; Lim, Sung-Il; Hur, Hor-Gil; Kim, Daesung; Park, Kihong
2010-08-01
Size-resolved bacterial concentrations in atmospheric aerosols sampled by using a six stage viable impactor at rice field, sanitary landfill, and waste incinerator sites were determined. Culture-based and Polymerase Chain Reaction (PCR) methods were used to identify the airborne bacteria. The culturable bacteria concentration in total suspended particles (TSP) was found to be the highest (848 Colony Forming Unit (CFU)/m(3)) at the sanitary landfill sampling site, while the rice field sampling site has the lowest (125 CFU/m(3)). The closed landfill would be the main source of the observed bacteria concentration at the sanitary landfill. The rice field sampling site was fully covered by rice grain with wetted conditions before harvest and had no significant contribution to the airborne bacteria concentration. This might occur because the dry conditions favor suspension of soil particles and this area had limited personnel and vehicle flow. The respirable fraction calculated by particles less than 3.3 mum was highest (26%) at the sanitary landfill sampling site followed by waste incinerator (19%) and rice field (10%), which showed a lower level of respiratory fraction compared to previous literature values. We identified 58 species in 23 genera of culturable bacteria, and the Microbacterium, Staphylococcus, and Micrococcus were the most abundant genera at the sanitary landfill, waste incinerator, and rice field sites, respectively. An antibiotic resistant test for the above bacteria (Micrococcus sp., Microbacterium sp., and Staphylococcus sp.) showed that the Staphylococcus sp. had the strongest resistance to both antibiotics (25.0% resistance for 32 microg ml(-1) of Chloramphenicol and 62.5% resistance for 4 microg ml(-1) of Gentamicin).
A cold finger cooling system for the efficient graphitisation of microgram-sized carbon samples
Yang, Bin; Smith, A. M.; Hua, Quan
2013-01-01
At ANSTO, we use the Bosch reaction to convert sample CO2 to graphite for production of our radiocarbon AMS targets. Key to the efficient graphitisation of ultra-small samples are the type of iron catalyst used and the effective trapping of water vapour during the reaction. Here we report a simple liquid nitrogen cooling system that enables us to rapidly adjust the temperature of the cold finger in our laser-heated microfurnace. This has led to an improvement in the graphitisation of microgram-sized carbon samples. This simple system uses modest amounts of liquid nitrogen (typically <200 mL/h during graphitisation) and is compact and reliable. We have used it to produce over 120 AMS targets containing between 5 and 20 μg of carbon, with conversion efficiencies for 5 μg targets ranging from 80% to 100%. In addition, this cooling system has been adapted for use with our conventional graphitisation reactors and has also improved their performance.
Effects of sample size on estimation of rainfall extremes at high temperatures
B. Boessenkool
2017-09-01
Full Text Available High precipitation quantiles tend to rise with temperature, following the so-called Clausius–Clapeyron (CC scaling. It is often reported that the CC-scaling relation breaks down and even reverts for very high temperatures. In our study, we investigate this reversal using observational climate data from 142 stations across Germany. One of the suggested meteorological explanations for the breakdown is limited moisture supply. Here we argue that, instead, it could simply originate from undersampling. As rainfall frequency generally decreases with higher temperatures, rainfall intensities as dictated by CC scaling are less likely to be recorded than for moderate temperatures. Empirical quantiles are conventionally estimated from order statistics via various forms of plotting position formulas. They have in common that their largest representable return period is given by the sample size. In small samples, high quantiles are underestimated accordingly. The small-sample effect is weaker, or disappears completely, when using parametric quantile estimates from a generalized Pareto distribution (GPD fitted with L moments. For those, we obtain quantiles of rainfall intensities that continue to rise with temperature.
Pitchaiah Mandava
Full Text Available OBJECTIVE: Clinical trial outcomes often involve an ordinal scale of subjective functional assessments but the optimal way to quantify results is not clear. In stroke, the most commonly used scale, the modified Rankin Score (mRS, a range of scores ("Shift" is proposed as superior to dichotomization because of greater information transfer. The influence of known uncertainties in mRS assessment has not been quantified. We hypothesized that errors caused by uncertainties could be quantified by applying information theory. Using Shannon's model, we quantified errors of the "Shift" compared to dichotomized outcomes using published distributions of mRS uncertainties and applied this model to clinical trials. METHODS: We identified 35 randomized stroke trials that met inclusion criteria. Each trial's mRS distribution was multiplied with the noise distribution from published mRS inter-rater variability to generate an error percentage for "shift" and dichotomized cut-points. For the SAINT I neuroprotectant trial, considered positive by "shift" mRS while the larger follow-up SAINT II trial was negative, we recalculated sample size required if classification uncertainty was taken into account. RESULTS: Considering the full mRS range, error rate was 26.1%±5.31 (Mean±SD. Error rates were lower for all dichotomizations tested using cut-points (e.g. mRS 1; 6.8%±2.89; overall p<0.001. Taking errors into account, SAINT I would have required 24% more subjects than were randomized. CONCLUSION: We show when uncertainty in assessments is considered, the lowest error rates are with dichotomization. While using the full range of mRS is conceptually appealing, a gain of information is counter-balanced by a decrease in reliability. The resultant errors need to be considered since sample size may otherwise be underestimated. In principle, we have outlined an approach to error estimation for any condition in which there are uncertainties in outcome assessment. We
What about N? A methodological study of sample-size reporting in focus group studies
Glenton Claire
2011-03-01
Full Text Available Abstract Background Focus group studies are increasingly published in health related journals, but we know little about how researchers use this method, particularly how they determine the number of focus groups to conduct. The methodological literature commonly advises researchers to follow principles of data saturation, although practical advise on how to do this is lacking. Our objectives were firstly, to describe the current status of sample size in focus group studies reported in health journals. Secondly, to assess whether and how researchers explain the number of focus groups they carry out. Methods We searched PubMed for studies that had used focus groups and that had been published in open access journals during 2008, and extracted data on the number of focus groups and on any explanation authors gave for this number. We also did a qualitative assessment of the papers with regard to how number of groups was explained and discussed. Results We identified 220 papers published in 117 journals. In these papers insufficient reporting of sample sizes was common. The number of focus groups conducted varied greatly (mean 8.4, median 5, range 1 to 96. Thirty seven (17% studies attempted to explain the number of groups. Six studies referred to rules of thumb in the literature, three stated that they were unable to organize more groups for practical reasons, while 28 studies stated that they had reached a point of saturation. Among those stating that they had reached a point of saturation, several appeared not to have followed principles from grounded theory where data collection and analysis is an iterative process until saturation is reached. Studies with high numbers of focus groups did not offer explanations for number of groups. Too much data as a study weakness was not an issue discussed in any of the reviewed papers. Conclusions Based on these findings we suggest that journals adopt more stringent requirements for focus group method
Zhong, Wei; Koopmeiners, Joseph S; Carlin, Bradley P
2013-11-01
Frequentist sample size determination for binary outcome data in a two-arm clinical trial requires initial guesses of the event probabilities for the two treatments. Misspecification of these event rates may lead to a poor estimate of the necessary sample size. In contrast, the Bayesian approach that considers the treatment effect to be random variable having some distribution may offer a better, more flexible approach. The Bayesian sample size proposed by (Whitehead et al., 2008) for exploratory studies on efficacy justifies the acceptable minimum sample size by a "conclusiveness" condition. In this work, we introduce a new two-stage Bayesian design with sample size reestimation at the interim stage. Our design inherits the properties of good interpretation and easy implementation from Whitehead et al. (2008), generalizes their method to a two-sample setting, and uses a fully Bayesian predictive approach to reduce an overly large initial sample size when necessary. Moreover, our design can be extended to allow patient level covariates via logistic regression, now adjusting sample size within each subgroup based on interim analyses. We illustrate the benefits of our approach with a design in non-Hodgkin lymphoma with a simple binary covariate (patient gender), offering an initial step toward within-trial personalized medicine. Copyright © 2013 Elsevier Inc. All rights reserved.
Lee, Paul H; Tse, Andy C Y
2017-05-01
There are limited data on the quality of reporting of information essential for replication of the calculation as well as the accuracy of the sample size calculation. We examine the current quality of reporting of the sample size calculation in randomized controlled trials (RCTs) published in PubMed and to examine the variation in reporting across study design, study characteristics, and journal impact factor. We also reviewed the targeted sample size reported in trial registries. We reviewed and analyzed all RCTs published in December 2014 with journals indexed in PubMed. The 2014 Impact Factors for the journals were used as proxies for their quality. Of the 451 analyzed papers, 58.1% reported an a priori sample size calculation. Nearly all papers provided the level of significance (97.7%) and desired power (96.6%), and most of the papers reported the minimum clinically important effect size (73.3%). The median (inter-quartile range) of the percentage difference of the reported and calculated sample size calculation was 0.0% (IQR -4.6%;3.0%). The accuracy of the reported sample size was better for studies published in journals that endorsed the CONSORT statement and journals with an impact factor. A total of 98 papers had provided targeted sample size on trial registries and about two-third of these papers (n=62) reported sample size calculation, but only 25 (40.3%) had no discrepancy with the reported number in the trial registries. The reporting of the sample size calculation in RCTs published in PubMed-indexed journals and trial registries were poor. The CONSORT statement should be more widely endorsed. Copyright © 2016 European Federation of Internal Medicine. Published by Elsevier B.V. All rights reserved.
Determination of reference limits: statistical concepts and tools for sample size calculation.
Wellek, Stefan; Lackner, Karl J; Jennen-Steinmetz, Christine; Reinhard, Iris; Hoffmann, Isabell; Blettner, Maria
2014-12-01
Reference limits are estimators for 'extreme' percentiles of the distribution of a quantitative diagnostic marker in the healthy population. In most cases, interest will be in the 90% or 95% reference intervals. The standard parametric method of determining reference limits consists of computing quantities of the form X̅±c·S. The proportion of covered values in the underlying population coincides with the specificity obtained when a measurement value falling outside the corresponding reference region is classified as diagnostically suspect. Nonparametrically, reference limits are estimated by means of so-called order statistics. In both approaches, the precision of the estimate depends on the sample size. We present computational procedures for calculating minimally required numbers of subjects to be enrolled in a reference study. The much more sophisticated concept of reference bands replacing statistical reference intervals in case of age-dependent diagnostic markers is also discussed.
Analyzing insulin samples by size-exclusion chromatography: a column degradation study.
Teska, Brandon M; Kumar, Amit; Carpenter, John F; Wempe, Michael F
2015-04-01
Investigating insulin analogs and probing their intrinsic stability at physiological temperature, we observed significant degradation in the size-exclusion chromatography (SEC) signal over a moderate number of insulin sample injections, which generated concerns about the quality of the separations. Therefore, our research goal was to identify the cause(s) for the observed signal degradation and attempt to mitigate the degradation in order to extend SEC column lifespan. In these studies, we used multiangle light scattering, nuclear magnetic resonance, and gas chromatography-mass spectrometry methods to evaluate column degradation. The results from these studies illustrate: (1) that zinc ions introduced by the insulin product produced the observed column performance issues; and (2) that including ethylenediaminetetraacetic acid, a zinc chelator, in the mobile phase helped to maintain column performance.
Sample Size Dependence of Second Magnetization Peak in Type-II Superconductors
无
2003-01-01
We show that the second magnetization peak (SMP), i. e., an increase in the magnetization hysteresis loop width in type-II superconductors,vanishes for samples smaller than a critical size. We argue that the SMP is not related to the critical current enhancement but can be well explained within a framework of the thermomagnetic flux-jump instability theory, where flux jumps reduce the absolute irreversible magnetization relative to the isothermal critical state value at low enough magnetic fields. The recovering of the isothermal critical state with increasing field leads to the SMP. The low-field SMP takes place in both low-Tc conventional and high-Tc unconventional superconductors. Our results show that the restoration of the isothermal critical state is responsible for the SMP occurrence in both cases.
Christian Damgaard
2011-12-01
Full Text Available Increasingly, the survival rates in experimental ecology are presented using odds ratios or log response ratios, but the use of ratio metrics has a problem when all the individuals have either died or survived in only one replicate. In the empirical ecological literature, the problem often has been ignored or circumvented by different, more or less ad hoc approaches. Here, it is argued that the best summary statistic for communicating ecological results of frequency data in studies with small unbalanced samples may be the mean of the posterior distribution of the survival rate. The developed approach may be particularly useful when effect size indexes, such as odds ratios, are needed to compare frequency data between treatments, sites or studies.
Vollert, Jan; Maier, Christoph; Attal, Nadine; Bennett, David L.H.; Bouhassira, Didier; Enax-Krumova, Elena K.; Finnerup, Nanna B.; Freynhagen, Rainer; Gierthmühlen, Janne; Haanpää, Maija; Hansson, Per; Hüllemann, Philipp; Jensen, Troels S.; Magerl, Walter; Ramirez, Juan D.; Rice, Andrew S.C.; Schuh-Hofer, Sigrid; Segerdahl, Märta; Serra, Jordi; Shillo, Pallai R.; Sindrup, Soeren; Tesfaye, Solomon; Themistocleous, Andreas C.; Tölle, Thomas R.; Treede, Rolf-Detlef; Baron, Ralf
2017-01-01
Abstract In a recent cluster analysis, it has been shown that patients with peripheral neuropathic pain can be grouped into 3 sensory phenotypes based on quantitative sensory testing profiles, which are mainly characterized by either sensory loss, intact sensory function and mild thermal hyperalgesia and/or allodynia, or loss of thermal detection and mild mechanical hyperalgesia and/or allodynia. Here, we present an algorithm for allocation of individual patients to these subgroups. The algorithm is nondeterministic—ie, a patient can be sorted to more than one phenotype—and can separate patients with neuropathic pain from healthy subjects (sensitivity: 78%, specificity: 94%). We evaluated the frequency of each phenotype in a population of patients with painful diabetic polyneuropathy (n = 151), painful peripheral nerve injury (n = 335), and postherpetic neuralgia (n = 97) and propose sample sizes of study populations that need to be screened to reach a subpopulation large enough to conduct a phenotype-stratified study. The most common phenotype in diabetic polyneuropathy was sensory loss (83%), followed by mechanical hyperalgesia (75%) and thermal hyperalgesia (34%, note that percentages are overlapping and not additive). In peripheral nerve injury, frequencies were 37%, 59%, and 50%, and in postherpetic neuralgia, frequencies were 31%, 63%, and 46%. For parallel study design, either the estimated effect size of the treatment needs to be high (>0.7) or only phenotypes that are frequent in the clinical entity under study can realistically be performed. For crossover design, populations under 200 patients screened are sufficient for all phenotypes and clinical entities with a minimum estimated treatment effect size of 0.5. PMID:28595241
Presentation of coefficient of variation for bioequivalence sample-size calculation .
Lee, Yi Lin; Mak, Wen Yao; Looi, Irene; Wong, Jia Woei; Yuen, Kah Hay
2017-07-01
The current study aimed to further contribute information on intrasubject coefficient of variation (CV) from 43 bioequivalence studies conducted by our center. Consistent with Yuen et al. (2001), current work also attempted to evaluate the effect of different parameters (AUC0-t, AUC0-∞, and Cmax) used in the estimation of the study power. Furthermore, we have estimated the number of subjects required for each study by looking at the values of intrasubject CV of AUC0-∞ and have also taken into consideration the minimum sample-size requirement set by the US FDA. A total of 37 immediate-release and 6 extended-release formulations from 28 different active pharmaceutical ingredients (APIs) were evaluated. Out of the total number of studies conducted, 10 studies did not achieve satisfactory statistical power on two or more parameters; 4 studies consistently scored poorly across all three parameters. In general, intrasubject CV values calculated from Cmax were more variable compared to either AUC0-t and AUC0-∞. 20 out of 43 studies did not achieve more than 80% power when the value was calculated from Cmax value, compared to only 11 (AUC0-∞) and 8 (AUC0-t) studies. This finding is consistent with Steinijans et al. (1995) [2] and Yuen et al. (2001) [3]. In conclusion, the CV values obtained from AUC0-t and AUC0-∞ were similar, while those derived from Cmax were consistently more variable. Hence, CV derived from AUC instead of Cmax should be used in sample-size calculation to achieve a sufficient, yet practical, test power. .
Measuring proteins with greater speed and resolution while reducing sample size.
Hsieh, Vincent H; Wyatt, Philip J
2017-08-30
A multi-angle light scattering (MALS) system, combined with chromatographic separation, directly measures the absolute molar mass, size and concentration of the eluate species. The measurement of these crucial properties in solution is essential in basic macromolecular characterization and all research and production stages of bio-therapeutic products. We developed a new MALS methodology that has overcome the long-standing, stubborn barrier to microliter-scale peak volumes and achieved the highest resolution and signal-to-noise performance of any MALS measurement. The novel design simultaneously facilitates online dynamic light scattering (DLS) measurements. As National Institute of Standards and Technology (NIST) new protein standard reference material (SRM 8671) is becoming the benchmark molecule against which many biomolecular analytical techniques are assessed and evaluated, we present its measurement results as a demonstration of the unique capability of our system to swiftly resolve and measure sharp (20~25 µL full-width-half-maximum) chromatography peaks. Precise measurements of protein mass and size can be accomplished 10 times faster than before with improved resolution. In the meantime the sample amount required for such measurements is reduced commensurately. These abilities will have far-reaching impacts at every stage of the development and production of biologics and bio-therapeutic formulations.
BAYESIAN BICLUSTERING FOR PATIENT STRATIFICATION.
Khakabimamaghani, Sahand; Ester, Martin
2016-01-01
The move from Empirical Medicine towards Personalized Medicine has attracted attention to Stratified Medicine (SM). Some methods are provided in the literature for patient stratification, which is the central task of SM, however, there are still significant open issues. First, it is still unclear if integrating different datatypes will help in detecting disease subtypes more accurately, and, if not, which datatype(s) are most useful for this task. Second, it is not clear how we can compare different methods of patient stratification. Third, as most of the proposed stratification methods are deterministic, there is a need for investigating the potential benefits of applying probabilistic methods. To address these issues, we introduce a novel integrative Bayesian biclustering method, called B2PS, for patient stratification and propose methods for evaluating the results. Our experimental results demonstrate the superiority of B2PS over a popular state-of-the-art method and the benefits of Bayesian approaches. Our results agree with the intuition that transcriptomic data forms a better basis for patient stratification than genomic data.
Christen, Hans M [ORNL; Okubo, Isao [ORNL; Rouleau, Christopher M [ORNL; Jellison Jr, Gerald Earle [ORNL; Puretzky, Alexander A [ORNL; Geohegan, David B [ORNL; Lowndes, Douglas H [ORNL
2005-01-01
Parallel (multi-sample) approaches, such as discrete combinatorial synthesis or continuous compositional-spread (CCS), can significantly increase the rate of materials discovery and process optimization. Here we review our generalized CCS method, based on pulsed-laser deposition, in which the synchronization between laser firing and substrate translation (behind a fixed slit aperture) yields the desired variations of composition and thickness. In situ alloying makes this approach applicable to the non-equilibrium synthesis of metastable phases. Deposition on a heater plate with a controlled spatial temperature variation can additionally be used for growth-temperature-dependence studies. Composition and temperature variations are controlled on length scales large enough to yield sample sizes sufficient for conventional characterization techniques (such as temperature-dependent measurements of resistivity or magnetic properties). This technique has been applied to various experimental studies, and we present here the results for the growth of electro-optic materials (Sr{sub x}Ba{sub 1-x}Nb{sub 2}O{sub 6}) and magnetic perovskites (Sr{sub 1-x}Ca{sub x}RuO{sub 3}), and discuss the application to the understanding and optimization of catalysts used in the synthesis of dense forests of carbon nanotubes.
Christen, Hans M.; Ohkubo, Isao; Rouleau, Christopher M.; Jellison, Gerald E., Jr.; Puretzky, Alex A.; Geohegan, David B.; Lowndes, Douglas H.
2005-01-01
Parallel (multi-sample) approaches, such as discrete combinatorial synthesis or continuous compositional-spread (CCS), can significantly increase the rate of materials discovery and process optimization. Here we review our generalized CCS method, based on pulsed-laser deposition, in which the synchronization between laser firing and substrate translation (behind a fixed slit aperture) yields the desired variations of composition and thickness. In situ alloying makes this approach applicable to the non-equilibrium synthesis of metastable phases. Deposition on a heater plate with a controlled spatial temperature variation can additionally be used for growth-temperature-dependence studies. Composition and temperature variations are controlled on length scales large enough to yield sample sizes sufficient for conventional characterization techniques (such as temperature-dependent measurements of resistivity or magnetic properties). This technique has been applied to various experimental studies, and we present here the results for the growth of electro-optic materials (SrxBa1-xNb2O6) and magnetic perovskites (Sr1-xCaxRuO3), and discuss the application to the understanding and optimization of catalysts used in the synthesis of dense forests of carbon nanotubes.
W. Holmes Finch
2016-05-01
Full Text Available Researchers and data analysts are sometimes faced with the problem of very small samples, where the number of variables approaches or exceeds the overall sample size; i.e. high dimensional data. In such cases, standard statistical models such as regression or analysis of variance cannot be used, either because the resulting parameter estimates exhibit very high variance and can therefore not be trusted, or because the statistical algorithm cannot converge on parameter estimates at all. There exist an alternative set of model estimation procedures, known collectively as regularization methods, which can be used in such circumstances, and which have been shown through simulation research to yield accurate parameter estimates. The purpose of this paper is to describe, for those unfamiliar with them, the most popular of these regularization methods, the lasso, and to demonstrate its use on an actual high dimensional dataset involving adults with autism, using the R software language. Results of analyses involving relating measures of executive functioning with a full scale intelligence test score are presented, and implications of using these models are discussed.
Weighted piecewise LDA for solving the small sample size problem in face verification.
Kyperountas, Marios; Tefas, Anastasios; Pitas, Ioannis
2007-03-01
A novel algorithm that can be used to boost the performance of face-verification methods that utilize Fisher's criterion is presented and evaluated. The algorithm is applied to similarity, or matching error, data and provides a general solution for overcoming the "small sample size" (SSS) problem, where the lack of sufficient training samples causes improper estimation of a linear separation hyperplane between the classes. Two independent phases constitute the proposed method. Initially, a set of weighted piecewise discriminant hyperplanes are used in order to provide a more accurate discriminant decision than the one produced by the traditional linear discriminant analysis (LDA) methodology. The expected classification ability of this method is investigated throughout a series of simulations. The second phase defines proper combinations for person-specific similarity scores and describes an outlier removal process that further enhances the classification ability. The proposed technique has been tested on the M2VTS and XM2VTS frontal face databases. Experimental results indicate that the proposed framework greatly improves the face-verification performance.
Żebrowska, Magdalena; Posch, Martin; Magirr, Dominic
2016-05-30
Consider a parallel group trial for the comparison of an experimental treatment to a control, where the second-stage sample size may depend on the blinded primary endpoint data as well as on additional blinded data from a secondary endpoint. For the setting of normally distributed endpoints, we demonstrate that this may lead to an inflation of the type I error rate if the null hypothesis holds for the primary but not the secondary endpoint. We derive upper bounds for the inflation of the type I error rate, both for trials that employ random allocation and for those that use block randomization. We illustrate the worst-case sample size reassessment rule in a case study. For both randomization strategies, the maximum type I error rate increases with the effect size in the secondary endpoint and the correlation between endpoints. The maximum inflation increases with smaller block sizes if information on the block size is used in the reassessment rule. Based on our findings, we do not question the well-established use of blinded sample size reassessment methods with nuisance parameter estimates computed from the blinded interim data of the primary endpoint. However, we demonstrate that the type I error rate control of these methods relies on the application of specific, binding, pre-planned and fully algorithmic sample size reassessment rules and does not extend to general or unplanned sample size adjustments based on blinded data. © 2015 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.
You-xin Shen
Full Text Available A large number of small-sized samples invariably shows that woody species are absent from forest soil seed banks, leading to a large discrepancy with the seedling bank on the forest floor. We ask: 1 Does this conventional sampling strategy limit the detection of seeds of woody species? 2 Are large sample areas and sample sizes needed for higher recovery of seeds of woody species? We collected 100 samples that were 10 cm (length × 10 cm (width × 10 cm (depth, referred to as larger number of small-sized samples (LNSS in a 1 ha forest plot, and placed them to germinate in a greenhouse, and collected 30 samples that were 1 m × 1 m × 10 cm, referred to as small number of large-sized samples (SNLS and placed them (10 each in a nearby secondary forest, shrub land and grass land. Only 15.7% of woody plant species of the forest stand were detected by the 100 LNSS, contrasting with 22.9%, 37.3% and 20.5% woody plant species being detected by SNLS in the secondary forest, shrub land and grassland, respectively. The increased number of species vs. sampled areas confirmed power-law relationships for forest stand, the LNSS and SNLS at all three recipient sites. Our results, although based on one forest, indicate that conventional LNSS did not yield a high percentage of detection for woody species, but SNLS strategy yielded a higher percentage of detection for woody species in the seed bank if samples were exposed to a better field germination environment. A 4 m2 minimum sample area derived from power equations is larger than the sampled area in most studies in the literature. Increased sample size also is needed to obtain an increased sample area if the number of samples is to remain relatively low.
Tsukakoshi, Yoshiki; Yasui, Akemi
2011-11-01
To give a quantitative guide to sample size allocation for developing sampling designs for a food composition survey, we discuss sampling strategies that consider the importance of each food; namely, consumption or production, variability of composition, and the restrictions within the available resources for sample collection and analysis are considered., Here we consider two strategies: 'proportional' and 'Neyman' are discussed. Both of these incorporate consumed quantity of foods, and we review some available statistics for allocation issues. The Neyman optimal strategy allocates less sample size for starch than proportional, because the former incorporates variability in the composition. Those strategies improved accuracy in dietary nutrient intake more than equal sample size allocation. Those strategies will be useful as we often face sample size allocation problems, wherein we decide whether to sample 'five white potatoes and five taros or nine white and one taros'. Allocating sufficient sample size for important foodstuffs is essential in assuring data quality. Nevertheless, the food composition table should be as comprehensive as possible.
Brus, D.J.; Nieuwenhuizen, W.; Koomen, A.J.M.
2006-01-01
Seventy-two squares of 100 ha were selected by stratified random sampling with probabilities proportional to size (pps) to survey landscape changes in the period 1996-2003. The area of the plots times the urbanization pressure was used as a size measure. The central question of this study is whether
Elsa Tavernier
Full Text Available We aimed to examine the extent to which inaccurate assumptions for nuisance parameters used to calculate sample size can affect the power of a randomized controlled trial (RCT. In a simulation study, we separately considered an RCT with continuous, dichotomous or time-to-event outcomes, with associated nuisance parameters of standard deviation, success rate in the control group and survival rate in the control group at some time point, respectively. For each type of outcome, we calculated a required sample size N for a hypothesized treatment effect, an assumed nuisance parameter and a nominal power of 80%. We then assumed a nuisance parameter associated with a relative error at the design stage. For each type of outcome, we randomly drew 10,000 relative errors of the associated nuisance parameter (from empirical distributions derived from a previously published review. Then, retro-fitting the sample size formula, we derived, for the pre-calculated sample size N, the real power of the RCT, taking into account the relative error for the nuisance parameter. In total, 23%, 0% and 18% of RCTs with continuous, binary and time-to-event outcomes, respectively, were underpowered (i.e., the real power was 90%. Even with proper calculation of sample size, a substantial number of trials are underpowered or overpowered because of imprecise knowledge of nuisance parameters. Such findings raise questions about how sample size for RCTs should be determined.
The impact of metrology study sample size on uncertainty in IAEA safeguards calculations
Burr Tom
2016-01-01
Full Text Available Quantitative conclusions by the International Atomic Energy Agency (IAEA regarding States' nuclear material inventories and flows are provided in the form of material balance evaluations (MBEs. MBEs use facility estimates of the material unaccounted for together with verification data to monitor for possible nuclear material diversion. Verification data consist of paired measurements (usually operators' declarations and inspectors' verification results that are analysed one-item-at-a-time to detect significant differences. Also, to check for patterns, an overall difference of the operator-inspector values using a “D (difference statistic” is used. The estimated DP and false alarm probability (FAP depend on the assumed measurement error model and its random and systematic error variances, which are estimated using data from previous inspections (which are used for metrology studies to characterize measurement error variance components. Therefore, the sample sizes in both the previous and current inspections will impact the estimated DP and FAP, as is illustrated by simulated numerical examples. The examples include application of a new expression for the variance of the D statistic assuming the measurement error model is multiplicative and new application of both random and systematic error variances in one-item-at-a-time testing.
Importance of sample size for the estimation of repeater F waves in amyotrophic lateral sclerosis.
Fang, Jia; Liu, Ming-Sheng; Guan, Yu-Zhou; Cui, Bo; Cui, Li-Ying
2015-02-20
In amyotrophic lateral sclerosis (ALS), repeater F waves are increased. Accurate assessment of repeater F waves requires an adequate sample size. We studied the F waves of left ulnar nerves in ALS patients. Based on the presence or absence of pyramidal signs in the left upper limb, the ALS patients were divided into two groups: One group with pyramidal signs designated as P group and the other without pyramidal signs designated as NP group. The Index repeating neurons (RN) and Index repeater F waves (Freps) were compared among the P, NP and control groups following 20 and 100 stimuli respectively. For each group, the Index RN and Index Freps obtained from 20 and 100 stimuli were compared. In the P group, the Index RN (P = 0.004) and Index Freps (P = 0.001) obtained from 100 stimuli were significantly higher than from 20 stimuli. For F waves obtained from 20 stimuli, no significant differences were identified between the P and NP groups for Index RN (P = 0.052) and Index Freps (P = 0.079); The Index RN (P waves obtained from 100 stimuli, the Index RN (P waves reflect increased excitability of motor neuron pool and indicate upper motor neuron dysfunction in ALS. For an accurate evaluation of repeater F waves in ALS patients especially those with moderate to severe muscle atrophy, 100 stimuli would be required.
Data with hierarchical structure: impact of intraclass correlation and sample size on type-I error.
Musca, Serban C; Kamiejski, Rodolphe; Nugier, Armelle; Méot, Alain; Er-Rafiy, Abdelatif; Brauer, Markus
2011-01-01
Least squares analyses (e.g., ANOVAs, linear regressions) of hierarchical data leads to Type-I error rates that depart severely from the nominal Type-I error rate assumed. Thus, when least squares methods are used to analyze hierarchical data coming from designs in which some groups are assigned to the treatment condition, and others to the control condition (i.e., the widely used "groups nested under treatment" experimental design), the Type-I error rate is seriously inflated, leading too often to the incorrect rejection of the null hypothesis (i.e., the incorrect conclusion of an effect of the treatment). To highlight the severity of the problem, we present simulations showing how the Type-I error rate is affected under different conditions of intraclass correlation and sample size. For all simulations the Type-I error rate after application of the popular Kish (1965) correction is also considered, and the limitations of this correction technique discussed. We conclude with suggestions on how one should collect and analyze data bearing a hierarchical structure.
Data with hierarchical structure: impact of intraclass correlation and sample size on Type-I error
Serban C Musca
2011-04-01
Full Text Available Least squares analyses (e.g., ANOVAs, linear regressions of hierarchical data leads to Type-I error rates that depart severely from the nominal Type-I error rate assumed. Thus, when least squares methods are used to analyze hierarchical data coming from designs in which some groups are assigned to the treatment condition, and others to the control condition (i.e., the widely used "groups nested under treatment" experimental design, the Type-I error rate is seriously inflated, leading too often to the incorrect rejection of the null hypothesis (i.e., the incorrect conclusion of an effect of the treatment. To highlight the severity of the problem, we present simulations showing how the Type-I error rate is affected under different conditions of intraclass correlation and sample size. For all simulations the Type-I error rate after application of the popular Kish (1965 correction is also considered, and the limitations of this correction technique discussed. We conclude with suggestions on how one should collect and analyze data bearing a hierarchical structure.
Nomura, S.; Ogata, Y.
2010-12-01
This study is concerned with the probability forecast by the Brownian Passage Time (BPT) model especially in case where only a few records of recurrent earthquakes from an active fault are available. We adopt the Bayesian predictive distribution that takes the relevant prior information and all possibilities for model parameters into account. We utilize the size of single-event displacements U and the slip rate V across the segment to calculate the mean recurrence time T=U/V that the past recurrence intervals are distributed around as Figure 1. We then make use of the best fitted prior distribution for the BPT variation coefficient (the shape parameter, α) selected by the Akaike Bayesian information criterion (ABIC), while the ERC uses the same common estimate α=0.24. Applying this prior distribution, we can see that α takes various values among the faults but has some locational tendencies from Figure 2. For example, α values tend to be higher in the center of Honshu island where the faults are densely populated. We compare the goodness of fit and probability forecasts between the conventional models and our proposed model by historical or simulated datasets. The Bayesian predictor shows very stable and superior performance for small samples or variant recurrence times. Figure 1: The relation between mean recurrence time from slip data and past recurrence intervals with error bars. Figure 2: The map of active faults in land and subduction-zones in Japan, whose colors show the Bayes estimates of variation coefficient α.
Rumen content stratification in the giraffe (Giraffa camelopardalis)
Sauer, Cathrine; Clauss, Marcus; Bertelsen, Mads F;
2016-01-01
Ruminants differ in the degree of rumen content stratification, with 'cattle-types' (i.e., the grazing and intermediate feeding ruminants) having stratified content, whereas 'moose-types' (i.e., the browsing ruminants) have unstratified content. The feeding ecology, as well as the digestive...... of these parameters, indicating homogenous rumen content in the giraffes. In addition to the digesta samples, samples of dorsal rumen, ventral rumen and atrium ruminis mucosa were collected and the papillary surface enlargement factor was determined, as a proxy for content stratification. The even rumen papillation...
Clanton, U. S.; Fletcher, C. R.
1976-01-01
The paper describes a Monte Carlo model for simulation of two-dimensional representations of thin sections of some of the more common igneous rock textures. These representations are extrapolated to three dimensions to develop a volume of 'rock'. The model (here applied to a medium-grained high-Ti basalt) can be used to determine a statistically significant sample for a lunar rock or to predict the probable errors in the oxide contents that can occur during the analysis of a sample that is not representative of the parent rock.
Social Stratification in Higher Education
Grodsky, Eric; Jackson, Erika
2009-01-01
Background/Context: Over the past half century, scholars in a variety of fields have contributed to our understanding of the relationship between higher education and social stratification. We review this literature, highlighting complementarities and inconsistencies. Purpose/Objective/Research Question/Focus of Study: We situate our review of the…
Evaluating methods to correct for population stratification when estimating paternity indexes.
Toscanini, Ulises; Garcia-Magariños, Manuel; Berardi, Gabriela; Egeland, Thore; Raimondi, Eduardo; Salas, Antonio
2012-01-01
The statistical interpretation of the forensic genetic evidence requires the use of allelic frequency estimates in the reference population for the studied markers. Differences in the genetic make up of the populations can be reflected in statistically different allelic frequency distributions. One can easily figure out that collecting such information for any given population is not always possible. Therefore, alternative approaches are needed in these cases in order to compensate for the lack of information. A number of statistics have been proposed to control for population stratification in paternity testing and forensic casework, Fst correction being the only one recommended by the forensic community. In this study we aimed to evaluate the performance of Fst to correct for population stratification in forensics. By way of simulations, we first tested the dependence of Fst on the relative sizes of the sub-populations, and second, we measured the effect of the Fst corrections on the Paternity Index (PI) values compared to the ones obtained when using the local reference database. The results provide clear-cut evidence that (i) Fst values are strongly dependent on the sampling scheme, and therefore, for most situations it would be almost impossible to estimate real values of Fst; and (ii) Fst corrections might unfairly correct PI values for stratification, suggesting the use of local databases whenever possible to estimate the frequencies of genetic profiles and PI values.
Evaluating methods to correct for population stratification when estimating paternity indexes.
Ulises Toscanini
Full Text Available The statistical interpretation of the forensic genetic evidence requires the use of allelic frequency estimates in the reference population for the studied markers. Differences in the genetic make up of the populations can be reflected in statistically different allelic frequency distributions. One can easily figure out that collecting such information for any given population is not always possible. Therefore, alternative approaches are needed in these cases in order to compensate for the lack of information. A number of statistics have been proposed to control for population stratification in paternity testing and forensic casework, Fst correction being the only one recommended by the forensic community. In this study we aimed to evaluate the performance of Fst to correct for population stratification in forensics. By way of simulations, we first tested the dependence of Fst on the relative sizes of the sub-populations, and second, we measured the effect of the Fst corrections on the Paternity Index (PI values compared to the ones obtained when using the local reference database. The results provide clear-cut evidence that (i Fst values are strongly dependent on the sampling scheme, and therefore, for most situations it would be almost impossible to estimate real values of Fst; and (ii Fst corrections might unfairly correct PI values for stratification, suggesting the use of local databases whenever possible to estimate the frequencies of genetic profiles and PI values.
Evaluating Methods to Correct for Population Stratification when Estimating Paternity Indexes
Garcia-Magariños, Manuel; Berardi, Gabriela; Egeland, Thore; Raimondi, Eduardo
2012-01-01
The statistical interpretation of the forensic genetic evidence requires the use of allelic frequency estimates in the reference population for the studied markers. Differences in the genetic make up of the populations can be reflected in statistically different allelic frequency distributions. One can easily figure out that collecting such information for any given population is not always possible. Therefore, alternative approaches are needed in these cases in order to compensate for the lack of information. A number of statistics have been proposed to control for population stratification in paternity testing and forensic casework, Fst correction being the only one recommended by the forensic community. In this study we aimed to evaluate the performance of Fst to correct for population stratification in forensics. By way of simulations, we first tested the dependence of Fst on the relative sizes of the sub-populations, and second, we measured the effect of the Fst corrections on the Paternity Index (PI) values compared to the ones obtained when using the local reference database. The results provide clear-cut evidence that (i) Fst values are strongly dependent on the sampling scheme, and therefore, for most situations it would be almost impossible to estimate real values of Fst; and (ii) Fst corrections might unfairly correct PI values for stratification, suggesting the use of local databases whenever possible to estimate the frequencies of genetic profiles and PI values. PMID:23226224
About 100 countries have established regulatory limits for aflatoxin in food and feeds. Because these limits vary widely among regulating countries, the Codex Committee on Food Additives and Contaminants (CCFAC) began work in 2004 to harmonize aflatoxin limits and sampling plans for aflatoxin in alm...
Tang, Yongqiang
2015-01-01
A sample size formula is derived for negative binomial regression for the analysis of recurrent events, in which subjects can have unequal follow-up time. We obtain sharp lower and upper bounds on the required size, which is easy to compute. The upper bound is generally only slightly larger than the required size, and hence can be used to approximate the sample size. The lower and upper size bounds can be decomposed into two terms. The first term relies on the mean number of events in each group, and the second term depends on two factors that measure, respectively, the extent of between-subject variability in event rates, and follow-up time. Simulation studies are conducted to assess the performance of the proposed method. An application of our formulae to a multiple sclerosis trial is provided.
Gibertini, Michael; Nations, Kari R; Whitaker, John A
2012-03-01
The high failure rate of antidepressant trials has spurred exploration of the factors that affect trial sensitivity. In the current analysis, Food and Drug Administration antidepressant drug registration trial data compiled by Turner et al. is extended to include the most recently approved antidepressants. The expanded dataset is examined to further establish the likely population effect size (ES) for monoaminergic antidepressants and to demonstrate the relationship between observed ES and sample size in trials on compounds with proven efficacy. Results indicate that the overall underlying ES for antidepressants is approximately 0.30, and that the variability in observed ES across trials is related to the sample size of the trial. The current data provide a unique real-world illustration of an often underappreciated statistical truism: that small N trials are more likely to mislead than to inform, and that by aligning sample size to the population ES, risks of both erroneously high and low effects are minimized. The results in the current study make this abstract concept concrete and will help drug developers arrive at informed gate decisions with greater confidence and fewer risks, improving the odds of success for future antidepressant trials.
Kostoulas, P.; Nielsen, Søren Saxmose; Browne, W. J.;
2013-01-01
and power when applied to these groups. We propose the use of the variance partition coefficient (VPC), which measures the clustering of infection/disease for individuals with a common risk profile. Sample size estimates are obtained separately for those groups that exhibit markedly different heterogeneity......SUMMARY Disease cases are often clustered within herds or generally groups that share common characteristics. Sample size formulae must adjust for the within-cluster correlation of the primary sampling units. Traditionally, the intra-cluster correlation coefficient (ICC), which is an average...
Reliable calculation in probabilistic logic: Accounting for small sample size and model uncertainty
Ferson, S. [Applied Biomathematics, Setauket, NY (United States)
1996-12-31
A variety of practical computational problems arise in risk and safety assessments, forensic statistics and decision analyses in which the probability of some event or proposition E is to be estimated from the probabilities of a finite list of related subevents or propositions F,G,H,.... In practice, the analyst`s knowledge may be incomplete in two ways. First, the probabilities of the subevents may be imprecisely known from statistical estimations, perhaps based on very small sample sizes. Second, relationships among the subevents may be known imprecisely. For instance, there may be only limited information about their stochastic dependencies. Representing probability estimates as interval ranges on has been suggested as a way to address the first source of imprecision. A suite of AND, OR and NOT operators defined with reference to the classical Frochet inequalities permit these probability intervals to be used in calculations that address the second source of imprecision, in many cases, in a best possible way. Using statistical confidence intervals as inputs unravels the closure properties of this approach however, requiring that probability estimates be characterized by a nested stack of intervals for all possible levels of statistical confidence, from a point estimate (0% confidence) to the entire unit interval (100% confidence). The corresponding logical operations implied by convolutive application of the logical operators for every possible pair of confidence intervals reduces by symmetry to a manageably simple level-wise iteration. The resulting calculus can be implemented in software that allows users to compute comprehensive and often level-wise best possible bounds on probabilities for logical functions of events.
Importance of Sample Size for the Estimation of Repeater F Waves in Amyotrophic Lateral Sclerosis
Jia Fang
2015-01-01
Full Text Available Background: In amyotrophic lateral sclerosis (ALS, repeater F waves are increased. Accurate assessment of repeater F waves requires an adequate sample size. Methods: We studied the F waves of left ulnar nerves in ALS patients. Based on the presence or absence of pyramidal signs in the left upper limb, the ALS patients were divided into two groups: One group with pyramidal signs designated as P group and the other without pyramidal signs designated as NP group. The Index repeating neurons (RN and Index repeater F waves (Freps were compared among the P, NP and control groups following 20 and 100 stimuli respectively. For each group, the Index RN and Index Freps obtained from 20 and 100 stimuli were compared. Results: In the P group, the Index RN (P = 0.004 and Index Freps (P = 0.001 obtained from 100 stimuli were significantly higher than from 20 stimuli. For F waves obtained from 20 stimuli, no significant differences were identified between the P and NP groups for Index RN (P = 0.052 and Index Freps (P = 0.079; The Index RN (P < 0.001 and Index Freps (P < 0.001 of the P group were significantly higher than the control group; The Index RN (P = 0.002 of the NP group was significantly higher than the control group. For F waves obtained from 100 stimuli, the Index RN (P < 0.001 and Index Freps (P < 0.001 of the P group were significantly higher than the NP group; The Index RN (P < 0.001 and Index Freps (P < 0.001 of the P and NP groups were significantly higher than the control group. Conclusions: Increased repeater F waves reflect increased excitability of motor neuron pool and indicate upper motor neuron dysfunction in ALS. For an accurate evaluation of repeater F waves in ALS patients especially those with moderate to severe muscle atrophy, 100 stimuli would be required.
Multilayer fabric stratification pipes for solar tanks
Andersen, Elsa; Furbo, Simon; Fan, Jianhua
2007-01-01
The thermal performance of solar heating systems is strongly influenced by the thermal stratification in the heat storage. The higher the degree of thermal stratification is, the higher the thermal performance of the solar heating systems. Thermal stratification in water storages can for instance...
Evaluation of 1H NMR relaxometry for the assessment of pore size distribution in soil samples
Jaeger, F.; Bowe, S.; As, van H.; Schaumann, G.E.
2009-01-01
1H NMR relaxometry is used in earth science as a non-destructive and time-saving method to determine pore size distributions (PSD) in porous media with pore sizes ranging from nm to mm. This is a broader range than generally reported for results from X-ray computed tomography (X-ray CT) scanning, wh
Grain size of loess and paleosol samples: what are we measuring?
Varga, György; Kovács, János; Szalai, Zoltán; Újvári, Gábor
2017-04-01
Particle size falling into a particularly narrow range is among the most important properties of windblown mineral dust deposits. Therefore, various aspects of aeolian sedimentation and post-depositional alterations can be reconstructed only from precise grain size data. Present study is aimed at (1) reviewing grain size data obtained from different measurements, (2) discussing the major reasons for disagreements between data obtained by frequently applied particle sizing techniques, and (3) assesses the importance of particle shape in particle sizing. Grain size data of terrestrial aeolian dust deposits (loess and paleosoil) were determined by laser scattering instruments (Fritsch Analysette 22 Microtec Plus, Horiba Partica La-950 v2 and Malvern Mastersizer 3000 with a Hydro Lv unit), while particles size and shape distributions were acquired by Malvern Morphologi G3-ID. Laser scattering results reveal that the optical parameter settings of the measurements have significant effects on the grain size distributions, especially for the fine-grained fractions (Innovation Office (Hungary) under contract NKFI 120620 is gratefully acknowledged. It was additionally supported (for G. Varga) by the Bolyai János Research Scholarship of the Hungarian Academy of Sciences.
Uyaguari-Diaz, Miguel I; Slobodan, Jared R; Nesbitt, Matthew J; Croxen, Matthew A; Isaac-Renton, Judith; Prystajecky, Natalie A; Tang, Patrick
2015-04-17
Next-generation sequencing of environmental samples can be challenging because of the variable DNA quantity and quality in these samples. High quality DNA libraries are needed for optimal results from next-generation sequencing. Environmental samples such as water may have low quality and quantities of DNA as well as contaminants that co-precipitate with DNA. The mechanical and enzymatic processes involved in extraction and library preparation may further damage the DNA. Gel size selection enables purification and recovery of DNA fragments of a defined size for sequencing applications. Nevertheless, this task is one of the most time-consuming steps in the DNA library preparation workflow. The protocol described here enables complete automation of agarose gel loading, electrophoretic analysis, and recovery of targeted DNA fragments. In this study, we describe a high-throughput approach to prepare high quality DNA libraries from freshwater samples that can be applied also to other environmental samples. We used an indirect approach to concentrate bacterial cells from environmental freshwater samples; DNA was extracted using a commercially available DNA extraction kit, and DNA libraries were prepared using a commercial transposon-based protocol. DNA fragments of 500 to 800 bp were gel size selected using Ranger Technology, an automated electrophoresis workstation. Sequencing of the size-selected DNA libraries demonstrated significant improvements to read length and quality of the sequencing reads.
Importance of Sample Size for the Estimation of Repeater F Waves in Amyotrophic Lateral Sclerosis
Jia Fang; Ming-Sheng Liu; Yu-Zhou Guan; Bo Cui; Li-Ying Cui
2015-01-01
Background:In amyotrophic lateral sclerosis (ALS),repeater F waves are increased.Accurate assessment of repeater F waves requires an adequate sample size.Methods:We studied the F waves of left ulnar nerves in ALS patients.Based on the presence or absence of pyramidal signs in the left upper limb,the ALS patients were divided into two groups:One group with pyramidal signs designated as P group and the other without pyramidal signs designated as NP group.The Index repeating neurons (RN) and Index repeater F waves (Freps) were compared among the P,NP and control groups following 20 and 100 stimuli respectively.For each group,the Index RN and Index Freps obtained from 20 and 100 stimuli were compared.Results:In the P group,the Index RN (P =0.004) and Index Freps (P =0.001) obtained from 100 stimuli were significantly higher than from 20 stimuli.For F waves obtained from 20 stimuli,no significant differences were identified between the P and NP groups for Index RN (P =0.052) and Index Freps (P =0.079); The Index RN (P ＜ 0.001) and Index Freps (P ＜ 0.001) of the P group were significantly higher than the control group; The Index RN (P =0.002) of the NP group was significantly higher than the control group.For F waves obtained from 100 stimuli,the Index RN (P ＜ 0.001) and Index Freps (P ＜ 0.001) of the P group were significantly higher than the NP group; The Index RN (P ＜ 0.001) and Index Freps (P ＜ 0.001) of the P and NP groups were significantly higher than the control group.Conclusions:Increased repeater F waves reflect increased excitability of motor neuron pool and indicate upper motor neuron dysfunction in ALS.For an accurate evaluation of repeater F waves in ALS patients especially those with moderate to severe muscle atrophy,100 stimuli would be required.
Bovens, M; Csesztregi, T; Franc, A; Nagy, J; Dujourdy, L
2014-01-01
The basic goal in sampling for the quantitative analysis of illicit drugs is to maintain the average concentration of the drug in the material from its original seized state (the primary sample) all the way through to the analytical sample, where the effect of particle size is most critical. The size of the largest particles of different authentic illicit drug materials, in their original state and after homogenisation, using manual or mechanical procedures, was measured using a microscope with a camera attachment. The comminution methods employed included pestle and mortar (manual) and various ball and knife mills (mechanical). The drugs investigated were amphetamine, heroin, cocaine and herbal cannabis. It was shown that comminution of illicit drug materials using these techniques reduces the nominal particle size from approximately 600 μm down to between 200 and 300 μm. It was demonstrated that the choice of 1 g increments for the primary samples of powdered drugs and cannabis resin, which were used in the heterogeneity part of our study (Part I) was correct for the routine quantitative analysis of illicit seized drugs. For herbal cannabis we found that the appropriate increment size was larger. Based on the results of this study we can generally state that: An analytical sample weight of between 20 and 35 mg of an illicit powdered drug, with an assumed purity of 5% or higher, would be considered appropriate and would generate an RSDsampling in the same region as the RSDanalysis for a typical quantitative method of analysis for the most common, powdered, illicit drugs. For herbal cannabis, with an assumed purity of 1% THC (tetrahydrocannabinol) or higher, an analytical sample weight of approximately 200 mg would be appropriate. In Part III we will pull together our homogeneity studies and particle size investigations and use them to devise sampling plans and sample preparations suitable for the quantitative instrumental analysis of the most common illicit
Core size effect on the dry and saturated ultrasonic pulse velocity of limestone samples.
Ercikdi, Bayram; Karaman, Kadir; Cihangir, Ferdi; Yılmaz, Tekin; Aliyazıcıoğlu, Şener; Kesimal, Ayhan
2016-12-01
This study presents the effect of core length on the saturated (UPVsat) and dry (UPVdry) P-wave velocities of four different biomicritic limestone samples, namely light grey (BL-LG), dark grey (BL-DG), reddish (BL-R) and yellow (BL-Y), using core samples having different lengths (25-125mm) at a constant diameter (54.7mm). The saturated P-wave velocity (UPVsat) of all core samples generally decreased with increasing the sample length. However, the dry P-wave velocity (UPVdry) of samples obtained from BL-LG and BL-Y limestones increased with increasing the sample length. In contrast to the literature, the dry P-wave velocity (UPVdry) values of core samples having a length of 75, 100 and 125mm were consistently higher (2.8-46.2%) than those of saturated (UPVsat). Chemical and mineralogical analyses have shown that the P wave velocity is very sensitive to the calcite and clay minerals potentially leading to the weakening/disintegration of rock samples in the presence of water. Severe fluctuations in UPV values were observed to occur between 25 and 75mm sample lengths, thereafter, a trend of stabilization was observed. The maximum variation of UPV values between the sample length of 75mm and 125mm was only 7.3%. Therefore, the threshold core sample length was interpreted as 75mm for UPV measurement in biomicritic limestone samples used in this study.
Lubbock, Alexander L R; Stewart, Grant D; O'Mahony, Fiach C; Laird, Alexander; Mullen, Peter; O'Donnell, Marie; Powles, Thomas; Harrison, David J; Overton, Ian M
2017-06-26
. Indeed, sample selection could change risk group assignment for 64% of patients, and prognostication with one sample per patient performed only slightly better than random expectation (median logHR = 0.109). Low grade tissue was associated with 3.5-fold greater variation in predicted risk than high grade (p = 0.044). This case study in mccRCC quantitatively demonstrates the critical importance of tumour sampling for the success of molecular biomarker studies research where ITH is a factor. The NEAT model shows promise for mccRCC prognostication and warrants follow-up in larger cohorts. Our work evidences actionable parameters to guide sample collection (tumour coverage, size, grade) to inform the development of reproducible molecular risk stratification methods.
Chan, A.W.; Hrobjartsson, A.; Jorgensen, K.J.
2008-01-01
in 1994-5 by the scientific-ethics committees for Copenhagen and Frederiksberg, Denmark (n=70). MAIN OUTCOME MEASURE: Proportion of protocols and publications that did not provide key information about sample size calculations and statistical methods; proportion of trials with discrepancies between......OBJECTIVE: To evaluate how often sample size calculations and methods of statistical analysis are pre-specified or changed in randomised trials. DESIGN: Retrospective cohort study. Data source Protocols and journal publications of published randomised parallel group trials initially approved...... information presented in the protocol and the publication. RESULTS: Only 11/62 trials described existing sample size calculations fully and consistently in both the protocol and the publication. The method of handling protocol deviations was described in 37 protocols and 43 publications. The method...
Kelley, Ken
2008-01-01
Methods of sample size planning are developed from the accuracy in parameter approach in the multiple regression context in order to obtain a sufficiently narrow confidence interval for the population squared multiple correlation coefficient when regressors are random. Approximate and exact methods are developed that provide necessary sample size so that the expected width of the confidence interval will be sufficiently narrow. Modifications of these methods are then developed so that necessary sample size will lead to sufficiently narrow confidence intervals with no less than some desired degree of assurance. Computer routines have been developed and are included within the MBESS R package so that the methods discussed in the article can be implemented. The methods and computer routines are demonstrated using an empirical example linking innovation in the health services industry with previous innovation, personality factors, and group climate characteristics.
Gerke, Oke; Poulsen, Mads Hvid; Bouchelouche, Kirsten
2009-01-01
PURPOSE: For certain cancer indications, the current patient evaluation strategy is a perfect but locally restricted gold standard procedure. If positron emission tomography/computed tomography (PET/CT) can be shown to be reliable within the gold standard region and if it can be argued that PET....../CT also performs well in adjacent areas, then sample sizes in accuracy studies can be reduced. PROCEDURES: Traditional standard power calculations for demonstrating sensitivities of both 80% and 90% are shown. The argument is then described in general terms and demonstrated by an ongoing study...... of metastasized prostate cancer. RESULTS: An added value in accuracy of PET/CT in adjacent areas can outweigh a downsized target level of accuracy in the gold standard region, justifying smaller sample sizes. CONCLUSIONS: If PET/CT provides an accuracy benefit in adjacent regions, then sample sizes can be reduced...
Draxler, Clemens; Alexandrowicz, Rainer W
2015-12-01
This paper refers to the exponential family of probability distributions and the conditional maximum likelihood (CML) theory. It is concerned with the determination of the sample size for three groups of tests of linear hypotheses, known as the fundamental trinity of Wald, score, and likelihood ratio tests. The main practical purpose refers to the special case of tests of the class of Rasch models. The theoretical background is discussed and the formal framework for sample size calculations is provided, given a predetermined deviation from the model to be tested and the probabilities of the errors of the first and second kinds.
Social Capital and Stratification of Young People
Alireza Behtoui
2013-08-01
Full Text Available This paper addresses the impact of social capital on the status attainment process of young people at the start of their careers and examines how social class, gender and ethnicity affect the accumulation of social capital and thereby labour market stratification of young people. A sample of young Swedes graduating from vocational schools and universities between 2005 and 2006, was surveyed via the telephone about their experiences acquiring jobs. Two research questions are posed: (i Which characteristics (class, gender and ethnicity affect young people's access to more social capital? (ii How is social capital rewarded in the labour market? The results show that being female, coming from the lower social classes and being a member of a stigmatized immigrant groupare associated with a substantial social capital deficit. When socioeconomic and demographic backgrounds as well as the human capital of respondents are controlled, social capital is positively associated with salary level. The results indicate that social capital is a significant factor in the stratification process of young people.
Homeopathy: statistical significance versus the sample size in experiments with Toxoplasma gondii
Ana LÃƒÂºcia Falavigna Guilherme
2011-09-01
, examined in its full length. This study was approved by the Ethics Committee for animal experimentation of the UEM - Protocol 036/2009. The data were compared using the tests Mann Whitney and Bootstrap [7] with the statistical software BioStat 5.0. Results and discussion: There was no significant difference when analyzed with the Mann-Whitney, even multiplying the "n" ten times (p=0.0618. The number of cysts observed in BIOT 200DH group was 4.5 Ã‚Â± 3.3 and 12.8 Ã‚Â± 9.7 in the CONTROL group. Table 1 shows the results obtained using the bootstrap analysis for each data changed from 2n until 2n+5, and their respective p-values. With the inclusion of more elements in the different groups, tested one by one, randomly, increasing gradually the samples, we observed the sample size needed to statistically confirm the results seen experimentally. Using 17 mice in group BIOT 200DH and 19 in the CONTROL group we have already observed statistical significance. This result suggests that experiments involving highly diluted substances and infection of mice with T. gondii should work with experimental groups with 17 animals at least. Despite the current and relevant ethical discussions about the number of animals used for experimental procedures the number of animals involved in each experiment must meet the characteristics of each item to be studied. In the case of experiments involving highly diluted substances, experimental animal models are still rudimentary and the biological effects observed appear to be also individualized, as described in literature for homeopathy [8]. The fact that the statistical significance was achieved by increasing the sample observed in this trial, tell us about a rare event, with a strong individual behavior, difficult to demonstrate in a result set, treated simply with a comparison of means or medians. Conclusion: Bootstrap seems to be an interesting methodology for the analysis of data obtained from experiments with highly diluted
无
2001-01-01
Studies were conducted on specific core collections constructedon the basis of different traits and sample size by the method of stepwise cluster with three sampling strategies based on genotypic values of cotton.A total of 21 traits (11 agronomy traits,5 fiber traits and 5 seed traits) were used to construct main core collections.Specific core collections,as representative of the initial collection,were constructed by agronomy,fiber or seed trait,respectively.As compared with the main core collection,specific core collections tended to have similar property for maintaining genetic diversity of agronomy,seed or fiber traits.Core collections developed by about sample size of 17% (P2=0.17) and 24% (P1= 0.24) with three sampling strategies could be quite representative of the initial collection.
Leinonen, Merja R; Raekallio, Marja R; Vainio, Outi M; Ruohoniemi, Mirja O; O'Brien, Robert T
2011-01-01
Contrast-enhanced ultrasound can be used to quantify tissue perfusion based on region of interest (ROI) analysis. The effect of the location and size of the ROI on the obtained perfusion parameters has been described in phantom, ex vivo and in vivo studies. We assessed the effects of location and size of the ROI on perfusion parameters in the renal cortex of 10 healthy, anesthetized cats using Definity contrast-enhanced ultrasound to estimate the importance of the ROI on quantification of tissue perfusion with contrast-enhanced ultrasound. Three separate sets of ROIs were placed in the renal cortex, varying in location, size or depth. There was a significant inverse association between increased depth or increased size of the ROI and peak intensity (P < 0.05). There was no statistically significant difference in the peak intensity between the ROIs placed in a row in the near field cortex. There was no significant difference in the ROIs with regard to arrival time, time to peak intensity and wash-in rate. When comparing two different ROIs in a patient with focal lesions, such as suspected neoplasia or infarction, the ROIs should always be placed at same depth and be as similar in size as possible.
Evaluating the performance of species richness estimators: sensitivity to sample grain size
Hortal, Joaquín; Borges, Paulo A. V.; Gaspar, Clara
2006-01-01
. Data obtained with standardized sampling of 78 transects in natural forest remnants of five islands were aggregated in seven different grains (i.e. ways of defining a single sample): islands, natural areas, transects, pairs of traps, traps, database records and individuals to assess the effect of using...... in biodiversity studies. Owing to their inherent formulas, several nonparametric and asymptotic estimators present insensitivity to differences in the way the samples are aggregated. Thus, they could be used to compare species richness scores obtained from different sampling strategies. Our results also point out...
Simon Boitard
2016-03-01
Full Text Available Inferring the ancestral dynamics of effective population size is a long-standing question in population genetics, which can now be tackled much more accurately thanks to the massive genomic data available in many species. Several promising methods that take advantage of whole-genome sequences have been recently developed in this context. However, they can only be applied to rather small samples, which limits their ability to estimate recent population size history. Besides, they can be very sensitive to sequencing or phasing errors. Here we introduce a new approximate Bayesian computation approach named PopSizeABC that allows estimating the evolution of the effective population size through time, using a large sample of complete genomes. This sample is summarized using the folded allele frequency spectrum and the average zygotic linkage disequilibrium at different bins of physical distance, two classes of statistics that are widely used in population genetics and can be easily computed from unphased and unpolarized SNP data. Our approach provides accurate estimations of past population sizes, from the very first generations before present back to the expected time to the most recent common ancestor of the sample, as shown by simulations under a wide range of demographic scenarios. When applied to samples of 15 or 25 complete genomes in four cattle breeds (Angus, Fleckvieh, Holstein and Jersey, PopSizeABC revealed a series of population declines, related to historical events such as domestication or modern breed creation. We further highlight that our approach is robust to sequencing errors, provided summary statistics are computed from SNPs with common alleles.
The stratification of regolith on celestial objects
Schräpler, Rainer; von Borstel, Ingo; Güttler, Carsten
2015-01-01
All atmosphere-less planetary bodies are covered with a dust layer, the so-called regolith, which determines the optical, mechanical and thermal properties of their surface. These properties depend on the regolith material, the size distribution of the particles it consists of, and the porosity to which these particles are packed. We performed experiments in parabolic flights to determine the gravity dependency of the packing density of regolith for solid-particle sizes of 60 $\\mu$m and 1 mm as well as for 100-250 $\\mu$m-sized agglomerates of 1.5 $\\mu$m-sized solid grains. We utilized g-levels between 0.7 m s$^{-2}$ and 18 m s$^{-2}$ and completed our measurements with experiments under normal gravity conditions. Based on previous experimental and theoretical literature and supported by our new experiments, we developed an analytical model to calculate the regolith stratification of celestial rocky and icy bodies and estimated the mechanical yields of the regolith under the weight of an astronaut and a spacec...
Method to study sample object size limit of small-angle x-ray scattering computed tomography
Choi, Mina; Ghammraoui, Bahaa; Badal, Andreu; Badano, Aldo
2016-03-01
Small-angle x-ray scattering (SAXS) imaging is an emerging medical tool that can be used for in vivo detailed tissue characterization and has the potential to provide added contrast to conventional x-ray projection and CT imaging. We used a publicly available MC-GPU code to simulate x-ray trajectories in a SAXS-CT geometry for a target material embedded in a water background material with varying sample sizes (1, 3, 5, and 10 mm). Our target materials were water solution of gold nanoparticle (GNP) spheres with a radius of 6 nm and a water solution with dissolved serum albumin (BSA) proteins due to their well-characterized scatter profiles at small angles and highly scattering properties. The background material was water. Our objective is to study how the reconstructed scatter profile degrades at larger target imaging depths and increasing sample sizes. We have found that scatter profiles of the GNP in water can still be reconstructed at depths up to 5 mm embedded at the center of a 10 mm sample. Scatter profiles of BSA in water were also reconstructed at depths up to 5 mm in a 10 mm sample but with noticeable signal degradation as compared to the GNP sample. This work presents a method to study the sample size limits for future SAXS-CT imaging systems.
45 CFR Appendix C to Part 1356 - Calculating Sample Size for NYTD Follow-Up Populations
2010-10-01
... Populations C Appendix C to Part 1356 Public Welfare Regulations Relating to Public Welfare (Continued) OFFICE... Follow-Up Populations 1. Using Finite Population Correction The Finite Population Correction (FPC) is applied when the sample is drawn from a population of one to 5,000 youth, because the sample is more...
The accuracy of instrumental neutron activation analysis of kilogram-size inhomogeneous samples.
Blaauw, M; Lakmaker, O; van Aller, P
1997-07-01
The feasibility of quantitative instrumental neutron activation analysis (INAA) of samples in the kilogram range without internal standardization has been demonstrated by Overwater et al. (Anal. Chem. 1996, 68, 341). In their studies, however, they demonstrated only the agreement between the "corrected" γ ray spectrum of homogeneous large samples and that of small samples of the same material. In this paper, the k(0) calibration of the IRI facilities for large samples is described, and, this time in terms of (trace) element concentrations, some of Overwater's results for homogeneous materials are presented again, as well as results obtained from inhomogeneous materials and subsamples thereof. It is concluded that large-sample INAA can be as accurate as ordinary INAA, even when applied to inhomogeneous materials.
(I Can't Get No) Saturation: A simulation and guidelines for sample sizes in qualitative research.
van Rijnsoever, Frank J
2017-01-01
I explore the sample size in qualitative research that is required to reach theoretical saturation. I conceptualize a population as consisting of sub-populations that contain different types of information sources that hold a number of codes. Theoretical saturation is reached after all the codes in the population have been observed once in the sample. I delineate three different scenarios to sample information sources: "random chance," which is based on probability sampling, "minimal information," which yields at least one new code per sampling step, and "maximum information," which yields the largest number of new codes per sampling step. Next, I use simulations to assess the minimum sample size for each scenario for systematically varying hypothetical populations. I show that theoretical saturation is more dependent on the mean probability of observing codes than on the number of codes in a population. Moreover, the minimal and maximal information scenarios are significantly more efficient than random chance, but yield fewer repetitions per code to validate the findings. I formulate guidelines for purposive sampling and recommend that researchers follow a minimum information scenario.
Does size matter? An investigation into the Rey Complex Figure in a pediatric clinical sample.
Loughan, Ashlee R; Perna, Robert B; Galbreath, Jennifer D
2014-01-01
The Rey Complex Figure Test (RCF) copy requires visuoconstructional skills and significant attentional, organizational, and problem-solving skills. Most scoring schemes codify a subset of the details involved in figure construction. Research is unclear regarding the meaning of figure size. The research hypothesis of our inquiry is that size of the RCF copy will have neuropsychological significance. Data from 95 children (43 girls, 52 boys; ages 6-18 years) with behavioral and academic issues revealed that larger figure drawings were associated with higher RCF total scores and significantly higher scores across many neuropsychological tests including the Wechsler Individual Achievement Test-Second Edition (WIAT-II) Word Reading (F = 5.448, p = .022), WIAT-II Math Reasoning (F = 6.365, p = .013), Children's Memory Scale Visual Delay (F = 4.015, p = .048), Trail-Making Test-Part A (F = 5.448, p = .022), and RCF Recognition (F = 4.862, p = .030). Results indicated that wider figures were associated with higher cognitive functioning, which may be part of an adaptive strategy in helping facilitate accurate and relative proportions of the complex details presented in the RCF. Overall, this study initiates the investigation of the RCF size and the relationship between size and a child's neuropsychological profile.
Algina, James; Keselman, H. J.
2008-01-01
Applications of distribution theory for the squared multiple correlation coefficient and the squared cross-validation coefficient are reviewed, and computer programs for these applications are made available. The applications include confidence intervals, hypothesis testing, and sample size selection. (Contains 2 tables.)
(I Can’t Get No) Saturation: A simulation and guidelines for sample sizes in qualitative research
van Rijnsoever, Frank J.|info:eu-repo/dai/nl/314100334
2017-01-01
I explore the sample size in qualitative research that is required to reach theoretical saturation. I conceptualize a population as consisting of sub-populations that contain different types of information sources that hold a number of codes. Theoretical saturation is reached after all the codes in
van Rijnsoever, F.J.
2015-01-01
This paper explores the sample size in qualitative research that is required to reach theoretical saturation. I conceptualize a population as consisting of sub-populations that contain different types of information sources that hold a number of codes. Theoretical saturation is reached after all the
Leon, Andrew C; Heo, Moonseong
2009-01-15
Mixed-effects linear regression models have become more widely used for analysis of repeatedly measured outcomes in clinical trials over the past decade. There are formulae and tables for estimating sample sizes required to detect the main effects of treatment and the treatment by time interactions for those models. A formula is proposed to estimate the sample size required to detect an interaction between two binary variables in a factorial design with repeated measures of a continuous outcome. The formula is based, in part, on the fact that the variance of an interaction is fourfold that of the main effect. A simulation study examines the statistical power associated with the resulting sample sizes in a mixed-effects linear regression model with a random intercept. The simulation varies the magnitude (Δ) of the standardized main effects and interactions, the intraclass correlation coefficient (ρ ), and the number (k) of repeated measures within-subject. The results of the simulation study verify that the sample size required to detect a 2 × 2 interaction in a mixed-effects linear regression model is fourfold that to detect a main effect of the same magnitude.
Shih, Weichung Joe; Li, Gang; Wang, Yining
2016-03-01
Sample size plays a crucial role in clinical trials. Flexible sample-size designs, as part of the more general category of adaptive designs that utilize interim data, have been a popular topic in recent years. In this paper, we give a comparative review of four related methods for such a design. The likelihood method uses the likelihood ratio test with an adjusted critical value. The weighted method adjusts the test statistic with given weights rather than the critical value. The dual test method requires both the likelihood ratio statistic and the weighted statistic to be greater than the unadjusted critical value. The promising zone approach uses the likelihood ratio statistic with the unadjusted value and other constraints. All four methods preserve the type-I error rate. In this paper we explore their properties and compare their relationships and merits. We show that the sample size rules for the dual test are in conflict with the rules of the promising zone approach. We delineate what is necessary to specify in the study protocol to ensure the validity of the statistical procedure and what can be kept implicit in the protocol so that more flexibility can be attained for confirmatory phase III trials in meeting regulatory requirements. We also prove that under mild conditions, the likelihood ratio test still preserves the type-I error rate when the actual sample size is larger than the re-calculated one.
Heymann, D.; Lakatos, S.; Walton, J. R.
1973-01-01
Review of the results of inert gas measurements performed on six grain-size fractions and two single particles from four samples of Luna 20 material. Presented and discussed data include the inert gas contents, element and isotope systematics, radiation ages, and Ar-36/Ar-40 systematics.
Heymann, D.; Lakatos, S.; Walton, J. R.
1973-01-01
Review of the results of inert gas measurements performed on six grain-size fractions and two single particles from four samples of Luna 20 material. Presented and discussed data include the inert gas contents, element and isotope systematics, radiation ages, and Ar-36/Ar-40 systematics.
van Rijnsoever, F.J.
2015-01-01
This paper explores the sample size in qualitative research that is required to reach theoretical saturation. I conceptualize a population as consisting of sub-populations that contain different types of information sources that hold a number of codes. Theoretical saturation is reached after all the
Hans T. Schreuder; Jin-Mann S. Lin; John Teply
2000-01-01
The Forest Inventory and Analysis units in the USDA Forest Service have been mandated by Congress to go to an annualized inventory where a certain percentage of plots, say 20 percent, will be measured in each State each year. Although this will result in an annual sample size that will be too small for reliable inference for many areas, it is a sufficiently large...
González-Vacarezza, N; Abad-Santos, F; Carcas-Sansuan, A; Dorado, P; Peñas-Lledó, E; Estévez-Carrizo, F; Llerena, A
2013-10-01
In bioequivalence studies, intra-individual variability (CV(w)) is critical in determining sample size. In particular, highly variable drugs may require enrollment of a greater number of subjects. We hypothesize that a strategy to reduce pharmacokinetic CV(w), and hence sample size and costs, would be to include subjects with decreased metabolic enzyme capacity for the drug under study. Therefore, two mirtazapine studies, two-way, two-period crossover design (n=68) were re-analysed to calculate the total CV(w) and the CV(w)s in three different CYP2D6 genotype groups (0, 1 and ≥ 2 active genes). The results showed that a 29.2 or 15.3% sample size reduction would have been possible if the recruitment had been of individuals carrying just 0 or 0 plus 1 CYP2D6 active genes, due to the lower CV(w). This suggests that there may be a role for pharmacogenetics in the design of bioequivalence studies to reduce sample size and costs, thus introducing a new paradigm for the biopharmaceutical evaluation of drug products.
Meyer, J. Patrick; Seaman, Michael A.
2013-01-01
The authors generated exact probability distributions for sample sizes up to 35 in each of three groups ("n" less than or equal to 105) and up to 10 in each of four groups ("n" less than or equal to 40). They compared the exact distributions to the chi-square, gamma, and beta approximations. The beta approximation was best in…
Siqueira, Arminda Lucia; Todd, Susan; Whitehead, Anne
2015-08-01
This paper presents an approximate closed form sample size formula for determining non-inferiority in active-control trials with binary data. We use the odds-ratio as the measure of the relative treatment effect, derive the sample size formula based on the score test and compare it with a second, well-known formula based on the Wald test. Both closed form formulae are compared with simulations based on the likelihood ratio test. Within the range of parameter values investigated, the score test closed form formula is reasonably accurate when non-inferiority margins are based on odds-ratios of about 0.5 or above and when the magnitude of the odds ratio under the alternative hypothesis lies between about 1 and 2.5. The accuracy generally decreases as the odds ratio under the alternative hypothesis moves upwards from 1. As the non-inferiority margin odds ratio decreases from 0.5, the score test closed form formula increasingly overestimates the sample size irrespective of the magnitude of the odds ratio under the alternative hypothesis. The Wald test closed form formula is also reasonably accurate in the cases where the score test closed form formula works well. Outside these scenarios, the Wald test closed form formula can either underestimate or overestimate the sample size, depending on the magnitude of the non-inferiority margin odds ratio and the odds ratio under the alternative hypothesis. Although neither approximation is accurate for all cases, both approaches lead to satisfactory sample size calculation for non-inferiority trials with binary data where the odds ratio is the parameter of interest. © The Author(s) 2014.
Ellison, Laura E.; Lukacs, Paul M.
2014-01-01
Concern for migratory tree-roosting bats in North America has grown because of possible population declines from wind energy development. This concern has driven interest in estimating population-level changes. Mark-recapture methodology is one possible analytical framework for assessing bat population changes, but sample size requirements to produce reliable estimates have not been estimated. To illustrate the sample sizes necessary for a mark-recapture-based monitoring program we conducted power analyses using a statistical model that allows reencounters of live and dead marked individuals. We ran 1,000 simulations for each of five broad sample size categories in a Burnham joint model, and then compared the proportion of simulations in which 95% confidence intervals overlapped between and among years for a 4-year study. Additionally, we conducted sensitivity analyses of sample size to various capture probabilities and recovery probabilities. More than 50,000 individuals per year would need to be captured and released to accurately determine 10% and 15% declines in annual survival. To detect more dramatic declines of 33% or 50% survival over four years, then sample sizes of 25,000 or 10,000 per year, respectively, would be sufficient. Sensitivity analyses reveal that increasing recovery of dead marked individuals may be more valuable than increasing capture probability of marked individuals. Because of the extraordinary effort that would be required, we advise caution should such a mark-recapture effort be initiated because of the difficulty in attaining reliable estimates. We make recommendations for what techniques show the most promise for mark-recapture studies of bats because some techniques violate the assumptions of mark-recapture methodology when used to mark bats.
Use of High-Frequency In-Home Monitoring Data May Reduce Sample Sizes Needed in Clinical Trials.
Hiroko H Dodge
Full Text Available Trials in Alzheimer's disease are increasingly focusing on prevention in asymptomatic individuals. This poses a challenge in examining treatment effects since currently available approaches are often unable to detect cognitive and functional changes among asymptomatic individuals. Resultant small effect sizes require large sample sizes using biomarkers or secondary measures for randomized controlled trials (RCTs. Better assessment approaches and outcomes capable of capturing subtle changes during asymptomatic disease stages are needed.We aimed to develop a new approach to track changes in functional outcomes by using individual-specific distributions (as opposed to group-norms of unobtrusive continuously monitored in-home data. Our objective was to compare sample sizes required to achieve sufficient power to detect prevention trial effects in trajectories of outcomes in two scenarios: (1 annually assessed neuropsychological test scores (a conventional approach, and (2 the likelihood of having subject-specific low performance thresholds, both modeled as a function of time.One hundred nineteen cognitively intact subjects were enrolled and followed over 3 years in the Intelligent Systems for Assessing Aging Change (ISAAC study. Using the difference in empirically identified time slopes between those who remained cognitively intact during follow-up (normal control, NC and those who transitioned to mild cognitive impairment (MCI, we estimated comparative sample sizes required to achieve up to 80% statistical power over a range of effect sizes for detecting reductions in the difference in time slopes between NC and MCI incidence before transition.Sample size estimates indicated approximately 2000 subjects with a follow-up duration of 4 years would be needed to achieve a 30% effect size when the outcome is an annually assessed memory test score. When the outcome is likelihood of low walking speed defined using the individual-specific distributions of
Sample Size Effect of Magnetomechanical Response for Magnetic Elastomers by Using Permanent Magnets
Tsubasa Oguro
2017-01-01
Full Text Available The size effect of magnetomechanical response of chemically cross-linked disk shaped magnetic elastomers placed on a permanent magnet has been investigated by unidirectional compression tests. A cylindrical permanent magnet with a size of 35 mm in diameter and 15 mm in height was used to create the magnetic field. The magnetic field strength was approximately 420 mT at the center of the upper surface of the magnet. The diameter of the magnetoelastic polymer disks was varied from 14 mm to 35 mm, whereas the height was kept constant (5 mm in the undeformed state. We have studied the influence of the disk diameter on the stress-strain behavior of the magnetoelastic in the presence and in the lack of magnetic field. It was found that the smallest magnetic elastomer with 14 mm diameter did not exhibit measurable magnetomechanical response due to magnetic field. On the opposite, the magnetic elastomers with diameters larger than 30 mm contracted in the direction parallel to the mechanical stress and largely elongated in the perpendicular direction. An explanation is put forward to interpret this size-dependent behavior by taking into account the nonuniform field distribution of magnetic field produced by the permanent magnet.
Gardi, Jonathan Eyal; Nyengaard, Jens Randel; Gundersen, Hans Jørgen Gottlieb
2008-01-01
of its entirely different sampling strategy, based on known but non-uniform sampling probabilities, the proportionator for the first time allows the real CE at the section level to be automatically estimated (not just predicted), unbiased - for all estimators and at no extra cost to the user.......The proportionator is a novel and radically different approach to sampling with microscopes based on well-known statistical theory (probability proportional to size - PPS sampling). It uses automatic image analysis, with a large range of options, to assign to every field of view in the section......, the desired number of fields are sampled automatically with probability proportional to the weight and presented to the expert observer. Using any known stereological probe and estimator, the correct count in these fields leads to a simple, unbiased estimate of the total amount of structure in the sections...
Gonzalez, Susana; Yu, Woojin M; Smith, Michael S; Slack, Kristen N; Rotterdam, Heidrun; Abrams, Julian A; Lightdale, Charles J
2010-11-01
Several types of forceps are available for use in sampling Barrett's esophagus (BE). Few data exist with regard to biopsy quality for histologic assessment. To evaluate sampling quality of 3 different forceps in patients with BE. Single-center, randomized clinical trial. Consecutive patients with BE undergoing upper endoscopy. Patients randomized to have biopsy specimens taken with 1 of 3 types of forceps: standard, large capacity, or jumbo. Specimen adequacy was defined a priori as a well-oriented biopsy sample 2 mm or greater in diameter and with at least muscularis mucosa present. A total of 65 patients were enrolled and analyzed (standard forceps, n = 21; large-capacity forceps, n = 21; jumbo forceps, n = 23). Compared with jumbo forceps, a significantly higher proportion of biopsy samples with large-capacity forceps were adequate (37.8% vs 25.2%, P = .002). Of the standard forceps biopsy samples, 31.9% were adequate, which was not significantly different from specimens taken with large-capacity (P = .20) or jumbo (P = .09) forceps. Biopsy specimens taken with jumbo forceps had the largest diameter (median, 3.0 mm vs 2.5 mm [standard] vs 2.8 mm [large capacity]; P = .0001). However, jumbo forceps had the lowest proportion of specimens that were well oriented (overall P = .001). Heterogeneous patient population precluded dysplasia detection analyses. Our results challenge the requirement of jumbo forceps and therapeutic endoscopes to properly perform the Seattle protocol. We found that standard and large-capacity forceps used with standard upper endoscopes produced biopsy samples at least as adequate as those obtained with jumbo forceps and therapeutic endoscopes in patients with BE. Copyright © 2010 American Society for Gastrointestinal Endoscopy. Published by Mosby, Inc. All rights reserved.
Pan, Bo; Shibutani, Yoji, E-mail: sibutani@mech.eng.osaka-u.ac.jp [Department of Mechanical Engineering, Osaka University, Suita 565-0871 (Japan); Zhang, Xu [State Key Laboratory for Strength and Vibration of Mechanical Structures, School of Aerospace, Xi' an Jiaotong University, Xi' an 710049 (China); School of Mechanics and Engineering Science, Zhengzhou University, Zhengzhou 450001 (China); Shang, Fulin [State Key Laboratory for Strength and Vibration of Mechanical Structures, School of Aerospace, Xi' an Jiaotong University, Xi' an 710049 (China)
2015-07-07
Recent research has explained that the steeply increasing yield strength in metals depends on decreasing sample size. In this work, we derive a statistical physical model of the yield strength of finite single-crystal micro-pillars that depends on single-ended dislocation pile-up inside the micro-pillars. We show that this size effect can be explained almost completely by considering the stochastic lengths of the dislocation source and the dislocation pile-up length in the single-crystal micro-pillars. The Hall–Petch-type relation holds even in a microscale single-crystal, which is characterized by its dislocation source lengths. Our quantitative conclusions suggest that the number of dislocation sources and pile-ups are significant factors for the size effect. They also indicate that starvation of dislocation sources is another reason for the size effect. Moreover, we investigated the explicit relationship between the stacking fault energy and the dislocation “pile-up” effect inside the sample: materials with low stacking fault energy exhibit an obvious dislocation pile-up effect. Our proposed physical model predicts a sample strength that agrees well with experimental data, and our model can give a more precise prediction than the current single arm source model, especially for materials with low stacking fault energy.
Usami, Satoshi
2014-12-01
Recent years have shown increased awareness of the importance of sample size determination in experimental research. Yet effective and convenient methods for sample size determination, especially in longitudinal experimental design, are still under development, and application of power analysis in applied research remains limited. This article presents a convenient method for sample size determination in longitudinal experimental research using a multilevel model. A fundamental idea of this method is transformation of model parameters (level 1 error variance [σ(2)], level 2 error variances [τ 00, τ 11] and its covariance [τ 01, τ 10], and a parameter representing experimental effect [δ]) into indices (reliability of measurement at the first time point [ρ 1], effect size at the last time point [Δ T ], proportion of variance of outcomes between the first and the last time points [k], and level 2 error correlation [r]) that are intuitively understandable and easily specified. To foster more convenient use of power analysis, numerical tables are constructed that refer to ANOVA results to investigate the influence on statistical power by respective indices.
Szymańska, Ewa; Brodrick, Emma; Williams, Mark; Davies, Antony N; van Manen, Henk-Jan; Buydens, Lutgarde M C
2015-01-20
Ion mobility spectrometry combined with multicapillary column separation (MCC-IMS) is a well-known technology for detecting volatile organic compounds (VOCs) in gaseous samples. Due to their large data size, processing of MCC-IMS spectra is still the main bottleneck of data analysis, and there is an increasing need for data analysis strategies in which the size of MCC-IMS data is reduced to enable further analysis. In our study, the first untargeted chemometric strategy is developed and employed in the analysis of MCC-IMS spectra from 264 breath and ambient air samples. This strategy does not comprise identification of compounds as a primary step but includes several preprocessing steps and a discriminant analysis. Data size is significantly reduced in three steps. Wavelet transform, mask construction, and sparse-partial least squares-discriminant analysis (s-PLS-DA) allow data size reduction with down to 50 variables relevant to the goal of analysis. The influence and compatibility of the data reduction tools are studied by applying different settings of the developed strategy. Loss of information after preprocessing is evaluated, e.g., by comparing the performance of classification models for different classes of samples. Finally, the interpretability of the classification models is evaluated, and regions of spectra that are related to the identification of potential analytical biomarkers are successfully determined. This work will greatly enable the standardization of analytical procedures across different instrumentation types promoting the adoption of MCC-IMS technology in a wide range of diverse application fields.
Early detection of nonnative alleles in fish populations: When sample size actually matters
Croce, Patrick Della; Poole, Geoffrey C.; Payne, Robert A.; Gresswell, Bob
2017-01-01
Reliable detection of nonnative alleles is crucial for the conservation of sensitive native fish populations at risk of introgression. Typically, nonnative alleles in a population are detected through the analysis of genetic markers in a sample of individuals. Here we show that common assumptions associated with such analyses yield substantial overestimates of the likelihood of detecting nonnative alleles. We present a revised equation to estimate the likelihood of detecting nonnative alleles in a population with a given level of admixture. The new equation incorporates the effects of the genotypic structure of the sampled population and shows that conventional methods overestimate the likelihood of detection, especially when nonnative or F-1 hybrid individuals are present. Under such circumstances—which are typical of early stages of introgression and therefore most important for conservation efforts—our results show that improved detection of nonnative alleles arises primarily from increasing the number of individuals sampled rather than increasing the number of genetic markers analyzed. Using the revised equation, we describe a new approach to determining the number of individuals to sample and the number of diagnostic markers to analyze when attempting to monitor the arrival of nonnative alleles in native populations.
Evaluation of pump pulsation in respirable size-selective sampling: part I. Pulsation measurements.
Lee, Eun Gyung; Lee, Larry; Möhlmann, Carsten; Flemmer, Michael M; Kashon, Michael; Harper, Martin
2014-01-01
Pulsations generated by personal sampling pumps modulate the airflow through the sampling trains, thereby varying sampling efficiencies, and possibly invalidating collection or monitoring. The purpose of this study was to characterize pulsations generated by personal sampling pumps relative to a nominal flow rate at the inlet of different respirable cyclones. Experiments were conducted using a factorial combination of 13 widely used sampling pumps (11 medium and 2 high volumetric flow rate pumps having a diaphragm mechanism) and 7 cyclones [10-mm nylon also known as Dorr-Oliver (DO), Higgins-Dewell (HD), GS-1, GS-3, Aluminum, GK2.69, and FSP-10]. A hot-wire anemometer probe cemented to the inlet of each cyclone type was used to obtain pulsation readings. The three medium flow rate pump models showing the highest, a midrange, and the lowest pulsations and two high flow rate pump models for each cyclone type were tested with dust-loaded filters (0.05, 0.21, and 1.25mg) to determine the effects of filter loading on pulsations. The effects of different tubing materials and lengths on pulsations were also investigated. The fundamental frequency range was 22-110 Hz and the magnitude of pulsation as a proportion of the mean flow rate ranged from 4.4 to 73.1%. Most pump/cyclone combinations generated pulse magnitudes ≥10% (48 out of 59 combinations), while pulse shapes varied considerably. Pulsation magnitudes were not considerably different for the clean and dust-loaded filters for the DO, HD, and Aluminum cyclones, but no consistent pattern was observed for the other cyclone types. Tubing material had less effect on pulsations than tubing length; when the tubing length was 183cm, pronounced damping was observed for a pump with high pulsation (>60%) for all tested tubing materials except for the Tygon Inert tubing. The findings in this study prompted a further study to determine the possibility of shifts in cyclone sampling efficiency due to sampling pump pulsations
Li, Chung-I; Shyr, Yu
2016-12-01
As RNA-seq rapidly develops and costs continually decrease, the quantity and frequency of samples being sequenced will grow exponentially. With proteomic investigations becoming more multivariate and quantitative, determining a study's optimal sample size is now a vital step in experimental design. Current methods for calculating a study's required sample size are mostly based on the hypothesis testing framework, which assumes each gene count can be modeled through Poisson or negative binomial distributions; however, these methods are limited when it comes to accommodating covariates. To address this limitation, we propose an estimating procedure based on the generalized linear model. This easy-to-use method constructs a representative exemplary dataset and estimates the conditional power, all without requiring complicated mathematical approximations or formulas. Even more attractive, the downstream analysis can be performed with current R/Bioconductor packages. To demonstrate the practicability and efficiency of this method, we apply it to three real-world studies, and introduce our on-line calculator developed to determine the optimal sample size for a RNA-seq study.
In situ detection of small-size insect pests sampled on traps using multifractal analysis
Xia, Chunlei; Lee, Jang-Myung; Li, Yan; Chung, Bu-Keun; Chon, Tae-Soo
2012-02-01
We introduce a multifractal analysis for detecting the small-size pest (e.g., whitefly) images from a sticky trap in situ. An automatic attraction system is utilized for collecting pests from greenhouse plants. We applied multifractal analysis to segment action of whitefly images based on the local singularity and global image characteristics. According to the theory of multifractal dimension, the candidate blobs of whiteflies are initially defined from the sticky-trap image. Two schemes, fixed thresholding and regional minima obtainment, were utilized for feature extraction of candidate whitefly image areas. The experiment was conducted with the field images in a greenhouse. Detection results were compared with other adaptive segmentation algorithms. Values of F measuring precision and recall score were higher for the proposed multifractal analysis (96.5%) compared with conventional methods such as Watershed (92.2%) and Otsu (73.1%). The true positive rate of multifractal analysis was 94.3% and the false positive rate minimal level at 1.3%. Detection performance was further tested via human observation. The degree of scattering between manual and automatic counting was remarkably higher with multifractal analysis (R2=0.992) compared with Watershed (R2=0.895) and Otsu (R2=0.353), ensuring overall detection of the small-size pests is most feasible with multifractal analysis in field conditions.
John M Lachin
Full Text Available Preservation of β-cell function as measured by stimulated C-peptide has recently been accepted as a therapeutic target for subjects with newly diagnosed type 1 diabetes. In recently completed studies conducted by the Type 1 Diabetes Trial Network (TrialNet, repeated 2-hour Mixed Meal Tolerance Tests (MMTT were obtained for up to 24 months from 156 subjects with up to 3 months duration of type 1 diabetes at the time of study enrollment. These data provide the information needed to more accurately determine the sample size needed for future studies of the effects of new agents on the 2-hour area under the curve (AUC of the C-peptide values. The natural log(x, log(x+1 and square-root (√x transformations of the AUC were assessed. In general, a transformation of the data is needed to better satisfy the normality assumptions for commonly used statistical tests. Statistical analysis of the raw and transformed data are provided to estimate the mean levels over time and the residual variation in untreated subjects that allow sample size calculations for future studies at either 12 or 24 months of follow-up and among children 8-12 years of age, adolescents (13-17 years and adults (18+ years. The sample size needed to detect a given relative (percentage difference with treatment versus control is greater at 24 months than at 12 months of follow-up, and differs among age categories. Owing to greater residual variation among those 13-17 years of age, a larger sample size is required for this age group. Methods are also described for assessment of sample size for mixtures of subjects among the age categories. Statistical expressions are presented for the presentation of analyses of log(x+1 and √x transformed values in terms of the original units of measurement (pmol/ml. Analyses using different transformations are described for the TrialNet study of masked anti-CD20 (rituximab versus masked placebo. These results provide the information needed to
Memory-Optimized Software Synthesis from Dataflow Program Graphs with Large Size Data Samples
Hyunok Oh
2003-05-01
Full Text Available In multimedia and graphics applications, data samples of nonprimitive type require significant amount of buffer memory. This paper addresses the problem of minimizing the buffer memory requirement for such applications in embedded software synthesis from graphical dataflow programs based on the synchronous dataflow (SDF model with the given execution order of nodes. We propose a memory minimization technique that separates global memory buffers from local pointer buffers: the global buffers store live data samples and the local buffers store the pointers to the global buffer entries. The proposed algorithm reduces 67% memory for a JPEG encoder, 40% for an H.263 encoder compared with unshared versions, and 22% compared with the previous sharing algorithm for the H.263 encoder. Through extensive buffer sharing optimization, we believe that automatic software synthesis from dataflow program graphs achieves the comparable code quality with the manually optimized code in terms of memory requirement.
Basic distribution free identification tests for small size samples of environmental data
Federico, A.G.; Musmeci, F. [ENEA, Centro Ricerche Casaccia, Rome (Italy). Dipt. Ambiente
1998-01-01
Testing two or more data sets for the hypothesis that they are sampled form the same population is often required in environmental data analysis. Typically the available samples have a small number of data and often then assumption of normal distributions is not realistic. On the other hand the diffusion of the days powerful Personal Computers opens new possible opportunities based on a massive use of the CPU resources. The paper reviews the problem introducing the feasibility of two non parametric approaches based on intrinsic equi probability properties of the data samples. The first one is based on a full re sampling while the second is based on a bootstrap approach. A easy to use program is presented. A case study is given based on the Chernobyl children contamination data. [Italiano] Nell`analisi di dati ambientali ricorre spesso il caso di dover sottoporre a test l`ipotesi di provenienza di due, o piu`, insiemi di dati dalla stessa popolazione. Tipicamente i dati disponibili sono pochi e spesso l`ipotesi di provenienza da distribuzioni normali non e` sostenibile. D`altra aprte la diffusione odierna di Personal Computer fornisce nuove possibili soluzioni basate sull`uso intensivo delle risorse della CPU. Il rapporto analizza il problema e presenta la possibilita` di utilizzo di due test non parametrici basati sulle proprieta` intrinseche di equiprobabilita` dei campioni. Il primo e` basato su una tecnica di ricampionamento esaustivo mentre il secondo su un approccio di tipo bootstrap. E` presentato un programma di semplice utilizzo e un caso di studio basato su dati di contaminazione di bambini a Chernobyl.
Second generation laser-heated microfurnace for the preparation of microgram-sized graphite samples
Yang, Bin; Smith, A. M.; Long, S.
2015-10-01
We present construction details and test results for two second-generation laser-heated microfurnaces (LHF-II) used to prepare graphite samples for Accelerator Mass Spectrometry (AMS) at ANSTO. Based on systematic studies aimed at optimising the performance of our prototype laser-heated microfurnace (LHF-I) (Smith et al., 2007 [1]; Smith et al., 2010 [2,3]; Yang et al., 2014 [4]), we have designed the LHF-II to have the following features: (i) it has a small reactor volume of 0.25 mL allowing us to completely graphitise carbon dioxide samples containing as little as 2 μg of C, (ii) it can operate over a large pressure range (0-3 bar) and so has the capacity to graphitise CO2 samples containing up to 100 μg of C; (iii) it is compact, with three valves integrated into the microfurnace body, (iv) it is compatible with our new miniaturised conventional graphitisation furnaces (MCF), also designed for small samples, and shares a common vacuum system. Early tests have shown that the extraneous carbon added during graphitisation in each LHF-II is of the order of 0.05 μg, assuming 100 pMC activity, similar to that of the prototype unit. We use a 'budget' fibre packaged array for the diode laser with custom built focusing optics. The use of a new infrared (IR) thermometer with a short focal length has allowed us to decrease the height of the light-proof safety enclosure. These innovations have produced a cheaper and more compact device. As with the LHF-I, feedback control of the catalyst temperature and logging of the reaction parameters is managed by a LabVIEW interface.
Second generation laser-heated microfurnace for the preparation of microgram-sized graphite samples
Yang, Bin; Smith, A.M.; Long, S.
2015-10-15
We present construction details and test results for two second-generation laser-heated microfurnaces (LHF-II) used to prepare graphite samples for Accelerator Mass Spectrometry (AMS) at ANSTO. Based on systematic studies aimed at optimising the performance of our prototype laser-heated microfurnace (LHF-I) (Smith et al., 2007 [1]; Smith et al., 2010 [2,3]; Yang et al., 2014 [4]), we have designed the LHF-II to have the following features: (i) it has a small reactor volume of 0.25 mL allowing us to completely graphitise carbon dioxide samples containing as little as 2 μg of C, (ii) it can operate over a large pressure range (0–3 bar) and so has the capacity to graphitise CO{sub 2} samples containing up to 100 μg of C; (iii) it is compact, with three valves integrated into the microfurnace body, (iv) it is compatible with our new miniaturised conventional graphitisation furnaces (MCF), also designed for small samples, and shares a common vacuum system. Early tests have shown that the extraneous carbon added during graphitisation in each LHF-II is of the order of 0.05 μg, assuming 100 pMC activity, similar to that of the prototype unit. We use a ‘budget’ fibre packaged array for the diode laser with custom built focusing optics. The use of a new infrared (IR) thermometer with a short focal length has allowed us to decrease the height of the light-proof safety enclosure. These innovations have produced a cheaper and more compact device. As with the LHF-I, feedback control of the catalyst temperature and logging of the reaction parameters is managed by a LabVIEW interface.
Picchini, Umberto; Forman, Julie Lyng
2016-01-01
In recent years, dynamical modelling has been provided with a range of breakthrough methods to perform exact Bayesian inference. However, it is often computationally unfeasible to apply exact statistical methodologies in the context of large data sets and complex models. This paper considers...... a nonlinear stochastic differential equation model observed with correlated measurement errors and an application to protein folding modelling. An approximate Bayesian computation (ABC)-MCMC algorithm is suggested to allow inference for model parameters within reasonable time constraints. The ABC algorithm...... applications. A simulation study is conducted to compare our strategy with exact Bayesian inference, the latter resulting two orders of magnitude slower than ABC-MCMC for the considered set-up. Finally, the ABC algorithm is applied to a large size protein data. The suggested methodology is fairly general...
Fang, J; Cui, L Y; Liu, M S; Guan, Y Z; Ding, Q Y; Du, H; Li, B H; Wu, S
2017-03-07
Objective: The study aimed to investigate whether sample sizes of F-wave study differed according to different nerves, different F-wave parameters, and amyotrophic lateral sclerosis(ALS) patients or healthy subjects. Methods: The F-waves in the median, ulnar, tibial, and deep peroneal nerves of 55 amyotrophic lateral sclerosis (ALS) patients and 52 healthy subjects were studied to assess the effect of sample size on the accuracy of measurements of the following F-wave parameters: F-wave minimum latency, maximum latency, mean latency, F-wave persistence, F-wave chronodispersion, mean and maximum F-wave amplitude. A hundred stimuli were used in F-wave study. The values obtained from 100 stimuli were considered "true" values and were compared with the corresponding values from smaller samples of 20, 40, 60 and 80 stimuli. F-wave parameters obtained from different sample sizes were compared between the ALS patients and the normal controls. Results: Significant differences were not detected with samples above 60 stimuli for chronodispersion in all four nerves in normal participants. Significant differences were not detected with samples above 40 stimuli for maximum F-wave amplitude in median, ulnar and tibial nerves in normal participants. When comparing ALS patients and normal controls, significant differences were detected in the maximum (median nerve, Z=-3.560, PF-wave latency (median nerve, Z=-3.243, PF-wave chronodispersion (Z=-3.152, PF-wave persistence in the median (Z=6.139, PF-wave amplitude in the tibial nerve(t=2.981, PF-wave amplitude in the ulnar (Z=-2.134, PF-wave persistence in tibial nerve (Z=2.119, PF-wave amplitude in ulnar (Z=-2.552, PF-wave amplitude in peroneal nerve (t=2.693, PF-wave study differed according to different nerves, different F-wave parameters , and ALS patients or healthy subjects.
Faye, C.B.; Amodeo, T.; Fréjafon, E. [Institut National de l' Environnement Industriel et des Risques (INERIS/DRC/CARA/NOVA), Parc Technologique Alata, BP 2, 60550 Verneuil-En-Halatte (France); Delepine-Gilon, N. [Institut des Sciences Analytiques, 5 rue de la Doua, 69100 Villeurbanne (France); Dutouquet, C., E-mail: christophe.dutouquet@ineris.fr [Institut National de l' Environnement Industriel et des Risques (INERIS/DRC/CARA/NOVA), Parc Technologique Alata, BP 2, 60550 Verneuil-En-Halatte (France)
2014-01-01
Pollution of water is a matter of concern all over the earth. Particles are known to play an important role in the transportation of pollutants in this medium. In addition, the emergence of new materials such as NOAA (Nano-Objects, their Aggregates and their Agglomerates) emphasizes the need to develop adapted instruments for their detection. Surveillance of pollutants in particulate form in waste waters in industries involved in nanoparticle manufacturing and processing is a telling example of possible applications of such instrumental development. The LIBS (laser-induced breakdown spectroscopy) technique coupled with the liquid jet as sampling mode for suspensions was deemed as a potential candidate for on-line and real time monitoring. With the final aim in view to obtain the best detection limits, the interaction of nanosecond laser pulses with the liquid jet was examined. The evolution of the volume sampled by laser pulses was estimated as a function of the laser energy applying conditional analysis when analyzing a suspension of micrometric-sized particles of borosilicate glass. An estimation of the sampled depth was made. Along with the estimation of the sampled volume, the evolution of the SNR (signal to noise ratio) as a function of the laser energy was investigated as well. Eventually, the laser energy and the corresponding fluence optimizing both the sampling volume and the SNR were determined. The obtained results highlight intrinsic limitations of the liquid jet sampling mode when using 532 nm nanosecond laser pulses with suspensions. - Highlights: • Micrometric-sized particles in suspensions are analyzed using LIBS and a liquid jet. • The evolution of the sampling volume is estimated as a function of laser energy. • The sampling volume happens to saturate beyond a certain laser fluence. • Its value was found much lower than the beam diameter times the jet thickness. • Particles proved not to be entirely vaporized.
Finch Stephen J
2005-04-01
Full Text Available Abstract Background Phenotype error causes reduction in power to detect genetic association. We present a quantification of phenotype error, also known as diagnostic error, on power and sample size calculations for case-control genetic association studies between a marker locus and a disease phenotype. We consider the classic Pearson chi-square test for independence as our test of genetic association. To determine asymptotic power analytically, we compute the distribution's non-centrality parameter, which is a function of the case and control sample sizes, genotype frequencies, disease prevalence, and phenotype misclassification probabilities. We derive the non-centrality parameter in the presence of phenotype errors and equivalent formulas for misclassification cost (the percentage increase in minimum sample size needed to maintain constant asymptotic power at a fixed significance level for each percentage increase in a given misclassification parameter. We use a linear Taylor Series approximation for the cost of phenotype misclassification to determine lower bounds for the relative costs of misclassifying a true affected (respectively, unaffected as a control (respectively, case. Power is verified by computer simulation. Results Our major findings are that: (i the median absolute difference between analytic power with our method and simulation power was 0.001 and the absolute difference was no larger than 0.011; (ii as the disease prevalence approaches 0, the cost of misclassifying a unaffected as a case becomes infinitely large while the cost of misclassifying an affected as a control approaches 0. Conclusion Our work enables researchers to specifically quantify power loss and minimum sample size requirements in the presence of phenotype errors, thereby allowing for more realistic study design. For most diseases of current interest, verifying that cases are correctly classified is of paramount importance.
Luo, Shezhou; Chen, Jing M; Wang, Cheng; Xi, Xiaohuan; Zeng, Hongcheng; Peng, Dailiang; Li, Dong
2016-05-30
Vegetation leaf area index (LAI), height, and aboveground biomass are key biophysical parameters. Corn is an important and globally distributed crop, and reliable estimations of these parameters are essential for corn yield forecasting, health monitoring and ecosystem modeling. Light Detection and Ranging (LiDAR) is considered an effective technology for estimating vegetation biophysical parameters. However, the estimation accuracies of these parameters are affected by multiple factors. In this study, we first estimated corn LAI, height and biomass (R2 = 0.80, 0.874 and 0.838, respectively) using the original LiDAR data (7.32 points/m2), and the results showed that LiDAR data could accurately estimate these biophysical parameters. Second, comprehensive research was conducted on the effects of LiDAR point density, sampling size and height threshold on the estimation accuracy of LAI, height and biomass. Our findings indicated that LiDAR point density had an important effect on the estimation accuracy for vegetation biophysical parameters, however, high point density did not always produce highly accurate estimates, and reduced point density could deliver reasonable estimation results. Furthermore, the results showed that sampling size and height threshold were additional key factors that affect the estimation accuracy of biophysical parameters. Therefore, the optimal sampling size and the height threshold should be determined to improve the estimation accuracy of biophysical parameters. Our results also implied that a higher LiDAR point density, larger sampling size and height threshold were required to obtain accurate corn LAI estimation when compared with height and biomass estimations. In general, our results provide valuable guidance for LiDAR data acquisition and estimation of vegetation biophysical parameters using LiDAR data.
LIU Yixing; CHENG Ping
2000-01-01
Cheng[1]gave the limit distribution of weighted PP Cramér-Von Mises test statistic when dimension and sample size tend to infinity simultaneously under the underlying distribution being uniform distribution on Sp-1 = {a:‖a‖ = 1, a ∈ SP-1}; this limit distribution is standard normal distribution N(0, 1). In this paper, we give the BerryEsseen bound of this statistic converging to normal distribution and the law of iterated logarithm.
Tai, Bee-Choo; Grundy, Richard; Machin, David
2011-03-15
To accurately model the cumulative need for radiotherapy in trials designed to delay or avoid irradiation among children with malignant brain tumor, it is crucial to account for competing events and evaluate how each contributes to the timing of irradiation. An appropriate choice of statistical model is also important for adequate determination of sample size. We describe the statistical modeling of competing events (A, radiotherapy after progression; B, no radiotherapy after progression; and C, elective radiotherapy) using proportional cause-specific and subdistribution hazard functions. The procedures of sample size estimation based on each method are outlined. These are illustrated by use of data comparing children with ependymoma and other malignant brain tumors. The results from these two approaches are compared. The cause-specific hazard analysis showed a reduction in hazards among infants with ependymoma for all event types, including Event A (adjusted cause-specific hazard ratio, 0.76; 95% confidence interval, 0.45-1.28). Conversely, the subdistribution hazard analysis suggested an increase in hazard for Event A (adjusted subdistribution hazard ratio, 1.35; 95% confidence interval, 0.80-2.30), but the reduction in hazards for Events B and C remained. Analysis based on subdistribution hazard requires a larger sample size than the cause-specific hazard approach. Notable differences in effect estimates and anticipated sample size were observed between methods when the main event showed a beneficial effect whereas the competing events showed an adverse effect on the cumulative incidence. The subdistribution hazard is the most appropriate for modeling treatment when its effects on both the main and competing events are of interest. Copyright © 2011 Elsevier Inc. All rights reserved.
Shuaicheng Guo
2017-03-01
Full Text Available Entrained air voids can improve the freeze-thaw durability of concrete, and also affect its mechanical and transport properties. Therefore, it is important to measure the air void structure and understand its influence on concrete performance for quality control. This paper aims to measure air void structure evolution at both early-age and hardened stages with the ultrasonic technique, and evaluates its influence on concrete properties. Three samples with different air entrainment agent content were specially prepared. The air void structure was determined with optimized inverse analysis by achieving the minimum error between experimental and theoretical attenuation. The early-age sample measurement showed that the air void content with the whole size range slightly decreases with curing time. The air void size distribution of hardened samples (at Day 28 was compared with American Society for Testing and Materials (ASTM C457 test results. The air void size distribution with different amount of air entrainment agent was also favorably compared. In addition, the transport property, compressive strength, and dynamic modulus of concrete samples were also evaluated. The concrete transport decreased with the curing age, which is in accordance with the air void shrinkage. The correlation between the early-age strength development and hardened dynamic modulus with the ultrasonic parameters was also evaluated. The existence of clustered air voids in the Interfacial Transition Zone (ITZ area was found to cause severe compressive strength loss. The results indicated that this developed ultrasonic technique has potential in air void size distribution measurement, and demonstrated the influence of air void structure evolution on concrete properties during both early-age and hardened stages.
Fraley, R Chris; Vazire, Simine
2014-01-01
The authors evaluate the quality of research reported in major journals in social-personality psychology by ranking those journals with respect to their N-pact Factors (NF)-the statistical power of the empirical studies they publish to detect typical effect sizes. Power is a particularly important attribute for evaluating research quality because, relative to studies that have low power, studies that have high power are more likely to (a) to provide accurate estimates of effects, (b) to produce literatures with low false positive rates, and (c) to lead to replicable findings. The authors show that the average sample size in social-personality research is 104 and that the power to detect the typical effect size in the field is approximately 50%. Moreover, they show that there is considerable variation among journals in sample sizes and power of the studies they publish, with some journals consistently publishing higher power studies than others. The authors hope that these rankings will be of use to authors who are choosing where to submit their best work, provide hiring and promotion committees with a superior way of quantifying journal quality, and encourage competition among journals to improve their NF rankings.
A Rounding by Sampling Approach to the Minimum Size k-Arc Connected Subgraph Problem
Laekhanukit, Bundit; Singh, Mohit
2012-01-01
In the k-arc connected subgraph problem, we are given a directed graph G and an integer k and the goal is the find a subgraph of minimum cost such that there are at least k-arc disjoint paths between any pair of vertices. We give a simple (1 + 1/k)-approximation to the unweighted variant of the problem, where all arcs of G have the same cost. This improves on the 1 + 2/k approximation of Gabow et al. [GGTW09]. Similar to the 2-approximation algorithm for this problem [FJ81], our algorithm simply takes the union of a k in-arborescence and a k out-arborescence. The main difference is in the selection of the two arborescences. Here, inspired by the recent applications of the rounding by sampling method (see e.g. [AGM+ 10, MOS11, OSS11, AKS12]), we select the arborescences randomly by sampling from a distribution on unions of k arborescences that is defined based on an extreme point solution of the linear programming relaxation of the problem. In the analysis, we crucially utilize the sparsity property of the ext...
RNA Profiling for Biomarker Discovery: Practical Considerations for Limiting Sample Sizes
Danny J. Kelly
2005-01-01
Full Text Available We have compared microarray data generated on Affymetrix™ chips from standard (8 micrograms or low (100 nanograms amounts of total RNA. We evaluated the gene signals and gene fold-change estimates obtained from the two methods and validated a subset of the results by real time, polymerase chain reaction assays. The correlation of low RNA derived gene signals to gene signals obtained from standard RNA was poor for less to moderately abundant genes. Genes with high abundance showed better correlation in signals between the two methods. The signal correlation between the low RNA and standard RNA methods was improved by including a reference sample in the microarray analysis. In contrast, the fold-change estimates for genes were better correlated between the two methods regardless of the magnitude of gene signals. A reference sample based method is suggested for studies that would end up comparing gene signal data from a combination of low and standard RNA templates; no such referencing appears to be necessary when comparing fold-changes of gene expression between standard and low template reactions.
Shrinkage-based diagonal Hotelling’s tests for high-dimensional small sample size data
Dong, Kai
2015-09-16
DNA sequencing techniques bring novel tools and also statistical challenges to genetic research. In addition to detecting differentially expressed genes, testing the significance of gene sets or pathway analysis has been recognized as an equally important problem. Owing to the “large pp small nn” paradigm, the traditional Hotelling’s T2T2 test suffers from the singularity problem and therefore is not valid in this setting. In this paper, we propose a shrinkage-based diagonal Hotelling’s test for both one-sample and two-sample cases. We also suggest several different ways to derive the approximate null distribution under different scenarios of pp and nn for our proposed shrinkage-based test. Simulation studies show that the proposed method performs comparably to existing competitors when nn is moderate or large, but it is better when nn is small. In addition, we analyze four gene expression data sets and they demonstrate the advantage of our proposed shrinkage-based diagonal Hotelling’s test.
Dealing with large sample sizes: comparison of a new one spot dot blot method to western blot.
Putra, Sulistyo Emantoko Dwi; Tsuprykov, Oleg; Von Websky, Karoline; Ritter, Teresa; Reichetzeder, Christoph; Hocher, Berthold
2014-01-01
Western blot is the gold standard method to determine individual protein expression levels. However, western blot is technically difficult to perform in large sample sizes because it is a time consuming and labor intensive process. Dot blot is often used instead when dealing with large sample sizes, but the main disadvantage of the existing dot blot techniques, is the absence of signal normalization to a housekeeping protein. In this study we established a one dot two development signals (ODTDS) dot blot method employing two different signal development systems. The first signal from the protein of interest was detected by horseradish peroxidase (HRP). The second signal, detecting the housekeeping protein, was obtained by using alkaline phosphatase (AP). Inter-assay results variations within ODTDS dot blot and western blot and intra-assay variations between both methods were low (1.04-5.71%) as assessed by coefficient of variation. ODTDS dot blot technique can be used instead of western blot when dealing with large sample sizes without a reduction in results accuracy.
Jamshid Jamali
2017-01-01
Full Text Available Evaluating measurement equivalence (also known as differential item functioning (DIF is an important part of the process of validating psychometric questionnaires. This study aimed at evaluating the multiple indicators multiple causes (MIMIC model for DIF detection when latent construct distribution is nonnormal and the focal group sample size is small. In this simulation-based study, Type I error rates and power of MIMIC model for detecting uniform-DIF were investigated under different combinations of reference to focal group sample size ratio, magnitude of the uniform-DIF effect, scale length, the number of response categories, and latent trait distribution. Moderate and high skewness in the latent trait distribution led to a decrease of 0.33% and 0.47% power of MIMIC model for detecting uniform-DIF, respectively. The findings indicated that, by increasing the scale length, the number of response categories and magnitude DIF improved the power of MIMIC model, by 3.47%, 4.83%, and 20.35%, respectively; it also decreased Type I error of MIMIC approach by 2.81%, 5.66%, and 0.04%, respectively. This study revealed that power of MIMIC model was at an acceptable level when latent trait distributions were skewed. However, empirical Type I error rate was slightly greater than nominal significance level. Consequently, the MIMIC was recommended for detection of uniform-DIF when latent construct distribution is nonnormal and the focal group sample size is small.
Distribution of human waste samples in relation to sizing waste processing in space
Parker, Dick; Gallagher, S. K.
1992-01-01
Human waste processing for closed ecological life support systems (CELSS) in space requires that there be an accurate knowledge of the quantity of wastes produced. Because initial CELSS will be handling relatively few individuals, it is important to know the variation that exists in the production of wastes rather than relying upon mean values that could result in undersizing equipment for a specific crew. On the other hand, because of the costs of orbiting equipment, it is important to design the equipment with a minimum of excess capacity because of the weight that extra capacity represents. A considerable quantity of information that had been independently gathered on waste production was examined in order to obtain estimates of equipment sizing requirements for handling waste loads from crews of 2 to 20 individuals. The recommended design for a crew of 8 should hold 34.5 liters per day (4315 ml/person/day) for urine and stool water and a little more than 1.25 kg per day (154 g/person/day) of human waste solids and sanitary supplies.
Prognostic stratification of ulcerated melanoma
Bønnelykke-Behrndtz, Marie L; Schmidt, Henrik; Christensen, Ib J
2014-01-01
OBJECTIVES: For patients with melanoma, ulceration is an important prognostic marker and interestingly also a predictive marker for the response of adjuvant interferon. A consensual definition and accurate assessment of ulceration are therefore crucial for proper staging and clinical management. We...... stratification of ulcerated lesions. METHODS: From H&E-stained sections, the status (presence vs absence), extent (percentage of the total tumor length), and type (infiltrative vs attenuative) of ulceration and epidermal involvement were evaluated from 385 patients with cutaneous melanoma. RESULTS: The presence...... of ulceration (hazard ratio [HR], 1.83), an attenuative type of ulceration (HR, 3.02), and excessive ulceration (HR, 3.57) were independent predictors of poor melanoma-specific survival. Further subdivision of minimal/moderate ulceration showed independent prognostic value only for lesions with epidermal...
The effect of sample size on fresh plasma thromboplastin ISI determination
Poller, L; Van Den Besselaar, A M; Jespersen, J;
1999-01-01
The possibility of reduction of numbers of fresh coumarin and normal plasmas has been studied in a multicentre manual prothrombin (PT) calibration of high international sensitivity index (ISI) rabbit and low ISI human reference thromboplastins at 14 laboratories. The number of calibrant plasmas...... was reduced progressively by a computer program which generated random numbers to provide 1000 different selections for each reduced sample at each participant laboratory. Results were compared with those of the full set of 20 normal and 60 coumarin plasma calibrations. With the human reagent, 20 coumarins...... and seven normals still achieved the W.H.O. precision limit (3% CV of the slope), but with the rabbit reagent reduction coumarins with 17 normal plasmas led to unacceptable CV. Little reduction of numbers from the full set of 80 fresh plasmas appears advisable. For maximum confidence, when calibrating...
V. Indira
2015-03-01
Full Text Available Hydraulic brake in automobile engineering is considered to be one of the important components. Condition monitoring and fault diagnosis of such a component is very essential for safety of passengers, vehicles and to minimize the unexpected maintenance time. Vibration based machine learning approach for condition monitoring of hydraulic brake system is gaining momentum. Training and testing the classifier are two important activities in the process of feature classification. This study proposes a systematic statistical method called power analysis to find the minimum number of samples required to train the classifier with statistical stability so as to get good classification accuracy. Descriptive statistical features have been used and the more contributing features have been selected by using C4.5 decision tree algorithm. The results of power analysis have also been verified using a decision tree algorithm namely, C4.5.
Historical Studies of Social Mobility and Stratification
Leeuwen, Marco H.D. van; Maas, Ineke
2010-01-01
This review discusses historical studies of social mobility and stratification. The focus is on changes in social inequality and mobility in past societies and their determinants. It discusses major historical sources, approaches, and results in the fields of social stratification (ranks and classes
Historical Studies of Social Mobility and Stratification
Leeuwen, Marco H.D. van; Maas, Ineke
2010-01-01
This review discusses historical studies of social mobility and stratification. The focus is on changes in social inequality and mobility in past societies and their determinants. It discusses major historical sources, approaches, and results in the fields of social stratification (ranks and classes
Dombrowski, Kirk; Khan, Bilal; Wendel, Travis; McLean, Katherine; Misshula, Evan; Curtis, Ric
2012-12-01
As part of a recent study of the dynamics of the retail market for methamphetamine use in New York City, we used network sampling methods to estimate the size of the total networked population. This process involved sampling from respondents' list of co-use contacts, which in turn became the basis for capture-recapture estimation. Recapture sampling was based on links to other respondents derived from demographic and "telefunken" matching procedures-the latter being an anonymized version of telephone number matching. This paper describes the matching process used to discover the links between the solicited contacts and project respondents, the capture-recapture calculation, the estimation of "false matches", and the development of confidence intervals for the final population estimates. A final population of 12,229 was estimated, with a range of 8235 - 23,750. The techniques described here have the special virtue of deriving an estimate for a hidden population while retaining respondent anonymity and the anonymity of network alters, but likely require larger sample size than the 132 persons interviewed to attain acceptable confidence levels for the estimate.
Küme, Tuncay; Şişman, Ali Rıza; Solak, Ahmet; Tuğlu, Birsen; Çinkooğlu, Burcu; Çoker, Canan
2012-01-01
Introductıon: We evaluated the effect of different syringe volume, needle size and sample volume on blood gas analysis in syringes washed with heparin. Materials and methods: In this multi-step experimental study, percent dilution ratios (PDRs) and final heparin concentrations (FHCs) were calculated by gravimetric method for determining the effect of syringe volume (1, 2, 5 and 10 mL), needle size (20, 21, 22, 25 and 26 G) and sample volume (0.5, 1, 2, 5 and 10 mL). The effect of different PDRs and FHCs on blood gas and electrolyte parameters were determined. The erroneous results from nonstandardized sampling were evaluated according to RiliBAK’s TEa. Results: The increase of PDRs and FHCs was associated with the decrease of syringe volume, the increase of needle size and the decrease of sample volume: from 2.0% and 100 IU/mL in 10 mL-syringe to 7.0% and 351 IU/mL in 1 mL-syringe; from 4.9% and 245 IU/mL in 26G to 7.6% and 380 IU/mL in 20 G with combined 1 mL syringe; from 2.0% and 100 IU/mL in full-filled sample to 34% and 1675 IU/mL in 0.5 mL suctioned sample into 10 mL-syringe. There was no statistical difference in pH; but the percent decreasing in pCO2, K+, iCa2+, iMg2+; the percent increasing in pO2 and Na+ were statistical significance compared to samples full-filled in syringes. The all changes in pH and pO2 were acceptable; but the changes in pCO2, Na+, K+ and iCa2+ were unacceptable according to TEa limits except fullfilled-syringes. Conclusions: The changes in PDRs and FHCs due nonstandardized sampling in syringe washed with liquid heparin give rise to erroneous test results for pCO2 and electrolytes. PMID:22838185
The design of high-temperature thermal conductivity measurements apparatus for thin sample size
Hadi Syamsul
2017-01-01
Full Text Available This study presents the designing, constructing and validating processes of thermal conductivity apparatus using steady-state heat-transfer techniques with the capability of testing a material at high temperatures. This design is an improvement from ASTM D5470 standard where meter-bars with the equal cross-sectional area were used to extrapolate surface temperature and measure heat transfer across a sample. There were two meter-bars in apparatus where each was placed three thermocouples. This Apparatus using a heater with a power of 1,000 watts, and cooling water to stable condition. The pressure applied was 3.4 MPa at the cross-sectional area of 113.09 mm2 meter-bar and thermal grease to minimized interfacial thermal contact resistance. To determine the performance, the validating process proceeded by comparing the results with thermal conductivity obtained by THB 500 made by LINSEIS. The tests showed the thermal conductivity of the stainless steel and bronze are 15.28 Wm-1K-1 and 38.01 Wm-1K-1 with a difference of test apparatus THB 500 are −2.55% and 2.49%. Furthermore, this apparatus has the capability to measure the thermal conductivity of the material to a temperature of 400°C where the results for the thermal conductivity of stainless steel is 19.21 Wm-1K-1 and the difference was 7.93%.
Bayer, Immanuel; Groth, Philip; Schneckener, Sebastian
2013-01-01
Model-based prediction is dependent on many choices ranging from the sample collection and prediction endpoint to the choice of algorithm and its parameters. Here we studied the effects of such choices, exemplified by predicting sensitivity (as IC50) of cancer cell lines towards a variety of compounds. For this, we used three independent sample collections and applied several machine learning algorithms for predicting a variety of endpoints for drug response. We compared all possible models for combinations of sample collections, algorithm, drug, and labeling to an identically generated null model. The predictability of treatment effects varies among compounds, i.e. response could be predicted for some but not for all. The choice of sample collection plays a major role towards lowering the prediction error, as does sample size. However, we found that no algorithm was able to consistently outperform the other and there was no significant difference between regression and two- or three class predictors in this experimental setting. These results indicate that response-modeling projects should direct efforts mainly towards sample collection and data quality, rather than method adjustment.
To discuss different calculation methods of sample size%样本含量估算方法探讨
喻宁芳
2014-01-01
目的：介绍和比较医学实验设计中不同的样本含量估算方法。方法：以PI3K抑制剂对小鼠气道炎症影响的实验研究*为例运用不同方法计算样本含量。结果：①公式法计算需12例②PASS软件Simple法计算需10例③Stata软件计算需8例，验算其检验效能：1-β>0.9结论：3种不同方法估算的样本含量都是合理有效的，实验研究人员可以以多种计算结果为依据，分析实验研究性质，综合考虑研究成本、可行性与伦理学要求对样本含量的影响确定合适的样本数。%Objective: To introduce and compare different calculation Methods of sample size in experiment design.Methods: As an example of PI3K inhibitor reduces respiratory tract inflammation in a murine model of Asthma.Results: In method of formula,12;in PASS software,8;in Stata software, 10.1-β>0.9.Conclusion: Proper analysis of the nature of research design,setting the correct parameters,Based on a variety of calculations to estimate the sample size, and then considering the research costs, feasibility and ethics requirements impact on sample size, and ultimately determine the most appropriate number of samples.
Sebastian Wilhelm
2015-12-01
Full Text Available The production of silica is performed by mixing an inorganic, silicate-based precursor and an acid. Monomeric silicic acid forms and polymerizes to amorphous silica particles. Both further polymerization and agglomeration of the particles lead to a gel network. Since polymerization continues after gelation, the gel network consolidates. This rather slow process is known as “natural syneresis” and strongly influences the product properties (e.g., agglomerate size, porosity or internal surface. “Enforced syneresis” is the superposition of natural syneresis with a mechanical, external force. Enforced syneresis may be used either for analytical or preparative purposes. Hereby, two open key aspects are of particular interest. On the one hand, the question arises whether natural and enforced syneresis are analogous processes with respect to their dependence on the process parameters: pH, temperature and sample size. On the other hand, a method is desirable that allows for correlating natural and enforced syneresis behavior. We can show that the pH-, temperature- and sample size-dependency of natural and enforced syneresis are indeed analogous. It is possible to predict natural syneresis using a correlative model. We found that our model predicts maximum volume shrinkages between 19% and 30% in comparison to measured values of 20% for natural syneresis.
Effects of strong stratification on equatorward dynamo wave propagation
Käpylä, Petri J; Cole, Elizabeth; Warnecke, Jörn; Brandenburg, Axel
2013-01-01
We present results from simulations of rotating magnetized turbulent convection in spherical wedge geometry representing parts of the latitudinal and longitudinal extents of a star. Here we consider a set of runs for which the density stratification is varied, keeping the Reynolds and Coriolis numbers at similar values. In the case of weak stratification we find quasi-steady solutions for moderate rotation and oscillatory dynamos with poleward migration of activity belts for more rapid rotation. For stronger stratification a similar transition as a function of the Coriolis number is found, but with an equatorward migrating branch near the equator. We test the domain size dependence of our results for a rapidly rotating run with equatorward migration by varying the longitudinal extent of our wedge. The energy of the axisymmetric mean magnetic field decreases as the domain size increases and we find that an m=1 mode is excited for a full 2pi phi-extent, reminiscent of the field configurations deduced from obser...
Chang, Ying-Jie; Shih, Yang-Hsin; Su, Chiu-Hun; Ho, Han-Chen
2017-01-15
Due to the widespread application of engineered nanoparticles, their potential risk to ecosystems and human health is of growing concern. Silver nanoparticles (Ag NPs) are one of the most extensively produced NPs. Thus, this study aims to develop a method to detect Ag NPs in different aquatic systems. In complex media, three emerging techniques are compared, including hydrodynamic chromatography (HDC), asymmetric flow field flow fractionation (AF4) and single particle inductively coupled plasma-mass spectrometry (SP-ICP-MS). The pre-treatment procedure of centrifugation is evaluated. HDC can estimate the Ag NP sizes, which were consistent with the results obtained from DLS. AF4 can also determine the size of Ag NPs but with lower recoveries, which could result from the interactions between Ag NPs and the working membrane. For the SP-ICP-MS, both the particle size and concentrations can be determined with high Ag NP recoveries. The particle size resulting from SP-ICP-MS also corresponded to the transmission electron microscopy observation (p>0.05). Therefore, HDC and SP-ICP-MS are recommended for environmental analysis of the samples after our established pre-treatment process. The findings of this study propose a preliminary technique to more accurately determine the Ag NPs in aquatic environments and to use this knowledge to evaluate the environmental impact of manufactured NPs. Copyright © 2016 Elsevier B.V. All rights reserved.
Michael B.C. Khoo
2013-11-01
Full Text Available The double sampling (DS X bar chart, one of the most widely-used charting methods, is superior for detecting small and moderate shifts in the process mean. In a right skewed run length distribution, the median run length (MRL provides a more credible representation of the central tendency than the average run length (ARL, as the mean is greater than the median. In this paper, therefore, MRL is used as the performance criterion instead of the traditional ARL. Generally, the performance of the DS X bar chart is investigated under the assumption of known process parameters. In practice, these parameters are usually estimated from an in-control reference Phase-I dataset. Since the performance of the DS X bar chart is significantly affected by estimation errors, we study the effects of parameter estimation on the MRL-based DS X bar chart when the in-control average sample size is minimised. This study reveals that more than 80 samples are required for the MRL-based DS X bar chart with estimated parameters to perform more favourably than the corresponding chart with known parameters.
Arbab, A
2014-10-01
The rice stem borer, Chilo suppressalis (Walker), feeds almost exclusively in paddy fields in most regions of the world. The study of its spatial distribution is fundamental for designing correct control strategies, improving sampling procedures, and adopting precise agricultural techniques. Field experiments were conducted during 2011 and 2012 to estimate the spatial distribution pattern of the overwintering larvae. Data were analyzed using five distribution indices and two regression models (Taylor and Iwao). All of the indices and Taylor's model indicated random spatial distribution pattern of the rice stem borer overwintering larvae. Iwao's patchiness regression was inappropriate for our data as shown by the non-homogeneity of variance, whereas Taylor's power law fitted the data well. The coefficients of Taylor's power law for a combined 2 years of data were a = -0.1118, b = 0.9202 ± 0.02, and r (2) = 96.81. Taylor's power law parameters were used to compute minimum sample size needed to estimate populations at three fixed precision levels, 5, 10, and 25% at 0.05 probabilities. Results based on this equation parameters suggesting that minimum sample sizes needed for a precision level of 0.25 were 74 and 20 rice stubble for rice stem borer larvae when the average larvae is near 0.10 and 0.20 larvae per rice stubble, respectively.
Hagell, Peter; Westergren, Albert
Sample size is a major factor in statistical null hypothesis testing, which is the basis for many approaches to testing Rasch model fit. Few sample size recommendations for testing fit to the Rasch model concern the Rasch Unidimensional Measurement Models (RUMM) software, which features chi-square and ANOVA/F-ratio based fit statistics, including Bonferroni and algebraic sample size adjustments. This paper explores the occurrence of Type I errors with RUMM fit statistics, and the effects of algebraic sample size adjustments. Data with simulated Rasch model fitting 25-item dichotomous scales and sample sizes ranging from N = 50 to N = 2500 were analysed with and without algebraically adjusted sample sizes. Results suggest the occurrence of Type I errors with N less then or equal to 500, and that Bonferroni correction as well as downward algebraic sample size adjustment are useful to avoid such errors, whereas upward adjustment of smaller samples falsely signal misfit. Our observations suggest that sample sizes around N = 250 to N = 500 may provide a good balance for the statistical interpretation of the RUMM fit statistics studied here with respect to Type I errors and under the assumption of Rasch model fit within the examined frame of reference (i.e., about 25 item parameters well targeted to the sample).
Dahlin, Jakob; Spanne, Mårten; Karlsson, Daniel; Dalene, Marianne; Skarping, Gunnar
2008-07-01
Isocyanates in the workplace atmosphere are typically present both in gas and particle phase. The health effects of exposure to isocyanates in gas phase and different particle size fractions are likely to be different due to their ability to reach different parts in the respiratory system. To reveal more details regarding the exposure to isocyanate aerosols, a denuder-impactor (DI) sampler for airborne isocyanates was designed. The sampler consists of a channel-plate denuder for collection of gaseous isocyanates, in series with three-cascade impactor stages with cut-off diameters (d(50)) of 2.5, 1.0 and 0.5 mum. An end filter was connected in series after the impactor for collection of particles smaller than 0.5 mum. The denuder, impactor plates and the end filter were impregnated with a mixture of di-n-butylamine (DBA) and acetic acid for derivatization of the isocyanates. During sampling, the reagent on the impactor plates and the end filter is continuously refreshed, due to the DBA release from the impregnated denuder plates. This secures efficient derivatization of all isocyanate particles. The airflow through the sampler was 5 l min(-1). After sampling, the samples containing the different size fractions were analyzed using liquid chromatography-mass spectrometry (LC-MS)/MS. The DBA impregnation was stable in the sampler for at least 1 week. After sampling, the DBA derivatives were stable for at least 3 weeks. Air sampling was performed in a test chamber (300 l). Isocyanate aerosols studied were thermal degradation products of different polyurethane polymers, spraying of isocyanate coating compounds and pure gas-phase isocyanates. Sampling with impinger flasks, containing DBA in toluene, with a glass fiber filter in series was used as a reference method. The DI sampler showed good compliance with the reference method, regarding total air levels. For the different aerosols studied, vast differences were revealed in the distribution of isocyanate in gas and
Vasiliu, Daniel; Clamons, Samuel; McDonough, Molly; Rabe, Brian; Saha, Margaret
2015-01-01
Global gene expression analysis using microarrays and, more recently, RNA-seq, has allowed investigators to understand biological processes at a system level. However, the identification of differentially expressed genes in experiments with small sample size, high dimensionality, and high variance remains challenging, limiting the usability of these tens of thousands of publicly available, and possibly many more unpublished, gene expression datasets. We propose a novel variable selection algorithm for ultra-low-n microarray studies using generalized linear model-based variable selection with a penalized binomial regression algorithm called penalized Euclidean distance (PED). Our method uses PED to build a classifier on the experimental data to rank genes by importance. In place of cross-validation, which is required by most similar methods but not reliable for experiments with small sample size, we use a simulation-based approach to additively build a list of differentially expressed genes from the rank-ordered list. Our simulation-based approach maintains a low false discovery rate while maximizing the number of differentially expressed genes identified, a feature critical for downstream pathway analysis. We apply our method to microarray data from an experiment perturbing the Notch signaling pathway in Xenopus laevis embryos. This dataset was chosen because it showed very little differential expression according to limma, a powerful and widely-used method for microarray analysis. Our method was able to detect a significant number of differentially expressed genes in this dataset and suggest future directions for investigation. Our method is easily adaptable for analysis of data from RNA-seq and other global expression experiments with low sample size and high dimensionality.
Daniel Vasiliu
Full Text Available Global gene expression analysis using microarrays and, more recently, RNA-seq, has allowed investigators to understand biological processes at a system level. However, the identification of differentially expressed genes in experiments with small sample size, high dimensionality, and high variance remains challenging, limiting the usability of these tens of thousands of publicly available, and possibly many more unpublished, gene expression datasets. We propose a novel variable selection algorithm for ultra-low-n microarray studies using generalized linear model-based variable selection with a penalized binomial regression algorithm called penalized Euclidean distance (PED. Our method uses PED to build a classifier on the experimental data to rank genes by importance. In place of cross-validation, which is required by most similar methods but not reliable for experiments with small sample size, we use a simulation-based approach to additively build a list of differentially expressed genes from the rank-ordered list. Our simulation-based approach maintains a low false discovery rate while maximizing the number of differentially expressed genes identified, a feature critical for downstream pathway analysis. We apply our method to microarray data from an experiment perturbing the Notch signaling pathway in Xenopus laevis embryos. This dataset was chosen because it showed very little differential expression according to limma, a powerful and widely-used method for microarray analysis. Our method was able to detect a significant number of differentially expressed genes in this dataset and suggest future directions for investigation. Our method is easily adaptable for analysis of data from RNA-seq and other global expression experiments with low sample size and high dimensionality.
Gardi, J E; Nyengaard, J R; Gundersen, H J G
2008-03-01
The proportionator is a novel and radically different approach to sampling with microscopes based on the well-known statistical theory (probability proportional to size-PPS sampling). It uses automatic image analysis, with a large range of options, to assign to every field of view in the section a weight proportional to some characteristic of the structure under study. A typical and very simple example, examined here, is the amount of color characteristic for the structure, marked with a stain with known properties. The color may be specific or not. In the recorded list of weights in all fields, the desired number of fields is sampled automatically with probability proportional to the weight and presented to the expert observer. Using any known stereological probe and estimator, the correct count in these fields leads to a simple, unbiased estimate of the total amount of structure in the sections examined, which in turn leads to any of the known stereological estimates including size distributions and spatial distributions. The unbiasedness is not a function of the assumed relation between the weight and the structure, which is in practice always a biased relation from a stereological (integral geometric) point of view. The efficiency of the proportionator depends, however, directly on this relation to be positive. The sampling and estimation procedure is simulated in sections with characteristics and various kinds of noises in possibly realistic ranges. In all cases examined, the proportionator is 2-15-fold more efficient than the common systematic, uniformly random sampling. The simulations also indicate that the lack of a simple predictor of the coefficient of error (CE) due to field-to-field variation is a more severe problem for uniform sampling strategies than anticipated. Because of its entirely different sampling strategy, based on known but non-uniform sampling probabilities, the proportionator for the first time allows the real CE at the section level to
Stratification, Hypothesis Testing, and Clinical Trial Simulation in Pediatric Drug Development
McMahon, Ann W.; Watt, Kevin; Wang, Jian; Green, Dionna; Tiwari, Ram; Burckart, Gilbert J.
2016-01-01
Background Pediatric drug development is plagued by small sample sizes, unvalidated clinical endpoints, and limited studies. Objectives The objective of this study was to determine whether age stratification within the pediatric population could be used to (1) assess response to a pharmacologic intervention and to (2) design future trials based upon published stratified disease data using clinical trial simulation (CTS). Methods Data available from the literature for Kawasaki disease (KD) was used in the model. Age-stratified CTS for a theoretical new drug was conducted. Results Population-specific differences due to age might affect trial success if not taken into account. CTS predicted inflammatory indices, and inclusion cutoff significantly altered the trial outcome. Finally, altered pharmacokinetics/pharmacodynamics in varying age groups of KD patients may alter drug exposure and response. Conclusions If assumptions regarding a pediatric disease process, such as KD, do not include age stratification with inclusion or response, then the wrong decision could result with regard to age-appropriateness or approval of a drug.
Hillson, Roger; Alejandre, Joel D; Jacobsen, Kathryn H; Ansumana, Rashid; Bockarie, Alfred S; Bangura, Umaru; Lamin, Joseph M; Stenger, David A
2015-01-01
There is a need for better estimators of population size in places that have undergone rapid growth and where collection of census data is difficult. We explored simulated estimates of urban population based on survey data from Bo, Sierra Leone, using two approaches: (1) stratified sampling from across 20 neighborhoods and (2) stratified single-stage cluster sampling of only four randomly-sampled neighborhoods. The stratification variables evaluated were (a) occupants per individual residence, (b) occupants per neighborhood, and (c) residential structures per neighborhood. For method (1), stratification variable (a) yielded the most accurate re-estimate of the current total population. Stratification variable (c), which can be estimated from aerial photography and zoning type verification, and variable (b), which could be ascertained by surveying a limited number of households, increased the accuracy of method (2). Small household-level surveys with appropriate sampling methods can yield reasonably accurate estimations of urban populations.
Optimization of stratification scheme for a fishery-independent survey with multiple objectives
XU Binduo; REN Yiping; CHEN Yong; XUE Ying; ZHANG Chongliang; WAN Rong
2015-01-01
Fishery-independent surveys are often used for collecting high quality biological and ecological data to support fisheries management. A careful optimization of fishery-independent survey design is necessary to improve the precision of survey estimates with cost-effective sampling efforts. We developed a simulation approach to evaluate and optimize the stratification scheme for a fishery-independent survey with multiple goals including estimation of abundance indices of individual species and species diversity indices. We compared the performances of the sampling designs with different stratification schemes for different goals over different months. Gains in precision of survey estimates from the stratification schemes were acquired compared to simple random sampling design for most indices. The stratification scheme with five strata performed the best. This study showed that the loss of precision of survey estimates due to the reduction of sampling efforts could be compensated by improved stratification schemes, which would reduce the cost and negative impacts of survey trawling on those species with low abundance in the fishery-independent survey. This study also suggests that optimization of a survey design differed with different survey objectives. A post-survey analysis can improve the stratification scheme of fishery-independent survey designs.
The inefficiency of re-weighted sampling and the curse of system size in high order path integration
Ceriotti, Michele; Riordan, Oliver; Manolopoulos, David E
2011-01-01
Computing averages over a target probability density by statistical re-weighting of a set of samples with a different distribution is a strategy which is commonly adopted in fields as diverse as atomistic simulation and finance. Here we present a very general analysis of the accuracy and efficiency of this approach, highlighting some of its weaknesses. We then give an example of how our results can be used, specifically to assess the feasibility of high-order path integral methods. We demonstrate that the most promising of these techniques -- which is based on re-weighted sampling -- is bound to fail as the size of the system is increased, because of the exponential growth of the statistical uncertainty in the re-weighted average.
Thermal Stratification in Vertical Mantle Tanks
Knudsen, Søren; Furbo, Simon
2001-01-01
It is well known that it is important to have a high degree of thermal stratification in the hot water storage tank to achieve a high thermal performance of SDHW systems. This study is concentrated on thermal stratification in vertical mantle tanks. Experiments based on typical operation conditions...... are carried out to investigate how the thermal stratification is affected by different placements of the mantle inlet. The heat transfer between the solar collector fluid in the mantle and the domestic water in the inner tank is analysed by CFD-simulations. Furthermore, the flow pattern in the vertical mantle...... tank is investigated....
Thompson, Steven K
2012-01-01
Praise for the Second Edition "This book has never had a competitor. It is the only book that takes a broad approach to sampling . . . any good personal statistics library should include a copy of this book." —Technometrics "Well-written . . . an excellent book on an important subject. Highly recommended." —Choice "An ideal reference for scientific researchers and other professionals who use sampling." —Zentralblatt Math Features new developments in the field combined with all aspects of obtaining, interpreting, and using sample data Sampling provides an up-to-date treat
de Winter, Joost C F; Gosling, Samuel D; Potter, Jeff
2016-09-01
The Pearson product–moment correlation coefficient (rp) and the Spearman rank correlation coefficient (rs) are widely used in psychological research. We compare rp and rs on 3 criteria: variability, bias with respect to the population value, and robustness to an outlier. Using simulations across low (N = 5) to high (N = 1,000) sample sizes we show that, for normally distributed variables, rp and rs have similar expected values but rs is more variable, especially when the correlation is strong. However, when the variables have high kurtosis, rp is more variable than rs. Next, we conducted a sampling study of a psychometric dataset featuring symmetrically distributed data with light tails, and of 2 Likert-type survey datasets, 1 with light-tailed and the other with heavy-tailed distributions. Consistent with the simulations, rp had lower variability than rs in the psychometric dataset. In the survey datasets with heavy-tailed variables in particular, rs had lower variability than rp, and often corresponded more accurately to the population Pearson correlation coefficient (Rp) than rp did. The simulations and the sampling studies showed that variability in terms of standard deviations can be reduced by about 20% by choosing rs instead of rp. In comparison, increasing the sample size by a factor of 2 results in a 41% reduction of the standard deviations of rs and rp. In conclusion, rp is suitable for light-tailed distributions, whereas rs is preferable when variables feature heavy-tailed distributions or when outliers are present, as is often the case in psychological research.
LI Xiao-ling; LU Yong-gen; LI Jin-quan; Xu Hai-ming; Muhammad Qasim SHAHID
2011-01-01
The development of a core collection could enhance the utilization of germplasm collections in crop improvement programs and simplify their management.Selection of an appropriate sampling strategy is an important prerequisite to construct a core collection with appropriate size in order to adequately represent the genetic spectrum and maximally capture the genetic diversity in available crop collections.The present study was initiated to construct nested core collections to determine the appropriate sample size to represent the genetic diversity of rice landrace collection based on 15 quantitative traits and 34 qualitative traits of 2 262 rice accessions.The results showed that 50-225 nested core collections,whose sampling rate was 2.2％-9.9％,were sufficient to maintain the maximum genetic diversity of the initial collections.Of these,150 accessions (6.6％) could capture the maximal genetic diversity of the initial collection.Three data types,i.e.qualitative traits (QT1),quantitative traits (QT2) and integrated qualitative and quantitative traits (QTT),were compared for their efficiency in constructing core collections based on the weighted pair-group average method combined with stepwise clustering and preferred sampling on adjusted Euclidean distances.Every combining scheme constructed eight rice core collections (225,200,175,150,125,100,75 and 50).The results showed that the QTT data was the best in constructing a core collection as indicated by the genetic diversity of core collections.A core collection constructed only on the information of QT1 could not represent the initial collection effectively.QTT should be used together to construct a productive core collection.
Particle stratification and penetration of a linear vibrating screen by the discrete element method
Xiao Jianzhang; Tong Xin
2012-01-01
A simulation of stratification and penetration was performed over a range of structural parameters that included screen width,aperture size,inclination angle,and wire diameter.The discrete element method (DEM) was used for the simulations.The terms stratification and penetration are defined and the change in fine particle concentration is discussed.Mathematical models relating fine particle ratio to time are established using the least squares method.The effect of structural parameters on fine particle ratio is analyzed.Stratification and penetration rate are discussed by considering the time derivative of the fine particle ratio.The conclusions are:an increase in inclination or wire diameter has a positive effect on particle stratifying; The optimal screen width is 40 mm for particle stratification; The inclination angle has a negative effect on the penetration; The effect of wire diameter and screen width on the penetration rate is negligible.
Cerit, Mahinur; Yücel, Cem; Göçün, Pınar Uyar; Poyraz, Aylar; Cerit, Ethem Turgay; Taneri, Ferit
2015-01-01
The aim of this study was to compare the diagnostic adequacy of thyroid samples obtained by aspiration or capillary biopsy techniques, with 22 or 27 gauge needles, and with or without on-site cytological analysis (OCA). Four hundred patients with thyroid nodules underwent ultrasound (US)-guided fine-needle biopsies. Patients were divided into eight groups according to needle size (22 vs. 27 gauge), biopsy technique (aspiration vs. capillary), and whether or not OCA was performed. Sample adequacy rates were calculated for each group and subgroups and compared using chi-square tests. When all nodes were evaluated (n = 400), the adequacy rate was significantly greater with the capillary than with the aspiration technique (97% vs. 91.5%, p = 0.032) and when OCA was than was not performed (97% vs. 91.5%, p = 0.032). When only solid nodules were evaluated (n = 205) the adequacy rate was also significantly greater with the capillary than with the aspiration technique (98.9% vs. 89.7%, p = 0.008) and when OCA was than was not performed (97.9% vs. 89.6%, p = 0.014). In contrast, the adequacy rate was similar for 22 and 27 gauge needles (94.2% vs. 93.1%, p = 0.733). Optimal results were obtained with the capillary technique and OCA. The capillary technique and OCA should be the preferred approach in thyroid nodule biopsy, optimising adequacy rates and patient comfort.
Development of Technologies for Early Detection and Stratification of Breast Cancer
2016-12-01
AWARD NUMBER: W81XWH-11-1-0814 TITLE: Development of Technologies for Early Detection and Stratification of Breast Cancer PRINCIPAL...Development of Technologies for Early Detection and 5a. CONTRACT NUMBER W81XWH-11-1-0814 Stratification of Breast Cancer 5b. GRANT NUMBER 5c...test can be implemented. We are also working to characterize breast cancer biopsy samples with single cell resolution to discover the nature of the
ECONOMIC STRATIFICATION AS A FACTOR IN THE AVAILABILITY OF DENTAL SERVICES
Kudinova Nadezhda Alekseevna
2013-02-01
Full Text Available Purpose to study the medical and social factors of availability of dental services among economic stratificated persons in the modern medical-demographic and socio-economic realities. Methodology historical, sociological, statistical. Results: Dental services as a sector of social production are highly cost. The state takes a large part of these expenditures. However, it is not possible to get high quality dental care (DC without the direct financial cost of patients. Author investigated the involvement of patients in the process of joint payment, developed a methodology for quantifying affordability (QA of DC for representatives of various segments of the population, based on a representative sample determined the total motivational area and the size of the values as a space in groups of dental patients, and the main factors that increase the QA DC. Practical implications public health and health care.
ECONOMIC STRATIFICATION AS A FACTOR IN THE AVAILABILITY OF DENTAL SERVICES
Надежда Алексеевна Кудинова
2013-04-01
Full Text Available Purpose to study the medical and social factors of availability of dental services among economic stratificated persons in the modern medical-demographic and socio-economic realities.Methodology historical, sociological, statistical. Results: Dental services as a sector of social production are highly cost. The state takes a large part of these expenditures. However, it is not possible to get high quality dental care (DC without the direct financial cost of patients. Author investigated the involvement of patients in the process of joint payment, developed a methodology for quantifying affordability (QA of DC for representatives of various segments of the population, based on a representative sample determined the total motivational area and the size of the values as a space in groups of dental patients, and the main factors that increase the QA DC.Practical implications public health and health care.DOI: http://dx.doi.org/10.12731/2218-7405-2013-2-23
François, Filip; Maenhaut, Willy; Colin, Jean-Louis; Losno, Remi; Schulz, Michael; Stahlschmidt, Thomas; Spokes, Lucinda; Jickells, Timothy
During an intercomparison field experiment, organized at the Atlantic coast station of Mace Head, Ireland, in April 1991, aerosol samples were collected by four research groups. A variety of samplers was used, combining both high- and low-volume devices, with different types of collection substrates: Hi-Vol Whatman 41 filter holders, single Nuclepore filters and stacked filter units, as well as PIXE cascade impactors. The samples were analyzed by each participating group, using in-house analytical techniques and procedures. The intercomparison of the daily concentrations for 15 elements, measured by two or more participants, revealed a good agreement for the low-volume samplers for the majority of the elements, but also indicated some specific analytical problems, owing to the very low concentrations of the non-sea-salt elements at the sampling site. With the Hi-Vol Whatman 41 filter sampler, on the other hand, much higher results were obtained in particular for the sea-salt and crustal elements. The discrepancy was dependent upon the wind speed and was attributed to a higher collection efficiency of the Hi-Vol sampler for the very coarse particles, as compared to the low-volume devices under high wind speed conditions. The elemental mass size distribution, as derived from parallel cascade impactor samplings by two groups, showed discrepancies in the submicrometer aerosol fraction, which were tentatively attributed to differences in stage cut-off diameters and/or to bounce-off or splintering effects on the quartz impactor slides used by one of the groups. However, the atmospheric concentrations (sums over all stages) were rather similar in the parallel impactor samples and were only slightly lower than those derived from stacked filter unit samples taken in parallel.
Sipola, Petri [Kuopio University Hospital, Department of Clinical Radiology, Kuopio (Finland); University of Eastern Finland, Institute of Clinical Medicine, Faculty of Health Sciences, Kuopio (Finland); Niemitukia, Lea H. [Kuopio University Hospital, Department of Clinical Radiology, Kuopio (Finland); Hyttinen, Mika M. [University of Eastern Finland, Institute of Biomedicine, Anatomy, Kuopio (Finland); Arokoski, Jari P.A. [Kuopio University Hospital, Department of Physical and Rehabilitation Medicine, Kuopio (Finland)
2011-04-15
To determine the number of participants required in controlled clinical trials investigating the progression of osteoarthritis (OA) of the hip as evaluated by the joint space width (JSW) on radiographs and to evaluate the reproducibility of the JSW measurement methods. Anteroposterior radiographs of hip were taken from 13 healthy volunteers and from 18 subjects with radiographic hip OA. The reproducibility of the JSW was determined from four segments using digital caliper measurements performed on film radiographs and using semiautomatic computerized image analysis of digitized images. Pearson correlation coefficient, coefficient of variability [CV (%)], and sample size values were calculated. It was found that 20 was a typical number of patients for a sufficiently powered study. The highest sample size was found in subjects with OA in the lateral segment. The reproducibility of the semiautomatic computerized method was not significantly better than the digital caliper method. The number of study subjects required to detect a significant joint space narrowing in follow-up studies is influenced by the baseline hip joint OA severity. The JSW measurements with computerized image analysis did not improve the reproducibility and thus performing JSW measurements with a digital caliper is acceptable. (orig.)
Mirante, Fátima; Alves, Célia; Pio, Casimiro; Pindado, Oscar; Perez, Rosa; Revuelta, M.a. Aranzazu; Artiñano, Begoña
2013-10-01
Madrid, the largest city of Spain, has some and unique air pollution problems, such as emissions from residential coal burning, a huge vehicle fleet and frequent African dust outbreaks, along with the lack of industrial emissions. The chemical composition of particulate matter (PM) was studied during summer and winter sampling campaigns, conducted in order to obtain size-segregated information at two different urban sites (roadside and urban background). PM was sampled with high volume cascade impactors, with 4 stages: 10-2.5, 2.5-1, 1-0.5 and extracted and organic compounds were identified and quantified by GC-MS. Alkanes, polycyclic aromatic hydrocarbons (PAHs), alcohols and fatty acids were chromatographically resolved. The PM1-2.5 was the fraction with the highest mass percentage of organics. Acids were the organic compounds that dominated all particle size fractions. Different organic compounds presented apparently different seasonal characteristics, reflecting distinct emission sources, such as vehicle exhausts and biogenic sources. The benzo[a]pyrene equivalent concentrations were lower than 1 ng m- 3. The estimated carcinogenic risk is low.
Alam, Murad; Rauf, Mutahir; Ali, Sana; Nodzenski, Michael; Minkis, Kira
2014-12-01
Dermatologic surgery is a fruitful research area that has spawned numerous randomized control trials (RCTs). To assess the quality of reporting of randomization, blinding, sample size, and power analysis in RCTs published in the journal Dermatologic Surgery. Randomized control trials published in Dermatologic Surgery between 1995 and 2012 were assessed regarding the quality of trial reporting. Data extraction performed independently by 2 data extractors. Dramatic increases in the numbers of RCTs in dermatologic surgery were noted in successive 5-year periods, from 39 in 1995 to 1999 to 66 in 2000 to 2004 and 131 in 2005 to 2009. The median Jadad score for articles from 1995 to 1999 was 1 and was 2 for articles since 2000. Subjects per study were 20 during 1995 to 1999, 25.5 from 2000 to 2004, and over 30 since 2005. Power analysis with sample size determination was reported in 0 articles during 1995 to 1999; greater than 13% of articles since 2005. Alpha level was specified for 37% of RCTs from 1995 to 1999 and 64% to 70% since 2005. During the last 20 years, the number of RCTs in Dermatologic Surgery has grown rapidly, almost doubling every 5 years, because the number of subjects per study has also increased and the quality of reporting has significantly improved.
Orth, Patrick; Zurakowski, David; Alini, Mauro; Cucchiarini, Magali; Madry, Henning
2013-11-01
Advanced tissue engineering approaches for articular cartilage repair in the knee joint rely on translational animal models. In these investigations, cartilage defects may be established either in one joint (unilateral design) or in both joints of the same animal (bilateral design). We hypothesized that a lower intraindividual variability following the bilateral strategy would reduce the number of required joints. Standardized osteochondral defects were created in the trochlear groove of 18 rabbits. In 12 animals, defects were produced unilaterally (unilateral design; n=12 defects), while defects were created bilaterally in 6 animals (bilateral design; n=12 defects). After 3 weeks, osteochondral repair was evaluated histologically applying an established grading system. Based on intra- and interindividual variabilities, required sample sizes for the detection of discrete differences in the histological score were determined for both study designs (α=0.05, β=0.20). Coefficients of variation (%CV) of the total histological score values were 1.9-fold increased following the unilateral design when compared with the bilateral approach (26 versus 14%CV). The resulting numbers of joints needed to treat were always higher for the unilateral design, resulting in an up to 3.9-fold increase in the required number of experimental animals. This effect was most pronounced for the detection of small-effect sizes and estimating large standard deviations. The data underline the possible benefit of bilateral study designs for the decrease of sample size requirements for certain investigations in articular cartilage research. These findings might also be transferred to other scoring systems, defect types, or translational animal models in the field of cartilage tissue engineering.
Effects of sample size on the second magnetization peak in Bi2Sr2CaCuO8+ at low temperatures
B Kalisky; A Shaulov; Y Yeshurun
2006-01-01
Effects of sample size on the second magnetization peak (SMP) in Bi2Sr2CaCuO8+ crystals are observed at low temperatures, above the temperature where the SMP totally disappears. In particular, the onset of the SMP shifts to lower fields as the sample size decreases - a result that could be interpreted as a size effect in the order-disorder vortex matter phase transition. However, local magnetic measurements trace this effect to metastable disordered vortex states, revealing the same order-disorder transition induction in samples of different size.
Temperature Stratification in a Cryogenic Fuel Tank
National Aeronautics and Space Administration — A reduced dynamical model describing temperature stratification effects driven by natural convection in a liquid hydrogen cryogenic fuel tank has been developed. It...
Microfluidic destabilization of viscous stratifications: Interfacial waves and droplets
Hu, Xiaoyi; Cubaud, Thomas
2016-11-01
Microfluidic two-fluid flows with large differences in viscosity are experimentally investigated to examine the role of fluid properties on hydrodynamic destabilization processes at the small scale. Two- and three-layer flow configurations are systematically studied in straight square microchannels using miscible and immiscible fluid pairs. We focus our attention on symmetric three-layer stratifications with a fast central stream made of low-viscosity fluid and a slow sheath flow composed of high-viscosity fluid. We quantify the influence of the capillary and the Reynolds numbers on the formation and evolution of droplets and wavy stratifications. Several functional relationships are developed for the morphology and dynamics of droplets and interfacial waves including size, celerity and frequency. In the wavy stratification regime, the formation and entrainment of thin viscous ligaments from wave crests display a rich variety of dynamics either in the presence or in the absence of interfacial tension between liquids. This work is supported by NSF (CBET-1150389).
Pure variation and organic stratification.
Rosanvallon, Jérôme
2012-09-01
The fundamental problem posed by Darwin distinguishes his theory from any transformism of the past as well as any evolutionism to come: since variation is inherent to the living, it is a question of explaining, not at all why the living varies, but instead why the living does not vary in all directions to the point of constituting a continuum of forms varying ad infinitum. What limits and stabilizes this intrinsically unlimited variation, allowing certain forms to subsist and multiply to the detriment of others, is natural selection. This double principle of intrinsic variation/extrinsic selection constitutes a vector for the unification of reality that underlies Jean-Jacques Kupiec's ontophylogenesis as well as Deleuze and Guattari's global philosophy of Nature. Therefore, everything would potentially tend to incessantly vary. The work of Kupiec and others identifies an intrinsic random variation within ontogenesis itself. For Deleuze and Guattari, it is nothing but the figure, already selected by the organic stratum, of a more fundamental or pure variation. But, in fact, nothing really varies incessantly: everything undergoes a selective pressure according to which nothing subsists as such except what manages to endure through invariance (physical stratum) or reproduction (organic stratum). Thus, organic stratification only retains from variation what ensures and augments this reproduction. In this sense, every organism stratifies, i.e. submits to its imperative of subsistence and reproduction, a body without organs that varies in itself and always tends to escape the organism, for better (intensifications of life) or worse (cancerous pathologies). Copyright © 2012 Elsevier Ltd. All rights reserved.
Quinto, Francesca; Lagos, Markus; Plaschke, Markus; Schaefer, Thorsten; Geckeis, Horst [Institute for Nuclear Waste Disposal, Karlsruhe Institute of Technology (Germany); Steier, Peter; Golser, Robin [VERA Laboratory, Faculty of Physics, University of Vienna (Austria)
2016-07-01
With the abundance sensitivities of AMS for U-236, Np-237 and Pu-239 relative to U-238 at levels lower than 1E-15, a simultaneous determination of several actinides without previous chemical separation from each other is possible. The actinides are extracted from the matrix elements via an iron hydroxide co-precipitation and the nuclides sequentially measured from the same sputter target. This simplified method allows for the use of non-isotopic tracers and consequently the determination of Np-237 and Am-243 for which isotopic tracers with the degree of purity required by ultra-trace mass-spectrometric analysis are not available. With detection limits of circa 1E+4 atoms in a sample, 1E+8 atoms are determined with circa 1 % relative uncertainty due to counting statistics. This allows for an unprecedented reduction of the sample size down to 100 ml of natural water. However, the use of non-isotopic tracers introduces a dominating uncertainty of up to 30 % related to the reproducibility of the results. The advantages and drawbacks of the novel method will be presented with the aid of recent results from the CFM Project at the Grimsel Test Site and from the investigation of global fallout in environmental samples.
Durney, Brandon C; Bachert, Beth A; Sloane, Hillary S; Lukomski, Slawomir; Landers, James P; Holland, Lisa A
2015-06-23
Phospholipid additives are a cost-effective medium to separate deoxyribonucleic acid (DNA) fragments and possess a thermally-responsive viscosity. This provides a mechanism to easily create and replace a highly viscous nanogel in a narrow bore capillary with only a 10°C change in temperature. Preparations composed of dimyristoyl-sn-glycero-3-phosphocholine (DMPC) and 1,2-dihexanoyl-sn-glycero-3-phosphocholine (DHPC) self-assemble, forming structures such as nanodisks and wormlike micelles. Factors that influence the morphology of a particular DMPC-DHPC preparation include the concentration of lipid in solution, the temperature, and the ratio of DMPC and DHPC. It has previously been established that an aqueous solution containing 10% phospholipid with a ratio of [DMPC]/[DHPC]=2.5 separates DNA fragments with nearly single base resolution for DNA fragments up to 500 base pairs in length, but beyond this size the resolution decreases dramatically. A new DMPC-DHPC medium is developed to effectively separate and size DNA fragments up to 1500 base pairs by decreasing the total lipid concentration to 2.5%. A 2.5% phospholipid nanogel generates a resolution of 1% of the DNA fragment size up to 1500 base pairs. This increase in the upper size limit is accomplished using commercially available phospholipids at an even lower material cost than is achieved with the 10% preparation. The separation additive is used to evaluate size markers ranging between 200 and 1500 base pairs in order to distinguish invasive strains of Streptococcus pyogenes and Aspergillus species by harnessing differences in gene sequences of collagen-like proteins in these organisms. For the first time, a reversible stacking gel is integrated in a capillary sieving separation by utilizing the thermally-responsive viscosity of these self-assembled phospholipid preparations. A discontinuous matrix is created that is composed of a cartridge of highly viscous phospholipid assimilated into a separation matrix
Sevelius, Jae M.
2017-01-01
Background. Transgender individuals have a gender identity that differs from the sex they were assigned at birth. The population size of transgender individuals in the United States is not well-known, in part because official records, including the US Census, do not include data on gender identity. Population surveys today more often collect transgender-inclusive gender-identity data, and secular trends in culture and the media have created a somewhat more favorable environment for transgender people. Objectives. To estimate the current population size of transgender individuals in the United States and evaluate any trend over time. Search methods. In June and July 2016, we searched PubMed, Cumulative Index to Nursing and Allied Health Literature, and Web of Science for national surveys, as well as “gray” literature, through an Internet search. We limited the search to 2006 through 2016. Selection criteria. We selected population-based surveys that used probability sampling and included self-reported transgender-identity data. Data collection and analysis. We used random-effects meta-analysis to pool eligible surveys and used meta-regression to address our hypothesis that the transgender population size estimate would increase over time. We used subsample and leave-one-out analysis to assess for bias. Main results. Our meta-regression model, based on 12 surveys covering 2007 to 2015, explained 62.5% of model heterogeneity, with a significant effect for each unit increase in survey year (F = 17.122; df = 1,10; b = 0.026%; P = .002). Extrapolating these results to 2016 suggested a current US population size of 390 adults per 100 000, or almost 1 million adults nationally. This estimate may be more indicative for younger adults, who represented more than 50% of the respondents in our analysis. Authors’ conclusions. Future national surveys are likely to observe higher numbers of transgender people. The large variety in questions used to ask
Hughes, William O.; McNelis, Anne M.
2010-01-01
The Earth Observing System (EOS) Terra spacecraft was launched on an Atlas IIAS launch vehicle on its mission to observe planet Earth in late 1999. Prior to launch, the new design of the spacecraft's pyroshock separation system was characterized by a series of 13 separation ground tests. The analysis methods used to evaluate this unusually large amount of shock data will be discussed in this paper, with particular emphasis on population distributions and finding statistically significant families of data, leading to an overall shock separation interface level. The wealth of ground test data also allowed a derivation of a Mission Assurance level for the flight. All of the flight shock measurements were below the EOS Terra Mission Assurance level thus contributing to the overall success of the EOS Terra mission. The effectiveness of the statistical methodology for characterizing the shock interface level and for developing a flight Mission Assurance level from a large sample size of shock data is demonstrated in this paper.
Mélachio, Tanekou Tito Trésor; Njiokou, Flobert; Ravel, Sophie; Simo, Gustave; Solano, Philippe; De Meeûs, Thierry
2015-07-01
Human and animal trypanosomiases are two major constraints to development in Africa. These diseases are mainly transmitted by tsetse flies in particular by Glossina palpalis palpalis in Western and Central Africa. To set up an effective vector control campaign, prior population genetics studies have proved useful. Previous studies on population genetics of G. p. palpalis using microsatellite loci showed high heterozygote deficits, as compared to Hardy-Weinberg expectations, mainly explained by the presence of null alleles and/or the mixing of individuals belonging to several reproductive units (Wahlund effect). In this study we implemented a system of trapping, consisting of a central trap and two to four satellite traps around the central one to evaluate a possible role of the Wahlund effect in tsetse flies from three Cameroon human and animal African trypanosomiases foci (Campo, Bipindi and Fontem). We also estimated effective population sizes and dispersal. No difference was observed between the values of allelic richness, genetic diversity and Wright's FIS, in the samples from central and from satellite traps, suggesting an absence of Wahlund effect. Partitioning of the samples with Bayesian methods showed numerous clusters of 2-3 individuals as expected from a population at demographic equilibrium with two expected offspring per reproducing female. As previously shown, null alleles appeared as the most probable factor inducing these heterozygote deficits in these populations. Effective population sizes varied from 80 to 450 individuals while immigration rates were between 0.05 and 0.43, showing substantial genetic exchanges between different villages within a focus. These results suggest that the "suppression" with establishment of physical barriers may be the best strategy for a vector control campaign in this forest context.
Valéria Schimitz Marodim
2000-10-01
Full Text Available Este estudo visa a estabelecer o delineamento experimental e o tamanho de amostra para a cultura da alface (Lactuca sativa em hidroponia, pelo sistema NFT (Nutrient film technique. O experimento foi conduzido no Laboratório de Cultivos Sem Solo/Hidroponia, no Departamento de Fitotecnia da Universidade Federal de Santa Maria e baseou-se em dados de massa de plantas. Os resultados obtidos mostraram que, usando estrutura de cultivo de alface em hidroponia sobre bancadas de fibrocimento com seis canais, o delineamento experimental adequado é blocos ao acaso se a unidade experimental for constituída de faixas transversais aos canais das bancadas, e deve ser inteiramente casualizado se a bancada for a unidade experimental; para a variável massa de plantas, o tamanho da amostra é de 40 plantas para uma semi-amplitude do intervalo de confiança em percentagem da média (d igual a 5% e de 7 plantas para um d igual a 20%.This study was carried out to establish the experimental design and sample size for hydroponic lettuce (Lactuca sativa crop under nutrient film technique. The experiment was conducted in the Laboratory of Hydroponic Crops of the Horticulture Department of the Federal University of Santa Maria. The evaluated traits were plant weight. Under hydroponic conditions on concrete bench with six ducts, the most indicated experimental design for lettuce is randomised blocks for duct transversal plots or completely randomised for bench plot. The sample size for plant weight should be 40 and 7 plants, respectively, for a confidence interval of mean percentage (d equal to 5% and 20%.
Li, Aifeng; Ma, Feifei; Song, Xiuli; Yu, Rencheng
2011-03-18
Solid-phase adsorption toxin tracking (SPATT) technology was developed as an effective passive sampling method for dissolved diarrhetic shellfish poisoning (DSP) toxins in seawater. HP20 and SP700 resins have been reported as preferred adsorption substrates for lipophilic algal toxins and are recommended for use in SPATT testing. However, information on the mechanism of passive adsorption by these polymeric resins is still limited. Described herein is a study on the adsorption of OA and DTX1 toxins extracted from Prorocentrum lima algae by HP20 and SP700 resins. The pore size distribution of the adsorbents was characterized by a nitrogen adsorption method to determine the relationship between adsorption and resin porosity. The Freundlich equation constant showed that the difference in adsorption capacity for OA and DTX1 toxins was not determined by specific surface area, but by the pore size distribution in particular, with micropores playing an especially important role. Additionally, it was found that differences in affinity between OA and DTX1 for aromatic resins were as a result of polarity discrepancies due to DTX1 having an additional methyl moiety.
Haugbøl, Steven; Pinborg, Lars H; Arfan, Haroon M
2006-01-01
PURPOSE: To determine the reproducibility of measurements of brain 5-HT2A receptors with an [18F]altanserin PET bolus/infusion approach. Further, to estimate the sample size needed to detect regional differences between two groups and, finally, to evaluate how partial volume correction affects...... reproducibility and the required sample size. METHODS: For assessment of the variability, six subjects were investigated with [18F]altanserin PET twice, at an interval of less than 2 weeks. The sample size required to detect a 20% difference was estimated from [18F]altanserin PET studies in 84 healthy subjects......% (range 5-12%), whereas in regions with a low receptor density, BP1 reproducibility was lower, with a median difference of 17% (range 11-39%). Partial volume correction reduced the variability in the sample considerably. The sample size required to detect a 20% difference in brain regions with high...
Liu, Xinyu; Wang, Yupeng; Sriram, T N
2014-06-14
Data on single-nucleotide polymorphisms (SNPs) have been found to be useful in predicting phenotypes ranging from an individual's class membership to his/her risk of developing a disease. In multi-class classification scenarios, clinical samples are often limited due to cost constraints, making it necessary to determine the sample size needed to build an accurate classifier based on SNPs. The performance of such classifiers can be assessed using the Area Under the Receiver Operating Characteristic (ROC) Curve (AUC) for two classes and the Volume Under the ROC hyper-Surface (VUS) for three or more classes. Sample size determination based on AUC or VUS would not only guarantee an overall correct classification rate, but also make studies more cost-effective. For coded SNP data from D(≥2) classes, we derive an optimal Bayes classifier and a linear classifier, and obtain a normal approximation to the probability of correct classification for each classifier. These approximations are then used to evaluate the associated AUCs or VUSs, whose accuracies are validated using Monte Carlo simulations. We give a sample size determination method, which ensures that the difference between the two approximate AUCs (or VUSs) is below a pre-specified threshold. The performance of our sample size determination method is then illustrated via simulations. For the HapMap data with three and four populations, a linear classifier is built using 92 independent SNPs and the required total sample sizes are determined for a continuum of threshold values. In all, four different sample size determination studies are conducted with the HapMap data, covering cases involving well-separated populations to poorly-separated ones. For multi-classes, we have developed a sample size determination methodology and illustrated its usefulness in obtaining a required sample size from the estimated learning curve. For classification scenarios, this methodology will help scientists determine whether a sample
Daouda Kassie; Anna Roudot; Nadine Dessay; Jean-Luc Piermay; Gerard Salem; Florence Fournet
2017-01-01
.... Methods This article describes the methodology used to develop a multi-stage sampling protocol to select a population for a demographic survey that investigates health disparities in the medium-sized...
U.S. Geological Survey, Department of the Interior — This data release serves as an archive of sediment physical properties and grain-size data for surficial samples collected offshore of Assateague Island, Maryland...
Dendukuri, Nandini; Bélisle, Patrick; Joseph, Lawrence
2010-11-20
Diagnostic tests rarely provide perfect results. The misclassification induced by imperfect sensitivities and specificities of diagnostic tests must be accounted for when planning prevalence studies or investigations into properties of new tests. The previous work has shown that applying a single imperfect test to estimate prevalence can often result in very large sample size requirements, and that sometimes even an infinite sample size is insufficient for precise estimation because the problem is non-identifiable. Adding a second test can sometimes reduce the sample size substantially, but infinite sample sizes can still occur as the problem remains non-identifiable. We investigate the further improvement possible when three diagnostic tests are to be applied. We first develop methods required for studies when three conditionally independent tests are available, using different Bayesian criteria. We then apply these criteria to prototypic scenarios, showing that large sample size reductions can occur compared to when only one or two tests are used. As the problem is now identifiable, infinite sample sizes cannot occur except in pathological situations. Finally, we relax the conditional independence assumption, demonstrating in this once again non-identifiable situation that sample sizes may substantially grow and possibly be infinite. We apply our methods to the planning of two infectious disease studies, the first designed to estimate the prevalence of Strongyloides infection, and the second relating to estimating the sensitivity of a new test for tuberculosis transmission. The much smaller sample sizes that are typically required when three as compared to one or two tests are used should encourage researchers to plan their studies using more than two diagnostic tests whenever possible. User-friendly software is available for both design and analysis stages greatly facilitating the use of these methods.
Rousing, Tine; Møller, Steen Henrik; Hansen, Steffen W
2012-01-01
" in validity, reliability as well as feasibility - the latter both as regards time and economy costs. This paper based on empiric data addressed the questions on needed sample size for a robust herd assessment of animal based measures. The animal based part of the full WelFur protocol including 9 animal based...... in herd prevalence of the mentioned parameters. Statistical analyses showed that a sample size of 125 adult mink was a robus estimate of the herd level of animal based measures....
Smedslund Geir; Zangi Heidi Andersen; Mowinckel Petter; Hagen Kåre Birger
2013-01-01
Abstract Background Patient reported outcomes are accepted as important outcome measures in rheumatology. The fluctuating symptoms in patients with rheumatic diseases have serious implications for sample size in clinical trials. We estimated the effects of measuring the outcome 1-5 times on the sample size required in a two-armed trial. Findings In a randomized controlled trial that evaluated the effects of a mindfulness-based group intervention for patients with inflammatory arthritis (n=71)...
Smith, Philip L; Lilburn, Simon D; Corbett, Elaine A; Sewell, David K; Kyllingsbæk, Søren
2016-09-01
We investigated the capacity of visual short-term memory (VSTM) in a phase discrimination task that required judgments about the configural relations between pairs of black and white features. Sewell et al. (2014) previously showed that VSTM capacity in an orientation discrimination task was well described by a sample-size model, which views VSTM as a resource comprised of a finite number of noisy stimulus samples. The model predicts the invariance of [Formula: see text] , the sum of squared sensitivities across items, for displays of different sizes. For phase discrimination, the set-size effect significantly exceeded that predicted by the sample-size model for both simultaneously and sequentially presented stimuli. Instead, the set-size effect and the serial position curves with sequential presentation were predicted by an attention-weighted version of the sample-size model, which assumes that one of the items in the display captures attention and receives a disproportionate share of resources. The choice probabilities and response time distributions from the task were well described by a diffusion decision model in which the drift rates embodied the assumptions of the attention-weighted sample-size model.
Atterton, Thomas; De Groote, Isabelle; Eliopoulos, Constantine
2016-10-01
The construction of the biological profile from human skeletal remains is the foundation of anthropological examination. However, remains may be fragmentary and the elements usually employed, such as the pelvis and skull, are not available. The clavicle has been successfully used for sex estimation in samples from Iran and Greece. In the present study, the aim was to test the suitability of the measurements used in those previous studies on a British Medieval population. In addition, the project tested whether discrimination between sexes was due to size or clavicular strength. The sample consisted of 23 females and 25 males of pre-determined sex from two medieval collections: Poulton and Gloucester. Six measurements were taken using an osteometric board, sliding calipers and graduated tape. In addition, putty rings and bi-planar radiographs were made and robusticity measures calculated. The resulting variables were used in stepwise discriminant analyses. The linear measurements allowed correct sex classification in 89.6% of all individuals. This demonstrates the applicability of the clavicle for sex estimation in British populations. The most powerful discriminant factor was maximum clavicular length and the best combination of factors was maximum clavicular length and circumference. This result is similar to that obtained by other studies. To further investigate the extent of sexual dimorphism of the clavicle, the biomechanical properties of the polar second moment of area J and the ratio of maximum to minimum bending rigidity are included in the analysis. These were found to have little influence when entered into the discriminant function analysis. Copyright © 2016 Elsevier GmbH. All rights reserved.
Fotini Kokou
2016-05-01
Full Text Available One of the main concerns in gene expression studies is the calculation of statistical significance which in most cases remains low due to limited sample size. Increasing biological replicates translates into more effective gains in power which, especially in nutritional experiments, is of great importance as individual variation of growth performance parameters and feed conversion is high. The present study investigates in the gilthead sea bream Sparus aurata, one of the most important Mediterranean aquaculture species. For 24 gilthead sea bream individuals (biological replicates the effects of gradual substitution of fish meal by plant ingredients (0% (control, 25%, 50% and 75% in the diets were studied by looking at expression levels of four immune-and stress-related genes in intestine, head kidney and liver. The present results showed that only the lowest substitution percentage is tolerated and that liver is the most sensitive tissue to detect gene expression variations in relation to fish meal substituted diets. Additionally the usage of three independent biological replicates were evaluated by calculating the averages of all possible triplets in order to assess the suitability of selected genes for stress indication as well as the impact of the experimental set up, thus in the present work the impact of FM substitution. Gene expression was altered depending of the selected biological triplicate. Only for two genes in liver (hsp70 and tgf significant differential expression was assured independently of the triplicates used. These results underlined the importance of choosing the adequate sample number especially when significant, but minor differences in gene expression levels are observed.
Reichmann, William M; LaValley, Michael P; Gagnon, David R; Losina, Elena
2013-02-15
Interaction in clinical trials presents challenges for design and appropriate sample size estimation. Here we considered interaction between treatment assignment and a dichotomous prognostic factor with a continuous outcome. Our objectives were to describe differences in power and sample size requirements across alternative distributions of a prognostic factor and magnitudes of the interaction effect, describe the effect of misspecification of the distribution of the prognostic factor on the power to detect an interaction effect, and discuss and compare three methods of handling the misspecification of the prognostic factor distribution. We examined the impact of the distribution of the dichotomous prognostic factor on power and sample size for the interaction effect using traditional one-stage sample size calculation. We varied the magnitude of the interaction effect, the distribution of the prognostic factor, and the magnitude and direction of the misspecification of the distribution of the prognostic factor. We compared quota sampling, modified quota sampling, and sample size re-estimation using conditional power as three strategies for ensuring adequate power and type I error in the presence of a misspecification of the prognostic factor distribution. The sample size required to detect an interaction effect with 80% power increases as the distribution of the prognostic factor becomes less balanced. Misspecification such that the actual distribution of the prognostic factor was more skewed than planned led to a decrease in power with the greatest loss in power seen as the distribution of the prognostic factor became less balanced. Quota sampling was able to maintain the empirical power at 80% and the empirical type I error at 5%. The performance of the modified quota sampling procedure was related to the percentage of trials switching the quota sampling scheme. Sample size re-estimation using conditional power was able to improve the empirical power under
A. Martín Andrés
2015-01-01
Full Text Available The Mantel-Haenszel test is the most frequent asymptotic test used for analyzing stratified 2 × 2 tables. Its exact alternative is the test of Birch, which has recently been reconsidered by Jung. Both tests have a conditional origin: Pearson’s chi-squared test and Fisher’s exact test, respectively. But both tests have the same drawback that the result of global test (the stratified test may not be compatible with the result of individual tests (the test for each stratum. In this paper, we propose to carry out the global test using a multiple comparisons method (MC method which does not have this disadvantage. By refining the method (MCB method an alternative to the Mantel-Haenszel and Birch tests may be obtained. The new MC and MCB methods have the advantage that they may be applied from an unconditional view, a methodology which until now has not been applied to this problem. We also propose some sample size calculation methods.
Neumann, Christoph; Taub, Margaret A; Younkin, Samuel G; Beaty, Terri H; Ruczinski, Ingo; Schwender, Holger
2014-11-01
Case-parent trio studies considering genotype data from children affected by a disease and their parents are frequently used to detect single nucleotide polymorphisms (SNPs) associated with disease. The most popular statistical tests for this study design are transmission/disequilibrium tests (TDTs). Several types of these tests have been developed, for example, procedures based on alleles or genotypes. Therefore, it is of great interest to examine which of these tests have the highest statistical power to detect SNPs associated with disease. Comparisons of the allelic and the genotypic TDT for individual SNPs have so far been conducted based on simulation studies, since the test statistic of the genotypic TDT was determined numerically. Recently, however, it has been shown that this test statistic can be presented in closed form. In this article, we employ this analytic solution to derive equations for calculating the statistical power and the required sample size for different types of the genotypic TDT. The power of this test is then compared with the one of the corresponding score test assuming the same mode of inheritance as well as the allelic TDT based on a multiplicative mode of inheritance, which is equivalent to the score test assuming an additive mode of inheritance. This is, thus, the first time the power of these tests are compared based on equations, yielding instant results and omitting the need for time-consuming simulation studies. This comparison reveals that these tests have almost the same power, with the score test being slightly more powerful.
Age-related changes in resting-state networks of a large sample size of healthy elderly.
Huang, Chun-Chao; Hsieh, Wen-Jin; Lee, Pei-Lin; Peng, Li-Ning; Liu, Li-Kuo; Lee, Wei-Ju; Huang, Jon-Kway; Chen, Liang-Kung; Lin, Ching-Po
2015-10-01
Population aging is burdening the society globally, and the evaluation of functional networks is the key toward understanding cognitive changes in normal aging. However, the effect of age on default mode subnetworks has not been documented well, and age-related changes in many resting-state networks remain debatable. The purpose of this study was to propose more precise results for these issues using a large sample size. We used group-level meta-ICA analysis and dual regression approach for identifying resting-state networks from functional magnetic resonance imaging data of 430 healthy elderly participants. Partial correlation was used to observe age-related correlations within and between resting-state networks. In the default mode network, only the ventral subnetwork negatively correlated with age. Age-related decrease in functional connectivity was also noted in the auditory, right frontoparietal, sensorimotor, and visual medial networks. Further, some age-related increases and decreases were observed for between-network correlations. The results of this study suggest that only the ventral default mode subnetwork had age-related decline in functional connectivity and several reverse patterns of resting-state networks for network development. Understanding age-related network changes may provide solutions for the impact of population aging and diagnosis of neurodegenerative diseases. © 2015 John Wiley & Sons Ltd.
Gruijter, de J.J.; Braak, ter C.J.F.
1992-01-01
Two fundamentally different sources of randomness exist on which design and inference in spatial sampling can be based: (a) variation that would occur on resampling the same spatial population with other sampling configurations generated by the same design, and (b) variation occurring on sampling
Long time durability tests of fabric inlet stratification pipes
Andersen, Elsa; Furbo, Simon
2008-01-01
The long time durability of seven different two layer fabric inlet stratification pipes for enhancing thermal stratification in hot water stores is investigated experimentally. Accelerated durability tests are carried out with the inlet stratification pipes both in a domestic hot water tank...... and that this destroys the capability of building up thermal stratification for the fabric inlet stratification pipe. The results also show that although dirt, algae etc. are deposited in the fabric pipes in the space heating tank, the capability of the fabric inlet stratifiers to build up thermal stratification...
Soo, Jhy-Charm; Lee, Eun Gyung; Lee, Larry A; Kashon, Michael L; Harper, Martin
2014-10-01
Lee et al. (Evaluation of pump pulsation in respirable size-selective sampling: part I. Pulsation measurements. Ann Occup Hyg 2014a;58:60-73) introduced an approach to measure pump pulsation (PP) using a real-world sampling train, while the European Standards (EN) (EN 1232-1997 and EN 12919-1999) suggest measuring PP using a resistor in place of the sampler. The goal of this study is to characterize PP according to both EN methods and to determine the relationship of PP between the published method (Lee et al., 2014a) and the EN methods. Additional test parameters were investigated to determine whether the test conditions suggested by the EN methods were appropriate for measuring pulsations. Experiments were conducted using a factorial combination of personal sampling pumps (six medium- and two high-volumetric flow rate pumps), back pressures (six medium- and seven high-flow rate pumps), resistors (two types), tubing lengths between a pump and resistor (60 and 90 cm), and different flow rates (2 and 2.5 l min(-1) for the medium- and 4.4, 10, and 11.2 l min(-1) for the high-flow rate pumps). The selection of sampling pumps and the ranges of back pressure were based on measurements obtained in the previous study (Lee et al., 2014a). Among six medium-flow rate pumps, only the Gilian5000 and the Apex IS conformed to the 10% criterion specified in EN 1232-1997. Although the AirChek XR5000 exceeded the 10% limit, the average PP (10.9%) was close to the criterion. One high-flow rate pump, the Legacy (PP=8.1%), conformed to the 10% criterion in EN 12919-1999, while the Elite12 did not (PP=18.3%). Conducting supplemental tests with additional test parameters beyond those used in the two subject EN standards did not strengthen the characterization of PPs. For the selected test conditions, a linear regression model [PPEN=0.014+0.375×PPNIOSH (adjusted R2=0.871)] was developed to determine the PP relationship between the published method (Lee et al., 2014a) and the EN methods
Franke, Karl-Josef; Szyrach, Mara; Nilius, Georg; Hetzel, Jürgen; Hetzel, Martin; Ruehle, Karl-Heinz; Enderle, Markus D
2009-08-01
Cryoextraction is a procedure for recanalization of obstructed airways caused by exophytic growing tumors. Biopsy samples obtained with this method can be used for histological diagnosis. The objective of this study was to evaluate the parameters influencing the size of cryobiopsies in an in vitro animal model. New flexible cryoprobes with different diameters were used to extract biopsies from lung tissue. These biopsies were compared with forceps biopsy (gold standard) in terms of the biopsy size. Tissue dependency of the biopsy size was analyzed by comparing biopsies taken from the lung, the liver, and gastric mucosa. The effect of contact pressure exerted by the tip of the cryoprobe on the tissue was analyzed on liver tissue separately. Biopsy size was estimated by measuring the weight and the diameter. Weight and diameter of cryobiopsies correlated positively with longer activation times and larger diameters of the cryoprobe. The weight of the biopsies was tissue dependent: lung biopsy diameter. The biopsy size increased when the probe was pressed on the tissue during cooling. Cryobiopsies can be taken from different tissue types with flexible cryoprobes. The size of the samples depends on tissue type, probe diameter, application time, and pressure exerted by the probe on the tissue. Even the cryoprobe with the smallest diameter can provide larger biopsies than a forceps biopsy in lung. It can be expected that the same parameters influence the sample size of biopsies in vivo.
Willruth, A M; Steinhard, J; Enzensberger, C; Axt-Fliedner, R; Gembruch, U; Doelle, A; Dimitriou, I; Fimmers, R; Bahlmann, F
2016-02-04
Purpose: To assess the time intervals of the cardiac cycle in healthy fetuses in the second and third trimester using color tissue Doppler imaging (cTDI) and to evaluate the influence of different sizes of sample gates on time interval values. Materials and Methods: Time intervals were measured from the cTDI-derived Doppler waveform using a small and large region of interest (ROI) in healthy fetuses. Results: 40 fetuses were included. The median gestational age at examination was 26 + 1 (range: 20 + 5 - 34 + 5) weeks. The median frame rate was 116/s (100 - 161/s) and the median heart rate 143 (range: 125 - 158) beats per minute (bpm). Using small and large ROIs, the second trimester right ventricular (RV) mean isovolumetric contraction times (ICTs) were 39.8 and 41.4 ms (p = 0.17), the mean ejection times (ETs) were 170.2 and 164.6 ms (p < 0.001), the mean isovolumetric relaxation times (IRTs) were 52.8 and 55.3 ms (p = 0.08), respectively. The left ventricular (LV) mean ICTs were 36.2 and 39.4 ms (p = 0.05), the mean ETs were 167.4 and 164.5 ms (p = 0.013), the mean IRTs were 53.9 and 57.1 ms (p = 0.05), respectively. The third trimester RV mean ICTs were 50.7 and 50.4 ms (p = 0.75), the mean ETs were 172.3 and 181.4 ms (p = 0.49), the mean IRTs were 50.2 and 54.6 ms (p = 0.03); the LV mean ICTs were 45.1 and 46.2 ms (p = 0.35), the mean ETs were 175.2 vs. 172.9 ms (p = 0.29), the mean IRTs were 47.1 and 50.0 ms (p = 0.01), respectively. Conclusion: Isovolumetric time intervals can be analyzed precisely and relatively independent of ROI size. In the near future, automatic time interval measurement using ultrasound systems will be feasible and the analysis of fetal myocardial function can become part of the clinical routine.
George, Goldy C.; Hoelscher, Deanna M.; Nicklas, Theresa A.; Kelder, Steven H.
2009-01-01
Objective: To examine diet- and body size-related attitudes and behaviors associated with supplement use in a representative sample of fourth-grade students in Texas. Design: Cross-sectional data from the School Physical Activity and Nutrition study, a probability-based sample of schoolchildren. Children completed a questionnaire that assessed…
Williamson, J B; Boehmer, U
1997-07-01
A number of studies have attempted to account for cross-national differences in life expectancy, but relatively few have focused on female life expectancy, and even fewer on the relevance of predictors linked to gender stratification theory. The present study seeks to assess the utility of gender stratification theory in accounting for cross-national differences in female life expectancy in less developed countries. An incremental model building strategy is used to develop a final model that combines predictors linked to both industrialism theory and gender stratification theory. The analysis is based on multiple regression and cross-sectional samples that vary in size from 40 to 97 countries. Evidence is presented that several aspects of women's status have a positive effect on female life expectancy. Indicators of women's educational status, women's economic status, and women's reproductive autonomy all prove to be important predictors of female life expectancy. Analysis of interaction effects suggests that the strength of the effects of some aspects of women's economic status and the effect of some aspects of health status on female life expectancy vary with the level of economic development. A comprehensive assessment of the relative strength of alternative measures of women's education is carried out, and evidence is presented that it does make a difference how the level of women's education is measured.
Trattner, Sigal; Cheng, Bin; Pieniazek, Radoslaw L.; Hoffmann, Udo; Douglas, Pamela S.; Einstein, Andrew J.
2014-01-01
Purpose: Effective dose (ED) is a widely used metric for comparing ionizing radiation burden between different imaging modalities, scanners, and scan protocols. In computed tomography (CT), ED can be estimated by performing scans on an anthropomorphic phantom in which metal-oxide-semiconductor field-effect transistor (MOSFET) solid-state dosimeters have been placed to enable organ dose measurements. Here a statistical framework is established to determine the sample size (number of scans) needed for estimating ED to a desired precision and confidence, for a particular scanner and scan protocol, subject to practical limitations. Methods: The statistical scheme involves solving equations which minimize the sample size required for estimating ED to desired precision and confidence. It is subject to a constrained variation of the estimated ED and solved using the Lagrange multiplier method. The scheme incorporates measurement variation introduced both by MOSFET calibration, and by variation in MOSFET readings between repeated CT scans. Sample size requirements are illustrated on cardiac, chest, and abdomen–pelvis CT scans performed on a 320-row scanner and chest CT performed on a 16-row scanner. Results: Sample sizes for estimating ED vary considerably between scanners and protocols. Sample size increases as the required precision or confidence is higher and also as the anticipated ED is lower. For example, for a helical chest protocol, for 95% confidence and 5% precision for the ED, 30 measurements are required on the 320-row scanner and 11 on the 16-row scanner when the anticipated ED is 4 mSv; these sample sizes are 5 and 2, respectively, when the anticipated ED is 10 mSv. Conclusions: Applying the suggested scheme, it was found that even at modest sample sizes, it is feasible to estimate ED with high precision and a high degree of confidence. As CT technology develops enabling ED to be lowered, more MOSFET measurements are needed to estimate ED with the same
Smedslund Geir
2013-02-01
Full Text Available Abstract Background Patient reported outcomes are accepted as important outcome measures in rheumatology. The fluctuating symptoms in patients with rheumatic diseases have serious implications for sample size in clinical trials. We estimated the effects of measuring the outcome 1-5 times on the sample size required in a two-armed trial. Findings In a randomized controlled trial that evaluated the effects of a mindfulness-based group intervention for patients with inflammatory arthritis (n=71, the outcome variables Numerical Rating Scales (NRS (pain, fatigue, disease activity, self-care ability, and emotional wellbeing and General Health Questionnaire (GHQ-20 were measured five times before and after the intervention. For each variable we calculated the necessary sample sizes for obtaining 80% power (α=.05 for one up to five measurements. Two, three, and four measures reduced the required sample sizes by 15%, 21%, and 24%, respectively. With three (and five measures, the required sample size per group was reduced from 56 to 39 (32 for the GHQ-20, from 71 to 60 (55 for pain, 96 to 71 (73 for fatigue, 57 to 51 (48 for disease activity, 59 to 44 (45 for self-care, and 47 to 37 (33 for emotional wellbeing. Conclusions Measuring the outcomes five times rather than once reduced the necessary sample size by an average of 27%. When planning a study, researchers should carefully compare the advantages and disadvantages of increasing sample size versus employing three to five repeated measurements in order to obtain the required statistical power.
NeCamp, Timothy; Kilbourne, Amy; Almirall, Daniel
2017-08-01
Cluster-level dynamic treatment regimens can be used to guide sequential treatment decision-making at the cluster level in order to improve outcomes at the individual or patient-level. In a cluster-level dynamic treatment regimen, the treatment is potentially adapted and re-adapted over time based on changes in the cluster that could be impacted by prior intervention, including aggregate measures of the individuals or patients that compose it. Cluster-randomized sequential multiple assignment randomized trials can be used to answer multiple open questions preventing scientists from developing high-quality cluster-level dynamic treatment regimens. In a cluster-randomized sequential multiple assignment randomized trial, sequential randomizations occur at the cluster level and outcomes are observed at the individual level. This manuscript makes two contributions to the design and analysis of cluster-randomized sequential multiple assignment randomized trials. First, a weighted least squares regression approach is proposed for comparing the mean of a patient-level outcome between the cluster-level dynamic treatment regimens embedded in a sequential multiple assignment randomized trial. The regression approach facilitates the use of baseline covariates which is often critical in the analysis of cluster-level trials. Second, sample size calculators are derived for two common cluster-randomized sequential multiple assignment randomized trial designs for use when the primary aim is a between-dynamic treatment regimen comparison of the mean of a continuous patient-level outcome. The methods are motivated by the Adaptive Implementation of Effective Programs Trial which is, to our knowledge, the first-ever cluster-randomized sequential multiple assignment randomized trial in psychiatry.
Theoretical comparison of solar water/space-heating combi systems and stratification design options
Andersen, Elsa; Furbo, Simon
2007-01-01
A theoretical analysis of differently designed solar combi systems is performed with weather data from the Danish Design Reference Year (55ºN). Three solar combi system designs found on the market are investigated. The investigation focuses on the influence of stratification on the thermal...... performance of using inlet stratification pipes at the different inlets. Also, it is investigated how the design of the space heating system, the control system of the solar collectors, and the system size influence the thermal performance of solar combi systems. The work is carried out within the Solar...
Stratification in Ap star atmospheres: Simulations
Cowley, Charles R.; Castelli, Fiorella
2017-01-01
It is now well established that the atmospheres of Ap stars can be chemically stratified (cf. Babel, A\\&A 258, 645, 1992; Ryabchikova et al. A\\&A 384, 545, 2002). The most convincing cases have been made with the profiles of very strong lines, such as Ca II K. Weaker line profiles are less obvious indicators. The collective behavior of sets or groups of lines have also been used. For example, if higher abundances are derived for strong lines in an atmosphere with zero microturbulence, one may assume that the absorbing species has been pushed into the higher photospheres. An example are the medium-strong Mn II lines in HgMn stars. In this paper, we probe this assumption by calculating line strengths with various assumed stratification models, and then determining abundances from those lines using an {\\bf unstratified} model with the same Teff and log(g). We use the model from Castelli, Kurucz \\& and Hubrig (A\\&A, 508, 401, 2009) for HR 6000, whose spectrum shows numerous indications of stratification. A variety of stratification models are considered, for example, ones where the majority of an absorbing species is concentrated above (or below) $log(\\tau_{5000}$ = -2.0. Cloud models are also investigated, where a species is concentrated within a range of photospheric depths. Curves of growth are generated in unstratified atmospheres for lines by holding the abundance fixed, and increasing log(gf). Similar curves are made in stratified models, and the ratios of strong to weak linesare compared with and without stratification. The effects of stratification on ionization are also investigated, as well as on the profiles of strong lines. We find, in agreement with previous work, that severe abundance jumps are sometimes required to account for some of the observed peculiarities.
Assessment of Stellar Stratification in Three Young Star Clusters in the Large Magellanic Cloud
Gouliermis, Dimitrios A; Xin, Yu; Rochau, Boyke
2010-01-01
(abridged) We present a comprehensive study of stellar stratification in young star clusters in the Large Magellanic Cloud (LMC). We apply our recently developed effective radius method for the assessment of stellar stratification on imaging data obtained with the Advanced Camera for Surveys of three young LMC clusters to characterize the phenomenon and develop a comparative scheme for its assessment in such clusters. The clusters of our sample, NGC 1983, NGC 2002 and NGC 2010, are selected on the basis of their youthfulness, and their variety in appearance, structure, stellar content, and surrounding stellar ambient. Our photometry is complete for magnitudes down to m_814 ~ 23 mag, allowing the calculation of the structural parameters of the clusters, the estimation of their ages and the determination of their stellar content. Our study shows that each cluster in our sample demonstrates stellar stratification in a quite different manner and at different degree from the others. Specifically, NGC 1983 shows to...
ShidaLIU; ZuguangZheng; 等
1996-01-01
We analyse the behavior of the nonlinear dynamical systems which are the truncated-spectrum model of the atmospheric turbulence equation.It shows that the chaos can appear in the Lorenz equation obtained by simple equations for the unstable stratification(Ri0),And the chaos can also appear in Burgers-Chao equations for the stable stratification(Ri>0,Ra<0),The atmospheric turbulence is intermittent in the stable stratified atmosphere.
Stratification-Based Outlier Detection over the Deep Web
Xuefeng Xian
2016-01-01
Full Text Available For many applications, finding rare instances or outliers can be more interesting than finding common patterns. Existing work in outlier detection never considers the context of deep web. In this paper, we argue that, for many scenarios, it is more meaningful to detect outliers over deep web. In the context of deep web, users must submit queries through a query interface to retrieve corresponding data. Therefore, traditional data mining methods cannot be directly applied. The primary contribution of this paper is to develop a new data mining method for outlier detection over deep web. In our approach, the query space of a deep web data source is stratified based on a pilot sample. Neighborhood sampling and uncertainty sampling are developed in this paper with the goal of improving recall and precision based on stratification. Finally, a careful performance evaluation of our algorithm confirms that our approach can effectively detect outliers in deep web.
Stratification-Based Outlier Detection over the Deep Web.
Xian, Xuefeng; Zhao, Pengpeng; Sheng, Victor S; Fang, Ligang; Gu, Caidong; Yang, Yuanfeng; Cui, Zhiming
2016-01-01
For many applications, finding rare instances or outliers can be more interesting than finding common patterns. Existing work in outlier detection never considers the context of deep web. In this paper, we argue that, for many scenarios, it is more meaningful to detect outliers over deep web. In the context of deep web, users must submit queries through a query interface to retrieve corresponding data. Therefore, traditional data mining methods cannot be directly applied. The primary contribution of this paper is to develop a new data mining method for outlier detection over deep web. In our approach, the query space of a deep web data source is stratified based on a pilot sample. Neighborhood sampling and uncertainty sampling are developed in this paper with the goal of improving recall and precision based on stratification. Finally, a careful performance evaluation of our algorithm confirms that our approach can effectively detect outliers in deep web.
A Comparative Review of Stratification Texts and Readers
Peoples, Clayton D.
2012-01-01
Social stratification is a core substantive area within sociology. There are a number of textbooks and readers available on the market that deal with this central topic. In this article, I conduct a comparative review of (a) four stratification textbooks and (b) four stratification readers. (Contains 2 tables.)
Gerrit eVoordouw
2016-03-01
Full Text Available Microbially-influenced corrosion (MIC contributes to the general corrosion rate (CR, which is typically measured with carbon steel coupons. Here we explore the use of carbon steel ball bearings, referred to as beads (55.0 ± 0.3 mg; Ø = 0.238 cm, for determining CRs. CRs for samples from an oil field in Oceania incubated with beads were determined by the weight loss method, using acid treatment to remove corrosion products. The release of ferrous and ferric iron was also measured and CRs based on weight loss and iron determination were in good agreement. Average CRs were 0.022 mm/yr for 8 produced waters with high numbers (105/ml of acid-producing bacteria (APB, but no sulfate-reducing bacteria (SRB. Average CRs were 0.009 mm/yr for 5 central processing facility (CPF waters, which had no APB or SRB due to weekly biocide treatment and 0.036 mm/yr for 2 CPF tank bottom sludges, which had high numbers of APB (106/ml and SRB (108/ml. Hence, corrosion monitoring with carbon steel beads indicated that biocide treatment of CPF waters decreased the CR, except where biocide did not penetrate. The CR for incubations with 20 ml of a produced water decreased from 0.061 to 0.007 mm/yr when increasing the number of beads from 1 to 40. CRs determined with beads were higher than those with coupons, possibly also due to a higher weight of iron per unit volume used in incubations with coupons. Use of 1 ml syringe columns, containing carbon steel beads and injected with 10 ml/day of SRB-containing medium for 256 days gave a CR of 0.11 mm/yr under flow conditions. The standard deviation of the distribution of residual bead weights, a measure for the unevenness of the corrosion, increased with increasing CR. The most heavily corroded beads showed significant pitting. Hence the use of uniformly sized carbon steel beads offers new opportunities for screening and monitoring of corrosion including determination of the distribution of corrosion rates, which allows
无
2007-01-01
Investigations into forest soils face the problem of the high level of spatial variability that is an inherent property of all forest soils. In order to investigate the effect of changes in residue management practices on soil properties in hoop pine (Araucaria cunninghamii Aiton ex A. Cunn.) plantations of subtropical Australia it was important to understand the intensity of sampling effort required to overcome the spatial variability induced by those changes. Harvest residues were formed into windrows to prevent nitrogen (N) losses through volatilisation and erosion that had previously occurred as a result of pile and burn operations. We selected second rotation (2R) hoop pine sites where the windrows (10-15 m apart) had been formed 1, 2 and 3 years prior to sampling in order to examine the spatial variability in soil carbon (C)and N and in potential mineralisable N (PMN) in the areas beneath and between (inter-) the windrows. We examined the implications of soil variability on the number of samples required to detect differences in means for specific soil properties,at different ages and at specified levels of accuracy. Sample size needed to accurately reflect differences between means was not affected by the position where the samples were taken relative to the windrows but differed according to the parameter to be sampled. The relative soil sampling size required for detecting differences between means of a soil property in the inter-windrow and beneath-windrow positions was highly dependent on the soil property assessed and the acceptable relative sampling error. An alternative strategy for soil sampling should be considered, if the estimated sample size exceeds 50 replications. The possible solution to this problem is collection of composite soil samples allowing a substantial reduction in the number of samples required for chemical analysis without loss in the precision of the mean estimates for a particular soil property.
Parshintsev, Jevgeni; Ruiz-Jimenez, Jose; Petäjä, Tuukka; Hartonen, Kari; Kulmala, Markku; Riekkola, Marja-Liisa
2011-07-01
In this research, the two most common filter media, quartz and Teflon, were tested to obtain information about the possible adsorption of gas-phase compounds onto filters during long sample collection of atmospheric aerosols. Particles of nanometer-size for off-line chemical characterization were collected using a recently introduced differential mobility analyzer for size separation. Samples were collected at an urban site (Helsinki, SMEARIII station) during spring 2010. Sampling time was 4 to 10 days for particles 50, 40, or 30 nm in diameter. Sample air flow was 4 L/min. The sampling setup was arranged so that two samples were obtained for each sampling period almost simultaneously: one containing particles and adsorbed gas-phase compounds and one containing adsorbed gas-phase compounds only. Filters were extracted and analyzed for the presence of selected carboxylic acids, polyols, nitrogen-containing compounds, and aldehydes. The results showed that, in quartz filter samples, gas-phase adsorption may be responsible for as much as 100% of some compound masses. Whether quartz or Teflon, simultaneous collection of gas-phase zero samples is essential during the whole sampling period. The dependence of the adsorption of gas-phase compounds on vapor pressure and the effect of adsorption on the deposited aerosol layer are discussed.
Rubinstein, Sidney M; van Eekelen, Rik; Oosterhuis, Teddy; de Boer, Michiel R; Ostelo, Raymond W J G; van Tulder, Maurits W
2014-10-01
The purpose of this study was to evaluate changes in methodological quality and sample size in randomized controlled trials (RCTs) of spinal manipulative therapy (SMT) for neck and low back pain over a specified period. A secondary purpose was to make recommendations for improvement for future SMT trials based upon our findings. Randomized controlled trials that examined the effect of SMT in adults with neck and/or low back pain and reported at least 1 patient-reported outcome measure were included. Studies were identified from recent Cochrane reviews of SMT, and an update of the literature was conducted (March 2013). Risk of bias was assessed using the 12-item criteria recommended by the Cochrane Back Review Group. In addition, sample size was examined. The relationship between the overall risk of bias and sample size over time was evaluated using regression analyses, and RCTs were grouped into periods (epochs) of approximately 5 years. In total, 105 RCTs were included, of which 41 (39%) were considered to have a low risk of bias. There is significant improvement in the mean risk of bias over time (P statistically (odds ratio, 2.1; confidence interval, 1.5-3.0). Sensitivity analyses suggest no appreciable difference between studies for neck or low back pain for risk of bias or sample size. Methodological quality of RCTs of SMT for neck and low back pain is improving, whereas overall sample size has shown only small and nonsignificant increases. There is an increasing trend among studies to conduct sample size calculations, which relate to statistical power. Based upon these findings, 7 areas of improvement for future SMT trials are suggested. Copyright © 2014 National University of Health Sciences. Published by Elsevier Inc. All rights reserved.
Risk stratification in emergency patients by copeptin
Iversen, Kasper; Gøtze, Jens P; Dalsgaard, Morten
2014-01-01
BACKGROUND: Rapid risk stratification is a core task in emergency medicine. Identifying patients at high and low risk shortly after admission could help clinical decision-making regarding treatment, level of observation, allocation of resources and post discharge follow-up. The purpose of the pre......BACKGROUND: Rapid risk stratification is a core task in emergency medicine. Identifying patients at high and low risk shortly after admission could help clinical decision-making regarding treatment, level of observation, allocation of resources and post discharge follow-up. The purpose...... to 0.1% (1/693) for patients with normal copeptin concentrations (that is, ≤11.3 pmol/L) (P figures for one-year mortality and for the entire...
Risk Stratification for Second Primary Lung Cancer.
Han, Summer S; Rivera, Gabriel A; Tammemägi, Martin C; Plevritis, Sylvia K; Gomez, Scarlett L; Cheng, Iona; Wakelee, Heather A
2017-09-01
Purpose This study estimated the 10-year risk of developing second primary lung cancer (SPLC) among survivors of initial primary lung cancer (IPLC) and evaluated the clinical utility of the risk prediction model for selecting eligibility criteria for screening. Methods SEER data were used to identify a population-based cohort of 20,032 participants diagnosed with IPLC between 1988 and 2003 and who survived ≥ 5 years after the initial diagnosis. We used a proportional subdistribution hazards model to estimate the 10-year risk of developing SPLC among survivors of lung cancer LC in the presence of competing risks. Considered predictors included age, sex, race, treatment, histology, stage, and extent of disease. We examined the risk-stratification ability of the prediction model and performed decision curve analysis to evaluate the clinical utility of the model by calculating its net benefit in varied risk thresholds for screening. Results Although the median 10-year risk of SPLC among survivors of LC was 8.36%, the estimated risk varied substantially (range, 0.56% to 14.3%) when stratified by age, histology, and extent of IPLC in the final prediction model. The stratification by deciles of estimated risk showed that the observed incidence of SPLC was significantly higher in the tenth-decile group (12.5%) versus the first-decile group (2.9%; P risk thresholds (1% to 11.5%) at which the clinical net benefit of the risk model was larger than those in hypothetical all-screening or no-screening scenarios. Conclusion The risk stratification approach in SPLC can be potentially useful for identifying survivors of LC to be screened by computed tomography. More comprehensive environmental and genetic data may help enhance the predictability and stratification ability of the risk model for SPLC.
Ellis, Alisha; Wheaton, Cathryn J.; Smith, Christopher G.
2017-01-01
This data release serves as an archive of sediment physical properties and grain-size data for surficial samples collected offshore of Assateague Island, Maryland and Virginia, for comparison with surficial estuarine and subaerial sedimentological samples collected and assessed following Hurricane Sandy (Ellis and others, 2015; Smith and others, 2015; Bernier and others, 2016). The sediment samples were collected by scientists from the U.S. Geological Survey (USGS) office in Woods Hole, Massachusetts while aboard the motor vessel (M/V) Scarlett Isabella as part of a larger effort to map the inner continental shelf (Pendleton and others, 2016). Following field work, the sediment samples were shipped to the USGS Coastal and Marine Science Center in St. Petersburg, Florida, where they were renamed for consistency with a previously existing naming scheme and processed for bulk density, loss on ignition (LOI), and grain-size. The grain-size subsamples were processed on a Coulter LS200 particle-size analyzer for consistency regarding methods and output statistics with related data sets from Chincoteague Bay and Assateague Island. For more information regarding sample collection and site information or the related data sets, refer to USGS data release Pendleton and others, 2016; for more information regarding processing methods refer to USGS Open-File Report 2015–1219.
Drainage and Stratification Kinetics of Foam Films
Zhang, Yiran; Sharma, Vivek
2014-03-01
Baking bread, brewing cappuccino, pouring beer, washing dishes, shaving, shampooing, whipping eggs and blowing bubbles all involve creation of aqueous foam films. Foam lifetime, drainage kinetics and stability are strongly influenced by surfactant type (ionic vs non-ionic), and added proteins, particles or polymers modify typical responses. The rate at which fluid drains out from a foam film, i.e. drainage kinetics, is determined in the last stages primarily by molecular interactions and capillarity. Interestingly, for certain low molecular weight surfactants, colloids and polyelectrolyte-surfactant mixtures, a layered ordering of molecules, micelles or particles inside the foam films leads to a stepwise thinning phenomena called stratification. Though stratification is observed in many confined systems including foam films containing particles or polyelectrolytes, films containing globular proteins seem not to show this behavior. Using a Scheludko-type cell, we experimentally study the drainage and stratification kinetics of horizontal foam films formed by protein-surfactant mixtures, and carefully determine how the presence of proteins influences the hydrodynamics and thermodynamics of foam films.
Hiroshi Nishiura
Full Text Available BACKGROUND: Seroepidemiological studies before and after the epidemic wave of H1N1-2009 are useful for estimating population attack rates with a potential to validate early estimates of the reproduction number, R, in modeling studies. METHODOLOGY/PRINCIPAL FINDINGS: Since the final epidemic size, the proportion of individuals in a population who become infected during an epidemic, is not the result of a binomial sampling process because infection events are not independent of each other, we propose the use of an asymptotic distribution of the final size to compute approximate 95% confidence intervals of the observed final size. This allows the comparison of the observed final sizes against predictions based on the modeling study (R = 1.15, 1.40 and 1.90, which also yields simple formulae for determining sample sizes for future seroepidemiological studies. We examine a total of eleven published seroepidemiological studies of H1N1-2009 that took place after observing the peak incidence in a number of countries. Observed seropositive proportions in six studies appear to be smaller than that predicted from R = 1.40; four of the six studies sampled serum less than one month after the reported peak incidence. The comparison of the observed final sizes against R = 1.15 and 1.90 reveals that all eleven studies appear not to be significantly deviating from the prediction with R = 1.15, but final sizes in nine studies indicate overestimation if the value R = 1.90 is used. CONCLUSIONS: Sample sizes of published seroepidemiological studies were too small to assess the validity of model predictions except when R = 1.90 was used. We recommend the use of the proposed approach in determining the sample size of post-epidemic seroepidemiological studies, calculating the 95% confidence interval of observed final size, and conducting relevant hypothesis testing instead of the use of methods that rely on a binomial proportion.
Halme, Alex S; Fritel, Xavier; Benedetti, Andrea; Eng, Ken; Tannenbaum, Cara
2015-03-01
Sample size calculations for treatment trials that aim to assess health-related quality-of-life (HRQOL) outcomes are often difficult to perform. Researchers must select a target minimal clinically important difference (MCID) in HRQOL for the trial, estimate the effect size of the intervention, and then consider the responsiveness of different HRQOL measures for detecting improvements. Generic preference-based HRQOL measures are usually less sensitive to gains in HRQOL than are disease-specific measures, but are nonetheless recommended to quantify an impact on HRQOL that can be translated into quality-adjusted life-years during cost-effectiveness analyses. Mapping disease-specific measures onto generic measures is a proposed method for yielding more efficient sample size requirements while retaining the ability to generate utility weights for cost-effectiveness analyses. This study sought to test this mapping strategy to calculate and compare the effect on sample size of three different methods. Three different methods were used for determining an MCID in HRQOL in patients with incontinence: 1) a global rating of improvement, 2) an incontinence-specific HRQOL instrument, and 3) a generic preference-based HRQOL instrument using mapping coefficients. The sample size required to detect a 20% difference in the MCID for the global rating of improvement was 52 per trial arm, 172 per arm for the incontinence-specific HRQOL outcome, and 500 per arm for the generic preference-based HRQOL outcome. We caution that treatment trials of conditions for which improvements are not easy to measure on generic HRQOL instruments will still require significantly greater sample size even when mapping functions are used to try to gain efficiency. Copyright © 2015 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.
PPOOLEX experiments on thermal stratification and mixing
Puustinen, M.; Laine, J.; Raesaenen, A. (Lappeenranta Univ. of Technology, Nuclear Safety Research Unit (Finland))
2009-08-15
The results of the thermal stratification experiments in 2008 with the PPOOLEX test facility are presented. PPOOLEX is a closed vessel divided into two compartments, dry well and wet well. Extra temperature measurements for capturing different aspects of the investigated phenomena were added before the experiments. The main purpose of the experiment series was to generate verification data for evaluating the capability of GOTHIC code to predict stratification and mixing phenomena. Altogether six experiments were carried out. Heat-up periods of several thousand seconds by steam injection into the dry well compartment and from there into the wet well water pool were recorded. The initial water bulk temperature was 20 deg. C. Cooling periods of several days were included in three experiments. A large difference between the pool bottom and top layer temperature was measured when small steam flow rates were used. With higher flow rates the mixing effect of steam discharge delayed the start of stratification until the pool bulk temperature exceeded 50 deg. C. The stratification process was also different in these two cases. With a small flow rate stratification was observed only above and just below the blowdown pipe outlet elevation. With a higher flow rate over a 30 deg. C temperature difference between the pool bottom and pipe outlet elevation was measured. Elevations above the pipe outlet indicated almost linear rise until the end of steam discharge. During the cooling periods the measurements of the bottom third of the pool first had an increasing trend although there was no heat input from outside. This was due to thermal diffusion downwards from the higher elevations. Heat-up in the gas space of the wet well was quite strong, first due to compression by pressure build-up and then by heat conduction from the hot dry well compartment via the intermediate floor and test vessel walls and by convection from the upper layers of the hot pool water. The gas space
On the formation of couplet-style stratifications
Yu-Hai WANG; Yan-Hong WANG; Li-Qun TANG
2008-01-01
Couplet-style stratifications refer to the sedimentary sequences that consist of alternating coarse-grain-dominated bed and fine-grain-dominated bed.with or without a sandwiched middle-sized bed.The formation mechanism is complicated due to the interplays of varying driving force(s),sediment supply & transport and topographic configurations.This paper presents a comprehensive overview of such characteristic bedforms,which emerge during the transport and deposition of non-uniform sediments.The leading formative models include water-stage variation,gravel-overpassing process,superimposition of bedload sheets and avalanching process(inverse grading).Each process might produce similar or distinct sedimentological features with respect to grading,matrix content,grain attitude,bounding faces between beds and internal longitudinal/transverse geometry.The couplet-style stratified strata might play an active role in landscape evolution.
Aukland, S M; Westerhausen, R; Plessen, K J
2011-01-01
BACKGROUND AND PURPOSE: Several studies suggest that VLBW is associated with a reduced CC size later in life. We aimed to clarify this in a prospective, controlled study of 19-year-olds, hypothesizing that those with LBWs had smaller subregions of CC than the age-matched controls, even after...
Park, Seung Shik; Kim, Young J; Kang, Chang Hee
2007-05-01
To analyze polycyclic aromatic hydrocarbons (PAHs) at an urban site in Seoul, South Korea, 24-hr ambient air PM2.5 samples were collected during five intensive sampling periods between November 1998 and December 1999. To determine the PAH size distribution, 3-day size-segregated aerosol samples were also collected in December 1999. Concentrations of the 16 PAHs in the PM2.5 particles ranged from 3.9 to 119.9 ng m(-3) with a mean of 24.3 ng m(-3). An exceptionally high concentration of PAHs( approximately 120 ng m(-3)) observed during a haze event in December 1999 was likely influenced more by diesel vehicle exhaust than by gasoline exhaust, as well as air stagnation, as evidenced by the low carbon monoxide/elemental carbon (CO/EC) ratio of 205 found in this study and results reported by previous studies. The total PAHs associated with the size-segregated particles showed unimodal distributions. Compared to the unimodal size distributions of PAHs with modal peaks at particles during transport to the sampling site. Further, the fraction of PAHs associated with coarse particles(> 1.8 microm) increased as the molecular weight of the PAHs decreased due to volatilization of fine particles followed by condensation onto coarse particles.
Robert D. Otto
2003-04-01
Full Text Available Wildlife radio-telemetry and tracking projects often determine a priori required sample sizes by statistical means or default to the maximum number that can be maintained within a limited budget. After initiation of such projects, little attention is focussed on effective sample size requirements, resulting in lack of statistical power. The Department of National Defence operates a base in Labrador, Canada for low level jet fighter training activities, and maintain a sample of satellite collars on the George River caribou (Rangifer tarandus caribou herd of the region for spatial avoidance mitiga¬tion purposes. We analysed existing location data, in conjunction with knowledge of life history, to develop estimates of satellite collar sample sizes required to ensure adequate mitigation of GRCH. We chose three levels of probability in each of six annual caribou seasons. Estimated number of collars required ranged from 15 to 52, 23 to 68, and 36 to 184 for 50%, 75%, and 90% probability levels, respectively, depending on season. Estimates can be used to make more informed decisions about mitigation of GRCH, and, generally, our approach provides a means to adaptively assess radio collar sam¬ple sizes for ongoing studies.
Spybrook, Jessaca; Puente, Anne Cullen; Lininger, Monica
2013-01-01
This article examines changes in the research design, sample size, and precision between the planning phase and implementation phase of group randomized trials (GRTs) funded by the Institute of Education Sciences. Thirty-eight GRTs funded between 2002 and 2006 were examined. Three studies revealed changes in the experimental design. Ten studies…
The objective of this research was to examine diet- and body size-related attitudes and behaviors associated with supplement use in a representative sample of fourth-grade students in Texas. The research design consisted of cross-sectional data from the School Physical Activity and Nutrition study, ...
Treen, Emily; Atanasova, Christina; Pitt, Leyland; Johnson, Michael
2016-01-01
Marketing instructors using simulation games as a way of inducing some realism into a marketing course are faced with many dilemmas. Two important quandaries are the optimal size of groups and how much of the students' time should ideally be devoted to the game. Using evidence from a very large sample of teams playing a simulation game, the study…
Thorlund, Kristian; Anema, Aranka; Mills, Edward
2010-01-01
To illustrate the utility of statistical monitoring boundaries in meta-analysis, and provide a framework in which meta-analysis can be interpreted according to the adequacy of sample size. To propose a simple method for determining how many patients need to be randomized in a future trial before ...
Daniele Tonina; Alberto Bellin
2008-01-01
Pore-scale dispersion (PSD), aquifer heterogeneity, sampling volume, and source size influence solute concentrations of conservative tracers transported in heterogeneous porous formations. In this work, we developed a new set of analytical solutions for the concentration ensemble mean, variance, and coefficient of variation (CV), which consider the effects of all these...
Rogan, Joanne C.; Keselman, H. J.
1977-01-01
The effects of variance heterogeneity on the empirical probability of a Type I error for the analysis of variance (ANOVA) F-test are examined. The rate of Type I error varies as a function of the degree of variance heterogeneity, and the ANOVA F-test is not always robust to variance heterogeneity when sample sizes are equal. (Author/JAC)
Evidence for a Global Sampling Process in Extraction of Summary Statistics of Item Sizes in a Set.
Tokita, Midori; Ueda, Sachiyo; Ishiguchi, Akira
2016-01-01
Several studies have shown that our visual system may construct a "summary statistical representation" over groups of visual objects. Although there is a general understanding that human observers can accurately represent sets of a variety of features, many questions on how summary statistics, such as an average, are computed remain unanswered. This study investigated sampling properties of visual information used by human observers to extract two types of summary statistics of item sets, average and variance. We presented three models of ideal observers to extract the summary statistics: a global sampling model without sampling noise, global sampling model with sampling noise, and limited sampling model. We compared the performance of an ideal observer of each model with that of human observers using statistical efficiency analysis. Results suggest that summary statistics of items in a set may be computed without representing individual items, which makes it possible to discard the limited sampling account. Moreover, the extraction of summary statistics may not necessarily require the representation of individual objects with focused attention when the sets of items are larger than 4.
Mulet, R.; Diaz, O.; Altshuler, E. [Superconductivity Laboratory, IMRE-Physics Faculty, University of Havana, La Habana (Cuba)
1997-10-01
The percolative character of the current paths and the self-field effects were considered to estimate optimal sample dimensions for the transport current of a granular superconductor by means of a Monte Carlo algorithm and critical-state model calculations. We showed that, under certain conditions, self-field effects are negligible and the J{sub c} dependence on sample dimensions is determined by the percolative character of the current. Optimal dimensions are demonstrated to be a function of the fraction of superconducting phase in the sample. (author)
Arnup, Sarah J; McKenzie, Joanne E; Hemming, Karla; Pilcher, David; Forbes, Andrew B
2017-08-15
In a cluster randomised crossover (CRXO) design, a sequence of interventions is assigned to a group, or 'cluster' of individuals. Each cluster receives each intervention in a separate period of time, forming 'cluster-periods'. Sample size calculations for CRXO trials need to account for both the cluster randomisation and crossover aspects of the design. Formulae are available for the two-period, two-intervention, cross-sectional CRXO design, however implementation of these formulae is known to be suboptimal. The aims of this tutorial are to illustrate the intuition behind the design; and provide guidance on performing sample size calculations. Graphical illustrations are used to describe the effect of the cluster randomisation and crossover aspects of the design on the correlation between individual responses in a CRXO trial. Sample size calculations for binary and continuous outcomes are illustrated using parameters estimated from the Australia and New Zealand Intensive Care Society - Adult Patient Database (ANZICS-APD) for patient mortality and length(s) of stay (LOS). The similarity between individual responses in a CRXO trial can be understood in terms of three components of variation: variation in cluster mean response; variation in the cluster-period mean response; and variation between individual responses within a cluster-period; or equivalently in terms of the correlation between individual responses in the same cluster-period (within-cluster within-period correlation, WPC), and between individual responses in the same cluster, but in different periods (within-cluster between-period correlation, BPC). The BPC lies between zero and the WPC. When the WPC and BPC are equal the precision gained by crossover aspect of the CRXO design equals the precision lost by cluster randomisation. When the BPC is zero there is no advantage in a CRXO over a parallel-group cluster randomised trial. Sample size calculations illustrate that small changes in the specification of
Dow Geoffrey S
2003-02-01
Full Text Available Abstract Background There is no known biochemical basis for the adverse neurological events attributed to mefloquine. Identification of genes modulated by toxic agents using microarrays may provide sufficient information to generate hypotheses regarding their mode of action. However, this utility may be compromised if sample sizes are too low or the filtering methods used to identify differentially expressed genes are inappropriate. Methods The transcriptional changes induced in rat neuroblastoma cells by a physiological dose of mefloquine (10 micro-molar were investigated using Affymetrix arrays. A large sample size was used (total of 16 arrays. Genes were ranked by P-value (t-test. RT-PCR was used to confirm (or reject the expression changes of several of the genes with the lowest P-values. Different P-value filtering methods were compared in terms of their ability to detect these differentially expressed genes. A retrospective power analysis was then performed to determine whether the use of lower sample sizes might also have detected those genes with altered transcription. Results Based on RT-PCR, mefloquine upregulated cJun, IkappaB and GADD153. Reverse Holm-Bonferroni P-value filtering was superior to other methods in terms of maximizing detection of differentially expressed genes but not those with unaltered expression. Reduction of total microarray sample size ( Conclusions Adequate sample sizes and appropriate selection of P-value filtering methods are essential for the reliable detection of differentially expressed genes. The changes in gene expression induced by mefloquine suggest that the ER might be a neuronal target of the drug.
Increasing the Knowledge of Stratification in Shallow Coastal Environments
Ojo, T.; Bonner, J.; Hodges, B.; Maidment, D.; Montagna, P.; Minsker, B.
2006-12-01
A testbed has been established using Corpus Christi Bay as an environmental field facility to study the phenomenon of hypoxia that has been observed to develop at certain periods during the year. Stratification affects vertical turbulent mixing of heat, momentum and mass (or constituents) within the water column, in turn influencing the transport of material. The mixing threshold is dependent on the value of the Richardson Number, Ri with inhibition due to stratification occurring at low values ( 0.25) of Ri. Corpus Christi Bay with average depth of ~3 m is the largest among a system of five bays has been known to stratify due to inflows of hypersaline water (up to 50 psu) from adjoining bays, the Laguna Madre and Oso Bay. Laguna Madre is separated from the Gulf of Mexico by a barrier island and becomes hypersaline because of the imbalance between inflow of freshwater and bay evaporation. Hypersalinity also occurs in Oso Bay due to anthropogenic forcing from a power plant that draws 400 MGD of cooling water from the upper Laguna Madre, discharging waste water into Oso Bay. Several wastewater treatment plants also discharge directly into Oso Bay or its tributary streams. The objective of this study is to develop a methodology for prescribing a set of parameters required for modeling and characterization of hypoxia in this shallow wind-driven bay. The extent to which Ri is dependent on external forcing at the surface boundary was measured using our fully instrumented sensor platforms. Each sensor platform includes sensors for synchronic near-surface meteorological (wind velocity, barometric pressure, air temperature) and water column oceanographic (current, water temperature, conductivity, particle size distribution, particulate concentration, dissolved oxygen, nutrient) variables. These were measured using fixed and mobile vertical profiling sensor platforms. A 2D hydrodynamic model was initially developed for the bay and results indicate that water mass is
Is risk stratification ever the same as 'profiling'?
Braithwaite, R Scott; Stevens, Elizabeth R; Caplan, Arthur
2016-05-01
Physicians engage in risk stratification as a normative part of their professional duties. Risk stratification has the potential to be beneficial in many ways, and implicit recognition of this potential benefit underlies its acceptance as a cornerstone of the medical profession. However, risk stratification also has the potential to be harmful. We argue that 'profiling' is a term that corresponds to risk stratification strategies in which there is concern that ethical harms exceed likely or proven benefits. In the case of risk stratification for health goals, this would occur most frequently if benefits were obtained by threats to justice, autonomy or privacy. We discuss implications of the potential overlap between risk stratification and profiling for researchers and for clinicians, and we consider whether there are salient characteristics that make a particular risk stratification algorithm more or less likely to overlap with profiling, such as whether the risk stratification algorithm is based on voluntary versus non-voluntary characteristics, based on causal versus non-causal characteristics, or based on signifiers of historical disadvantage. We also discuss the ethical challenges created when a risk stratification scheme helps all subgroups but some more than others, or when risk stratification harms some subgroups but benefits the aggregate group.
Wang, Lei; Li, Zhenyu; Jiang, Jia; An, Taiyu; Qin, Hongwei; Hu, Jifan
2017-01-01
In the present work, we demonstrate that ferromagnetic resonance and magneto-permittivity resonance can be observed in appropriate microwave frequencies at room temperature for multiferroic nano-BiFeO3/paraffin composite sample with an appropriate sample-thickness (such as 2 mm). Ferromagnetic resonance originates from the room-temperature weak ferromagnetism of nano-BiFeO3. The observed magneto-permittivity resonance in multiferroic nano-BiFeO3 is connected with the dynamic magnetoelectric coupling through Dzyaloshinskii-Moriya (DM) magnetoelectric interaction or the combination of magnetostriction and piezoelectric effects. In addition, we experimentally observed the resonance of negative imaginary permeability for nano BiFeO3/paraffin toroidal samples with longer sample thicknesses D=3.7 and 4.9 mm. Such resonance of negative imaginary permeability belongs to sample-size resonance.
All-reflective UV-VIS-NIR transmission and fluorescence spectrometer for μm-sized samples
Friedrich O. Kirchner
2014-07-01
Full Text Available We report on an optical transmission spectrometer optimized for tiny samples. The setup is based on all-reflective parabolic optics and delivers broadband operation from 215 to 1030 nm. A fiber-coupled light source is used for illumination and a fiber-coupled miniature spectrometer for detection. The diameter of the probed area is less than 200 μm for all wavelengths. We demonstrate the capability to record transmission, absorption, reflection, fluorescence and refractive indices of tiny and ultrathin sample flakes with this versatile device. The performance is validated with a solid state wavelength standard and with dye solutions.
Steven, E.; Jobiliong, E.; Eugenio, P. M.; Brooks, J. S.
2012-04-01
A procedure for fabricating adhesive stamp electrodes based on gold coated adhesive tape used to measure electronic transport properties of supra-micron samples in the lateral range 10-100 μm and thickness >1 μm is described. The electrodes can be patterned with a ˜4 μm separation by metal deposition through a mask using Nephila clavipes spider dragline silk fibers. Ohmic contact is made by adhesive lamination of a sample onto the patterned electrodes. The performance of the electrodes with temperature and magnetic field is demonstrated for the quasi-one-dimensional organic conductor (TMTSF)2PF6 and single crystal graphite, respectively.
Lugo, Jorge; Sosa, Victor
1999-10-01
The repulsion force between a cylindrical superconductor in the Meissner state and a small permanent magnet was calculated under the assumption that the superconductor was formed by a continuous array of dipoles distributed in the finite volume of the sample. After summing up the dipole-dipole interactions with the magnet, we obtained analytical expressions for the levitation force as a function of the superconductor-magnet distance, radius and thickness of the sample. We analyzed two configurations, with the magnet in a horizontal or vertical orientation.
Morecroft Michael D
2001-07-01
Full Text Available Abstract Background The Resource Dispersion Hypothesis (RDH proposes a mechanism for the passive formation of social groups where resources are dispersed, even in the absence of any benefits of group living per se. Despite supportive modelling, it lacks empirical testing. The RDH predicts that, rather than Territory Size (TS increasing monotonically with Group Size (GS to account for increasing metabolic needs, TS is constrained by the dispersion of resource patches, whereas GS is independently limited by their richness. We conducted multiple-year tests of these predictions using data from the long-term study of badgers Meles meles in Wytham Woods, England. The study has long failed to identify direct benefits from group living and, consequently, alternative explanations for their large group sizes have been sought. Results TS was not consistently related to resource dispersion, nor was GS consistently related to resource richness. Results differed according to data groupings and whether territories were mapped using minimum convex polygons or traditional methods. Habitats differed significantly in resource availability, but there was also evidence that food resources may be spatially aggregated within habitat types as well as between them. Conclusions This is, we believe, the largest ever test of the RDH and builds on the long-term project that initiated part of the thinking behind the hypothesis. Support for predictions were mixed and depended on year and the method used to map territory borders. We suggest that within-habitat patchiness, as well as model assumptions, should be further investigated for improved tests of the RDH in the future.
Thompson, Amanda L; Adair, Linda S; Bentley, Margaret E
2013-03-01
The prevalence of overweight among infants and toddlers has increased dramatically in the past three decades, highlighting the importance of identifying factors contributing to early excess weight gain, particularly in high-risk groups. Parental feeding styles and the attitudes and behaviors that characterize parental approaches to maintaining or modifying children's eating behavior are an important behavioral component shaping early obesity risk. Using longitudinal data from the Infant Care and Risk of Obesity Study, a cohort study of 217 African-American mother-infant pairs with feeding styles, dietary recalls, and anthropometry collected from 3 to 18 months of infant age, we examined the relationship between feeding styles, infant diet, and weight-for-age and sum of skinfolds. Longitudinal mixed models indicated that higher pressuring and indulgent feeding style scores were positively associated with greater infant energy intake, reduced odds of breastfeeding, and higher levels of age-inappropriate feeding of liquids and solids, whereas restrictive feeding styles were associated with lower energy intake, higher odds of breastfeeding, and reduced odds of inappropriate feeding. Pressuring and restriction were also oppositely related to infant size with pressuring associated with lower infant weight-for-age and restriction with higher weight-for-age and sum of skinfolds. Infant size also predicted maternal feeding styles in subsequent visits indicating that the relationship between size and feeding styles is likely bidirectional. Our results suggest that the degree to which parents are pressuring or restrictive during feeding shapes the early feeding environment and, consequently, may be an important environmental factor in the development of obesity. Copyright © 2012 The Obesity Society.
Plasma ion stratification by weak planar shocks
Simakov, Andrei N.; Keenan, Brett D.; Taitano, William T.; Chacón, Luis
2017-09-01
We derive fluid equations for describing steady-state planar shocks of a moderate strength ( 0 shock Mach number) propagating through an unmagnetized quasineutral collisional plasma comprising two separate ion species. In addition to the standard fluid shock quantities, such as the total mass density, mass-flow velocity, and electron and average ion temperatures, the equations describe shock stratification in terms of variations in the relative concentrations and temperatures of the two ion species along the shock propagation direction. We have solved these equations analytically for weak shocks ( 0 shocks, and they have been used to verify kinetic simulations of shocks in multi-ion plasmas.
Damiani, Rick [National Renewable Energy Lab. (NREL), Golden, CO (United States)
2016-02-08
This manual summarizes the theory and preliminary verifications of the JacketSE module, which is an offshore jacket sizing tool that is part of the Wind-Plant Integrated System Design & Engineering Model toolbox. JacketSE is based on a finite-element formulation and on user-prescribed inputs and design standards' criteria (constraints). The physics are highly simplified, with a primary focus on satisfying ultimate limit states and modal performance requirements. Preliminary validation work included comparing industry data and verification against ANSYS, a commercial finite-element analysis package. The results are encouraging, and future improvements to the code are recommended in this manual.
Analysis of Nb3Sn Strand Microstructure After Full-size SULTAN Test of ITER TF Conductor Sample
Kaverin, D.; Potanina, L.; Shutov, K.; Vysotsky, V.; Tronza, V.; Mitin, A.; Abdyukhanov, I.; Alekseev, M.
The study of defects generated in superconducting filaments of Nb3Sn strands under electromagnetic and thermal cycling was carried out for the TFRF3 cable-in-conduit-conductor (CICC) sample that passed final testing inthe SULTAN test facility. The TFRF3 sample was manufactured forthe qualification of the RF Toroidal Field (TF) CICC. The strand samples were taken from different locations in the cross-section of TFRF3 and different positions along its axis in relation to background magnetic field. Qualitative and quantitative analysis of defects were carried out using metallographic analysis of images obtained by Laser Scanning Microscope. We analyzed number, type, and distribution of defects in filaments of the Nb3Sn strand samples extracted from different petals of TFRF3 in dependence on thestrand location in the cross-section (the center of petal, nearby the spiral, nearby the outer jacket) in the high field zone (HFZ). The results about the defects amount and their distribution are presented and discussed.
Lorenzo, C; Carretero, J M; Arsuaga, J L; Gracia, A; Martínez, I
1998-05-01
A sexual dimorphism more marked than in living humans has been claimed for European Middle Pleistocene humans, Neandertals and prehistoric modern humans. In this paper, body size and cranial capacity variation are studied in the Sima de los Huesos Middle Pleistocene sample. This is the largest sample of non-modern humans found to date from one single site, and with all skeletal elements represented. Since the techniques available to estimate the degree of sexual dimorphism in small palaeontological samples are all unsatisfactory, we have used the bootstraping method to asses the magnitude of the variation in the Sima de los Huesos sample compared to modern human intrapopulational variation. We analyze size variation without attempting to sex the specimens a priori. Anatomical regions investigated are scapular glenoid fossa; acetabulum; humeral proximal and distal epiphyses; ulnar proximal epiphysis; radial neck; proximal femur; humeral, femoral, ulnar and tibial shaft; lumbosacral joint; patella; calcaneum; and talar trochlea. In the Sima de los Huesos sample only the humeral midshaft perimeter shows an unusual high variation (only when it is expressed by the maximum ratio, not by the coefficient of variation). In spite of that the cranial capacity range at Sima de los Huesos almost spans the rest of the European and African Middle Pleistocene range. The maximum ratio is in the central part of the distribution of modern human samples. Thus, the hypothesis of a greater sexual dimorphism in Middle Pleistocene populations than in modern populations is not supported by either cranial or postcranial evidence from Sima de los Huesos.