Averaging and sampling for magnetic-observatory hourly data
J. J. Love
2010-11-01
Full Text Available A time and frequency-domain analysis is made of the effects of averaging and sampling methods used for constructing magnetic-observatory hourly data values. Using 1-min data as a proxy for continuous, geomagnetic variation, we construct synthetic hourly values of two standard types: instantaneous "spot" measurements and simple 1-h "boxcar" averages. We compare these average-sample types with others: 2-h average, Gaussian, and "brick-wall" low-frequency-pass. Hourly spot measurements provide a statistically unbiased representation of the amplitude range of geomagnetic-field variation, but as a representation of continuous field variation over time, they are significantly affected by aliasing, especially at high latitudes. The 1-h, 2-h, and Gaussian average-samples are affected by a combination of amplitude distortion and aliasing. Brick-wall values are not affected by either amplitude distortion or aliasing, but constructing them is, in an operational setting, relatively more difficult than it is for other average-sample types. It is noteworthy that 1-h average-samples, the present standard for observatory hourly data, have properties similar to Gaussian average-samples that have been optimized for a minimum residual sum of amplitude distortion and aliasing. For 1-h average-samples from medium and low-latitude observatories, the average of the combination of amplitude distortion and aliasing is less than the 5.0 nT accuracy standard established by Intermagnet for modern 1-min data. For medium and low-latitude observatories, average differences between monthly means constructed from 1-min data and monthly means constructed from any of the hourly average-sample types considered here are less than the 1.0 nT resolution of standard databases. We recommend that observatories and World Data Centers continue the standard practice of reporting simple 1-h-average hourly values.
29 CFR 778.100 - The maximum-hours provisions.
2010-07-01
... POLICY OR INTERPRETATION NOT DIRECTLY RELATED TO REGULATIONS OVERTIME COMPENSATION The Overtime Pay... hours and overtime compensation for employees who are within the general coverage of the Act and are not... permitted for the employment of such employees in any workweek without extra compensation for overtime,...
10 CFR 170.20 - Average cost per professional staff-hour.
2010-01-01
... Provisions § 170.20 Average cost per professional staff-hour. Fees for permits, licenses, amendments, renewals, special projects, 10 CFR part 55 re-qualification and replacement examinations and tests, other... 10 Energy 2 2010-01-01 2010-01-01 false Average cost per professional staff-hour. 170.20...
Maximum Likelihood Estimation of Multivariate Autoregressive-Moving Average Models.
1977-02-01
maximizing the same have been proposed i) in time domain by Box and Jenkins [41. Astrom [3J, Wilson [23 1, and Phadke [161, and ii) in frequency domain by...moving average residuals and other convariance matrices with linear structure ”, Anna/s of Staustics, 3. 3. Astrom , K. J. (1970), Introduction to
MAXIMUM LIKELIHOOD ESTIMATION FOR PERIODIC AUTOREGRESSIVE MOVING AVERAGE MODELS.
Vecchia, A.V.
1985-01-01
A useful class of models for seasonal time series that cannot be filtered or standardized to achieve second-order stationarity is that of periodic autoregressive moving average (PARMA) models, which are extensions of ARMA models that allow periodic (seasonal) parameters. An approximation to the exact likelihood for Gaussian PARMA processes is developed, and a straightforward algorithm for its maximization is presented. The algorithm is tested on several periodic ARMA(1, 1) models through simulation studies and is compared to moment estimation via the seasonal Yule-Walker equations. Applicability of the technique is demonstrated through an analysis of a seasonal stream-flow series from the Rio Caroni River in Venezuela.
Stone, Wesley W.; Gilliom, Robert J.; Crawford, Charles G.
2008-01-01
Regression models were developed for predicting annual maximum and selected annual maximum moving-average concentrations of atrazine in streams using the Watershed Regressions for Pesticides (WARP) methodology developed by the National Water-Quality Assessment Program (NAWQA) of the U.S. Geological Survey (USGS). The current effort builds on the original WARP models, which were based on the annual mean and selected percentiles of the annual frequency distribution of atrazine concentrations. Estimates of annual maximum and annual maximum moving-average concentrations for selected durations are needed to characterize the levels of atrazine and other pesticides for comparison to specific water-quality benchmarks for evaluation of potential concerns regarding human health or aquatic life. Separate regression models were derived for the annual maximum and annual maximum 21-day, 60-day, and 90-day moving-average concentrations. Development of the regression models used the same explanatory variables, transformations, model development data, model validation data, and regression methods as those used in the original development of WARP. The models accounted for 72 to 75 percent of the variability in the concentration statistics among the 112 sampling sites used for model development. Predicted concentration statistics from the four models were within a factor of 10 of the observed concentration statistics for most of the model development and validation sites. Overall, performance of the models for the development and validation sites supports the application of the WARP models for predicting annual maximum and selected annual maximum moving-average atrazine concentration in streams and provides a framework to interpret the predictions in terms of uncertainty. For streams with inadequate direct measurements of atrazine concentrations, the WARP model predictions for the annual maximum and the annual maximum moving-average atrazine concentrations can be used to characterize
44 CFR 353.5 - Average cost per FEMA professional staff-hour.
2010-10-01
... 44 Emergency Management and Assistance 1 2010-10-01 2010-10-01 false Average cost per FEMA professional staff-hour. 353.5 Section 353.5 Emergency Management and Assistance FEDERAL EMERGENCY MANAGEMENT... of exercise objectives and scenarios, pre-exercise logistics, exercise conduct and...
Estimation of the diffuse radiation fraction for hourly, daily and monthly-average global radiation
Erbs, D. G.; Klein, S. A.; Duffie, J. A.
1982-01-01
Hourly pyrheliometer and pyranometer data from four U.S. locations are used to establish a relationship between the hourly diffuse fraction and the hourly clearness index. This relationship is compared to the relationship established by Orgill and Hollands (1977) and to a set of data from Highett, Australia, and agreement is within a few percent in both cases. The transient simulation program TRNSYS is used to calculate the annual performance of solar energy systems using several correlations. For the systems investigated, the effect of simulating the random distribution of the hourly diffuse fraction is negligible. A seasonally dependent daily diffuse correlation is developed from the data, and this daily relationship is used to derive a correlation for the monthly-average diffuse fraction.
Time series models to simulate and forecast hourly averaged wind speeds in Quetta, Pakistan
Lalarukh Kamal [Balochistan University, Quetta (Pakistan). Dept. of Mathematics; Yasmin Zahra Jafri [Balochistan University, Quetta (Pakistan). Dept. of Statistics
1997-07-01
Stochastic simulation and forecast models of hourly average wind speeds are presented. Time series models take into account several basic features of wind speed data including autocorrelation, non-Gaussian distribution and diurnal nonstationarity. The positive correlation between consecutive wind speed observations is taken into account by fitting an ARMA (p,q) process to wind speed data transformed to make their distribution approximately Gaussian and standardized to remove scattering of transformed data. Diurnal variations have been taken into account to observe forecasts and its dependence on lead times. We find the ARMA (p,q) model suitable for prediction intervals and probability forecasts. (author)
A rural-urban comparison of allied health professionals' average hourly wage.
Richardson, Indira; Slifkin, Rebecca; Randolph, Randy; Holmes, George M
2010-01-01
Nationwide, demand for allied health services is projected to grow significantly in the next several decades, and there is evidence that allied health shortages already exist in many states. Given the longstanding history of health professional shortages in rural areas, the existing and impending shortages in allied health professions may be particularly acute in these areas. To assess whether rural areas are potentially at a recruiting disadvantage because of relative wages, this report uses data from the Bureau of Labor Statistics to describe the extent to which rural-urban differentials exist in wages for eight allied health professions, focusing on professions that are both likely to be found in rural communities and have adequate data to support hourly wage estimates. Overall the data show that the national average wage of each of the eight allied health professions is higher in metropolitan than nonmetropolitan areas. On average, the unadjusted rural hourly wage is 10.3% less than the urban wage, although the extent of the difference varies by profession and by geographic area. Adjustment for the cost of living narrows the discrepancy, but does not eliminate it. It is likely that rural providers in areas with the greatest wage discrepancies find it more difficult to recruit allied health professionals, but the extent to which this is the case needs to be assessed through further research with data on workforce vacancy rates.
A new solar signal: Average maximum sunspot magnetic fields independent of activity cycle
Livingston, William
2016-01-01
Over the past five years, 2010-2015, we have observed, in the near infrared (IR), the maximum magnetic field strengths for 4145 sunspot umbrae. Herein we distinguish field strengths from field flux. (Most solar magnetographs measure flux). Maximum field strength in umbrae is co-spatial with the position of umbral minimum brightness (Norton and Gilman, 2004). We measure field strength by the Zeeman splitting of the Fe 15648.5 A spectral line. We show that in the IR no cycle dependence on average maximum field strength (2050 G) has been found +/- 20 Gauss. A similar analysis of 17,450 spots observed by the Helioseismic and Magnetic Imager onboard the Solar Dynamics Observatory reveal the same cycle independence +/- 0.18 G., or a variance of 0.01%. This is found not to change over the ongoing 2010-2015 minimum to maximum cycle. Conclude the average maximum umbral fields on the Sun are constant with time.
The effects of disjunct sampling and averaging time on maximum mean wind speeds
Larsén, Xiaoli Guo; Mann, J.
2006-01-01
Conventionally, the 50-year wind is calculated on basis of the annual maxima of consecutive 10-min averages. Very often, however, the averages are saved with a temporal spacing of several hours. We call it disjunct sampling. It may also happen that the wind speeds are averaged over a longer time...... period before being saved. In either case, the extreme wind will be underestimated. This paper investigates the effects of the disjunct sampling interval and the averaging time on the attenuation of the extreme wind estimation by means of a simple theoretical approach as well as measurements...
Variability of maximum and mean average temperature across Libya (1945-2009)
Ageena, I.; Macdonald, N.; Morse, A. P.
2014-08-01
Spatial and temporal variability in daily maximum and mean average daily temperature, monthly maximum and mean average monthly temperature for nine coastal stations during the period 1956-2009 (54 years), and annual maximum and mean average temperature for coastal and inland stations for the period 1945-2009 (65 years) across Libya are analysed. During the period 1945-2009, significant increases in maximum temperature (0.017 °C/year) and mean average temperature (0.021 °C/year) are identified at most stations. Significantly, warming in annual maximum temperature (0.038 °C/year) and mean average annual temperatures (0.049 °C/year) are observed at almost all study stations during the last 32 years (1978-2009). The results show that Libya has witnessed a significant warming since the middle of the twentieth century, which will have a considerable impact on societies and the ecology of the North Africa region, if increases continue at current rates.
Computational complexity of some maximum average weight problems with precedence constraints
Faigle, Ulrich; Kern, Walter
1994-01-01
Maximum average weight ideal problems in ordered sets arise from modeling variants of the investment problem and, in particular, learning problems in the context of concepts with tree-structured attributes in artificial intelligence. Similarly, trying to construct tests with high reliability leads t
Recovery of the histogram of hourly ozone distribution from weekly average concentrations
Olcese, Luis E. [Departamento de Fisico Quimica/INFIQC, Facultad de Ciencias Quimicas, Universidad Nacional de Cordoba, 5000 Cordoba (Argentina)]. E-mail: lolcese@fcq.unc.edu.ar; Toselli, Beatriz M. [Departamento de Fisico Quimica/INFIQC, Facultad de Ciencias Quimicas, Universidad Nacional de Cordoba, 5000 Cordoba (Argentina)
2006-05-15
A simple method is presented for estimating hourly distribution of air pollutants, based on data collected by passive sensors on a weekly or bi-weekly basis with no need for previous measurements at a site. In order for this method to be applied to locations where no hourly records are available, reference data from other sites are required to generate calibration histograms. The proposed procedure allows one to obtain the histogram of hourly ozone values during a given week with an error of about 30%, which is good considering the simplicity of this approach. This method can be a valuable tool for sites that lack previous hourly records of pollutant ambient concentrations, where it can be used to verify compliance with regulations or to estimate the AOT40 index with an acceptable degree of exactitude. - The histogram of hourly ozone distribution can be obtained based on passive sensor data.
Analytical expressions for maximum wind turbine average power in a Rayleigh wind regime
Carlin, P.W.
1996-12-01
Average or expectation values for annual power of a wind turbine in a Rayleigh wind regime are calculated and plotted as a function of cut-out wind speed. This wind speed is expressed in multiples of the annual average wind speed at the turbine installation site. To provide a common basis for comparison of all real and imagined turbines, the Rayleigh-Betz wind machine is postulated. This machine is an ideal wind machine operating with the ideal Betz power coefficient of 0.593 in a Rayleigh probability wind regime. All other average annual powers are expressed in fractions of that power. Cases considered include: (1) an ideal machine with finite power and finite cutout speed, (2) real machines operating in variable speed mode at their maximum power coefficient, and (3) real machines operating at constant speed.
Recovery of the histogram of hourly ozone distribution from weekly average concentrations.
Olcese, Luis E; Toselli, Beatriz M
2006-05-01
A simple method is presented for estimating hourly distribution of air pollutants, based on data collected by passive sensors on a weekly or bi-weekly basis with no need for previous measurements at a site. In order for this method to be applied to locations where no hourly records are available, reference data from other sites are required to generate calibration histograms. The proposed procedure allows one to obtain the histogram of hourly ozone values during a given week with an error of about 30%, which is good considering the simplicity of this approach. This method can be a valuable tool for sites that lack previous hourly records of pollutant ambient concentrations, where it can be used to verify compliance with regulations or to estimate the AOT40 index with an acceptable degree of exactitude.
Sadjadi, Firooz A; Mahalanobis, Abhijit
2006-05-01
We report the development of a technique for adaptive selection of polarization ellipse tilt and ellipticity angles such that the target separation from clutter is maximized. From the radar scattering matrix [S] and its complex components, in phase and quadrature phase, the elements of the Mueller matrix are obtained. Then, by means of polarization synthesis, the radar cross section of the radar scatters are obtained at different transmitting and receiving polarization states. By designing a maximum average correlation height filter, we derive a target versus clutter distance measure as a function of four transmit and receive polarization state angles. The results of applying this method on real synthetic aperture radar imagery indicate a set of four transmit and receive angles that lead to maximum target versus clutter discrimination. These optimum angles are different for different targets. Hence, by adaptive control of the state of polarization of polarimetric radar, one can noticeably improve the discrimination of targets from clutter.
29 CFR 553.230 - Maximum hours standards for work periods of 7 to 28 days-section 7(k).
2010-07-01
...-section 7(k). 553.230 Section 553.230 Labor Regulations Relating to Labor (Continued) WAGE AND HOUR... Compensation Rules § 553.230 Maximum hours standards for work periods of 7 to 28 days—section 7(k). (a) For... 28 consecutive days, no overtime compensation is required under section 7(k) until the number...
Kruse, Marcelo Lapa; Kruse, José Cláudio Lupi; Leiria, Tiago Luiz Luz; Pires, Leonardo Martins; Gensas, Caroline Saltz; Gomes, Daniel Garcia; Boris, Douglas; Mantovani, Augusto; Lima, Gustavo Glotz de
2014-12-01
Occurrences of asymptomatic atrial fibrillation (AF) are common. It is important to identify AF because it increases morbidity and mortality. 24-hour Holter has been used to detect paroxysmal AF (PAF). The objective of this study was to investigate the relationship between occurrence of PAF in 24-hour Holter and the symptoms of the population studied. Cross-sectional study conducted at a cardiology hospital. 11,321 consecutive 24-hour Holter tests performed at a referral service were analyzed. Patients with pacemakers or with AF throughout the recording were excluded. There were 75 tests (0.67%) with PAF. The mean age was 67 ± 13 years and 45% were female. The heart rate (HR) over the 24 hours was a minimum of 45 ± 8 bpm, mean of 74 ± 17 bpm and maximum of 151 ± 32 bpm. Among the tests showing PAF, only 26% had symptoms. The only factor tested that showed a correlation with symptomatic AF was maximum HR (165 ± 34 versus 147 ± 30 bpm) (P = 0.03). Use of beta blockers had a protective effect against occurrence of PAF symptoms (odds ratio: 0.24, P = 0.031). PAF is a rare event in 24-hour Holter. The maximum HR during the 24 hours was the only factor correlated with symptomatic AF, and use of beta blockers had a protective effect against AF symptom occurrence.
Zhang Zhang
2009-06-01
Full Text Available A major analytical challenge in computational biology is the detection and description of clusters of specified site types, such as polymorphic or substituted sites within DNA or protein sequences. Progress has been stymied by a lack of suitable methods to detect clusters and to estimate the extent of clustering in discrete linear sequences, particularly when there is no a priori specification of cluster size or cluster count. Here we derive and demonstrate a maximum likelihood method of hierarchical clustering. Our method incorporates a tripartite divide-and-conquer strategy that models sequence heterogeneity, delineates clusters, and yields a profile of the level of clustering associated with each site. The clustering model may be evaluated via model selection using the Akaike Information Criterion, the corrected Akaike Information Criterion, and the Bayesian Information Criterion. Furthermore, model averaging using weighted model likelihoods may be applied to incorporate model uncertainty into the profile of heterogeneity across sites. We evaluated our method by examining its performance on a number of simulated datasets as well as on empirical polymorphism data from diverse natural alleles of the Drosophila alcohol dehydrogenase gene. Our method yielded greater power for the detection of clustered sites across a breadth of parameter ranges, and achieved better accuracy and precision of estimation of clusters, than did the existing empirical cumulative distribution function statistics.
Zhang Zhang
2009-06-01
Full Text Available A major analytical challenge in computational biology is the detection and description of clusters of specified site types, such as polymorphic or substituted sites within DNA or protein sequences. Progress has been stymied by a lack of suitable methods to detect clusters and to estimate the extent of clustering in discrete linear sequences, particularly when there is no a priori specification of cluster size or cluster count. Here we derive and demonstrate a maximum likelihood method of hierarchical clustering. Our method incorporates a tripartite divide-and-conquer strategy that models sequence heterogeneity, delineates clusters, and yields a profile of the level of clustering associated with each site. The clustering model may be evaluated via model selection using the Akaike Information Criterion, the corrected Akaike Information Criterion, and the Bayesian Information Criterion. Furthermore, model averaging using weighted model likelihoods may be applied to incorporate model uncertainty into the profile of heterogeneity across sites. We evaluated our method by examining its performance on a number of simulated datasets as well as on empirical polymorphism data from diverse natural alleles of the Drosophila alcohol dehydrogenase gene. Our method yielded greater power for the detection of clustered sites across a breadth of parameter ranges, and achieved better accuracy and precision of estimation of clusters, than did the existing empirical cumulative distribution function statistics.
Wage and Labor Standards Administration (DOL), Washington, DC.
This report describes the 1966 amendments to the Fair Labor Standards Act and summarizes the findings of three 1969 studies of the economic effects of these amendments. The studies found that economic growth continued through the third phase of the amendments, beginning February 1, 1969, despite increased wage and hours restrictions for recently…
Malone, Stephen M.; McGue, Matt; Iacono, William G.
2010-01-01
Background: The maximum number of alcoholic drinks consumed in a single 24-hr period is an alcoholism-related phenotype with both face and empirical validity. It has been associated with severity of withdrawal symptoms and sensitivity to alcohol, genes implicated in alcohol metabolism, and amplitude of a measure of brain activity associated with…
Curtis, Gary P.; Lu, Dan; Ye, Ming
2015-01-01
While Bayesian model averaging (BMA) has been widely used in groundwater modeling, it is infrequently applied to groundwater reactive transport modeling because of multiple sources of uncertainty in the coupled hydrogeochemical processes and because of the long execution time of each model run. To resolve these problems, this study analyzed different levels of uncertainty in a hierarchical way, and used the maximum likelihood version of BMA, i.e., MLBMA, to improve the computational efficiency. This study demonstrates the applicability of MLBMA to groundwater reactive transport modeling in a synthetic case in which twenty-seven reactive transport models were designed to predict the reactive transport of hexavalent uranium (U(VI)) based on observations at a former uranium mill site near Naturita, CO. These reactive transport models contain three uncertain model components, i.e., parameterization of hydraulic conductivity, configuration of model boundary, and surface complexation reactions that simulate U(VI) adsorption. These uncertain model components were aggregated into the alternative models by integrating a hierarchical structure into MLBMA. The modeling results of the individual models and MLBMA were analyzed to investigate their predictive performance. The predictive logscore results show that MLBMA generally outperforms the best model, suggesting that using MLBMA is a sound strategy to achieve more robust model predictions relative to a single model. MLBMA works best when the alternative models are structurally distinct and have diverse model predictions. When correlation in model structure exists, two strategies were used to improve predictive performance by retaining structurally distinct models or assigning smaller prior model probabilities to correlated models. Since the synthetic models were designed using data from the Naturita site, the results of this study are expected to provide guidance for real-world modeling. Limitations of applying MLBMA to the
U.S. Geological Survey, Department of the Interior — This data set represents the average monthly maximum temperature in Celsius multiplied by 100 for 2002 compiled for every catchment of NHDPlus for the conterminous...
LIU Li-wei; LIU Hong; CHEN Guo-liang; HUANG Yi-ling; HAN Lu-lu; XU Zhi-min; JIANG Xiong-jing; LI Yi-shi
2010-01-01
Background Genetic factors can influence antihypertensive response to metoprolol, and many studies focused on the relationship between the genotype in β1-adrenergic receptor and blood pressure (BP), little was known about the association of angiotensin-converting enzyme (ACE) genotype with the therapeutic result of metoprolol. The present study aimed to investigate whether the ACE gene insertion (I) / deletion (D) polymorphism Is related to the response to metoprolol in Chinese Han hypertensive patients.Methods Ninety-six patients with essential hypertension received metoprolol (100 mg once daily) as monotherapy for 8 weeks. Twenty-four hours ambulatory blood pressure monitoring and dynamic electrocardiogram were performed before and after treatment. Genotyping analysis was performed using PCR. The association of the ACE gene I/D polymorphism with variations in BP and heart rate (HR) was observed after the 8-week treatment.Results The patients with ACE gene II polymorphism showed greater reduction in 24-hour average HR than those with ID or DD polymorphisms (P=0.045), no effect of this genotype on the reduction in seating HR or in BP was observed. After adjusting for age, gender, body mass index, BP and HR at baseline, the ACE gene I/D polymorphism was still an independent predictor for variations in 24-hour average HR.Conclusions The II polymorphism in ACE gene could be a candidate predictor for greater reduction in 24-hour average HR in Chinese Han hypertensive patients treated by metoprolol. Greater benefits would be obtained by patients with II polymorphism from the treatment with metoprolol. Larger studies are warranted to validate this finding.
6-hour maximum rain in Friuli Venezia Giulia: Climatology and ECMWF-based forecasts
Manzato, Agostino; Cicogna, Andrea; Pucillo, Arturo
2016-03-01
Friuli Venezia Giulia (FVG) is a region in Italy with very complex orography, having an annual rainfall amount that varies from about 900 mm on the coast to more than 3200 mm in the Julian Prealps. A network of 104 raingauges placed around the FVG territory was used to extract the absolute maximum rain accumulated every 6 h, during the period 16 February 2006 to 15 February 2015 (9 years). Interannual, annual, weekly and daily cycles of three classes of rain intensities are analyzed, finding that significant rainfalls (MaxRain > 5 mm) are more frequent in the May to mid-August period, while the heaviest rainfalls (> 40 mm) are more probable between May and the beginning of December, with a peak at the very beginning of November. ECMWF 6-h forecasts at 18 gridpoints (spaced at 0.25°) above the FVG region are studied for the same period, to find the maximum 6-h rain forecasted by the ECMWF model from + 6 to + 48 h and correlate it with the observed maximum rain of all the 104 raingauges. It is found that the correlation coefficient R is higher at 0000-0600 UTC and minimum at 1800-0000 UTC, while the BIAS is always negative (underestimation), varying between - 3.5 and - 6.9 mm. Looking at more homogeneous subareas, ECMWF has a much worse BIAS and RMSE for the Prealps zone, while its correlation coefficient is lower for the coastal and plains zones. For comparison, a similar exercise is repeated using a LAM model (ALADIN-ARSO), finding better BIAS and RMSE, but a lower skill for the mean correlation coefficient. Hence, a linear statistical method (multiregression with exhaustive input selection) for forecasting the maximum 6-h rain using as candidate predictors the direct model output (absolute values, anomalies, standardized values, plus mean, max and SD in time and space) is developed independently for four different sub-regions and two periods of the year starting from the ECMWF forecast. It is found that the strong BIAS in the Prealpine area can easily be removed
Probable maximum precipitation 24 hours estimation: A case study of Zanjan province of Iran
Azim Shirdeli
2012-10-01
Full Text Available One of the primary concerns in designing civil structures such as water storage dams and irrigation and drainage networks is to find economic scale based on possibility of natural incidents such as floods, earthquake, etc. Probable maximum precipitation (PMP is one of well known methods, which helps design a civil structure, properly. In this paper, we study the maximum one-day precipitation using 17 to 50 years of information in 13 stations located in province of Zanjan, Iran. The proposed study of this paper uses two Hershfield methods, where the first one yields 18.17 to 18.48 for precipitation where the PMP24 was between 170.14 mm and 255.28 mm. The second method reports precipitation between 2.29 and 4.95 while PMP24 was between 62.33 mm and 92.08 mm. In addition, when the out of range data were deleted from the study of the second method, precipitation rates were calculated between 2.29 and 4.31 while PMP24 was between 76.08 mm and 117.28 mm. The preliminary results indicate that the second Hershfield method provide more stable results than the first one.
Sung Woo Park
2015-03-01
Full Text Available The safety of a multi-span waler beam subjected simultaneously to a distributed load and deflections at its supports can be secured by limiting the maximum stress of the beam to a specific value to prevent the beam from reaching a limit state for failure or collapse. Despite the fact that the vast majority of accidents on construction sites occur at waler beams in retaining wall systems, no safety monitoring model that can consider deflections at the supports of the beam is available. In this paper, a maximum stress estimation model for a waler beam based on average strains measured from vibrating wire strain gauges (VWSGs, the most frequently used sensors in construction field, is presented. The model is derived by defining the relationship between the maximum stress and the average strains measured from VWSGs. In addition to the maximum stress, support reactions, deflections at supports, and the magnitudes of distributed loads for the beam structure can be identified by the estimation model using the average strains. Using simulation tests on two multi-span beams, the performance of the model is evaluated by estimating maximum stress, deflections at supports, support reactions, and the magnitudes of distributed loads.
Cavalli, Andrea; Camilloni, Carlo; Vendruscolo, Michele
2013-03-07
In order to characterise the dynamics of proteins, a well-established method is to incorporate experimental parameters as replica-averaged structural restraints into molecular dynamics simulations. Here, we justify this approach in the case of interproton distance information provided by nuclear Overhauser effects by showing that it generates ensembles of conformations according to the maximum entropy principle. These results indicate that the use of replica-averaged structural restraints in molecular dynamics simulations, given a force field and a set of experimental data, can provide an accurate approximation of the unknown Boltzmann distribution of a system.
U.S. Geological Survey, Department of the Interior — This data set represents the 30-year (1971-2000) average annual maximum temperature in Celsius multiplied by 100 compiled for every catchment of NHDPlus for the...
Sung Woo Park; Byung Kwan Oh; Hyo Seon Park
2015-01-01
The safety of a multi-span waler beam subjected simultaneously to a distributed load and deflections at its supports can be secured by limiting the maximum stress of the beam to a specific value to prevent the beam from reaching a limit state for failure or collapse. Despite the fact that the vast majority of accidents on construction sites occur at waler beams in retaining wall systems, no safety monitoring model that can consider deflections at the supports of the beam is available. In this...
G. M. J. HASAN
2014-10-01
Full Text Available Climate, one of the major controlling factors for well-being of the inhabitants in the world, has been changing in accordance with the natural forcing and manmade activities. Bangladesh, the most densely populated countries in the world is under threat due to climate change caused by excessive use or abuse of ecology and natural resources. This study checks the rainfall patterns and their associated changes in the north-eastern part of Bangladesh mainly Sylhet city through statistical analysis of daily rainfall data during the period of 1957 - 2006. It has been observed that a good correlation exists between the monthly mean and daily maximum rainfall. A linear regression analysis of the data is found to be significant for all the months. Some key statistical parameters like the mean values of Coefficient of Variability (CV, Relative Variability (RV and Percentage Inter-annual Variability (PIV have been studied and found to be at variance. Monthly, yearly and seasonal variation of rainy days also analysed to check for any significant changes.
Favret, Eduardo A; Fuentes, Néstor O; Molina, Ana M; Setten, Lorena M
2008-10-01
During the last few years, RIMAPS technique has been used to characterize the micro-relief of metallic surfaces and recently also applied to biological surfaces. RIMAPS is an image analysis technique which uses the rotation of an image and calculates its average power spectrum. Here, it is presented as a tool for describing the morphology of the trichodium net found in some grasses, which is developed on the epidermal cells of the lemma. Three different species of grasses (herbarium samples) are analyzed: Podagrostis aequivalvis (Trin.) Scribn. & Merr., Bromidium hygrometricum (Nees) Nees & Meyen and Bromidium ramboi (Parodi) Rúgolo. Simple schemes representing the real microstructure of the lemma are proposed and studied. RIMAPS spectra of both the schemes and the real microstructures are compared. These results allow inferring how similar the proposed geometrical schemes are to the real microstructures. Each geometrical pattern could be used as a reference for classifying other species. Finally, this kind of analysis is used to determine the morphology of the trichodium net of Agrostis breviculmis Hitchc. As the dried sample had shrunk and the microstructure was not clear, two kinds of morphology are proposed for the trichodium net of Agrostis L., one elliptical and the other rectilinear, the former being the most suitable.
Mossetti, Stefano; de Bartolo, Daniela; Veronese, Ivan; Cantone, Marie Claire; Cosenza, Cristina; Nava, Elisa
2016-12-01
International and national organizations have formulated guidelines establishing limits for occupational and residential electromagnetic field (EMF) exposure at high-frequency fields. Italian legislation fixed 20 V/m as a limit for public protection from exposure to EMFs in the frequency range 0.1 MHz-3 GHz and 6 V/m as a reference level. Recently, the law was changed and the reference level must now be evaluated as the 24-hour average value, instead of the previous highest 6 minutes in a day. The law refers to a technical guide (CEI 211-7/E published in 2013) for the extrapolation techniques that public authorities have to use when assessing exposure for compliance with limits. In this work, we present measurements carried out with a vectorial spectrum analyzer to identify technical critical aspects in these extrapolation techniques, when applied to UMTS and LTE signals. We focused also on finding a good balance between statistically significant values and logistic managements in control activity, as the signal trend in situ is not known. Measurements were repeated several times over several months and for different mobile companies. The outcome presented in this article allowed us to evaluate the reliability of the extrapolation results obtained and to have a starting point for defining operating procedures.
Aminah, Agustin Siti; Pawitan, Gandhi; Tantular, Bertho
2017-03-01
So far, most of the data published by Statistics Indonesia (BPS) as data providers for national statistics are still limited to the district level. Less sufficient sample size for smaller area levels to make the measurement of poverty indicators with direct estimation produced high standard error. Therefore, the analysis based on it is unreliable. To solve this problem, the estimation method which can provide a better accuracy by combining survey data and other auxiliary data is required. One method often used for the estimation is the Small Area Estimation (SAE). There are many methods used in SAE, one of them is Empirical Best Linear Unbiased Prediction (EBLUP). EBLUP method of maximum likelihood (ML) procedures does not consider the loss of degrees of freedom due to estimating β with β ^. This drawback motivates the use of the restricted maximum likelihood (REML) procedure. This paper proposed EBLUP with REML procedure for estimating poverty indicators by modeling the average of household expenditures per capita and implemented bootstrap procedure to calculate MSE (Mean Square Error) to compare the accuracy EBLUP method with the direct estimation method. Results show that EBLUP method reduced MSE in small area estimation.
Wage and Labor Standards Administration (DOL), Washington, DC.
The Fair Labor Standards Act's 1966 amendments extended coverage to all non-Federal hospitals. Using data on employment, hours, wages, and supplementary benefits from one payroll period in March 1969, this report describes the impact of the increased coverage. Although 19 percent of the nonsupervisory employees were earning less than $1.30 an hour…
DeVore, Matthew S; Gull, Stephen F; Johnson, Carey K
2012-04-05
We describe a method for analysis of single-molecule Förster resonance energy transfer (FRET) burst measurements using classic maximum entropy. Classic maximum entropy determines the Bayesian inference for the joint probability describing the total fluorescence photons and the apparent FRET efficiency. The method was tested with simulated data and then with DNA labeled with fluorescent dyes. The most probable joint distribution can be marginalized to obtain both the overall distribution of fluorescence photons and the apparent FRET efficiency distribution. This method proves to be ideal for determining the distance distribution of FRET-labeled biomolecules, and it successfully predicts the shape of the recovered distributions.
高艳普; 王向东; 王冬青
2015-01-01
An algorithm of maximum likelihood method for parameters estimate was presented aimed at multivariable controlled autoregressive moving average (CARMA-like).The algorithm transform the CARMA-like system into m identification models (m is the output numbers),each of which only had a parameter vector which needed to be esti-mated,and then through maximum likelihood method for estimating parameter vectors of each identification model,and all parameters estimate of the system were obtained.Simulation results verified the effectiveness of the proposed algo-rithm.%提出了一种针对多变量受控自回归滑动平均（controlled autoregressive moving average system-like，CARMA-like）系统的极大似然参数估计算法。将 CARMA-like 系统分解成为 m 个辨识模型（m 是输出量的个数），使每一个辨识模型仅包含一个需要估计的参数向量，通过极大似然方法估计每个辨识模型的参数向量，从而得到整个系统的参数估计值。仿真结果验证了该算法的有效性。
Unsupervised/supervised learning concept for 24-hour load forecasting
Djukanovic, M. (Electrical Engineering Inst. ' Nikola Tesla' , Belgrade (Yugoslavia)); Babic, B. (Electrical Power Industry of Serbia, Belgrade (Yugoslavia)); Sobajic, D.J.; Pao, Y.-H. (Case Western Reserve Univ., Cleveland, OH (United States). Dept. of Electrical Engineering and Computer Science)
1993-07-01
An application of artificial neural networks in short-term load forecasting is described. An algorithm using an unsupervised/supervised learning concept and historical relationship between the load and temperature for a given season, day type and hour of the day to forecast hourly electric load with a lead time of 24 hours is proposed. An additional approach using functional link net, temperature variables, average load and last one-hour load of previous day is introduced and compared with the ANN model with one hidden layer load forecast. In spite of limited available weather variables (maximum, minimum and average temperature for the day) quite acceptable results have been achieved. The 24-hour-ahead forecast errors (absolute average) ranged from 2.78% for Saturdays and 3.12% for working days to 3.54% for Sundays. (Author)
Mary Hokazono
Full Text Available CONTEXT AND OBJECTIVE: Transcranial Doppler (TCD detects stroke risk among children with sickle cell anemia (SCA. Our aim was to evaluate TCD findings in patients with different sickle cell disease (SCD genotypes and correlate the time-averaged maximum mean (TAMM velocity with hematological characteristics. DESIGN AND SETTING: Cross-sectional analytical study in the Pediatric Hematology sector, Universidade Federal de São Paulo. METHODS: 85 SCD patients of both sexes, aged 2-18 years, were evaluated, divided into: group I (62 patients with SCA/Sß0 thalassemia; and group II (23 patients with SC hemoglobinopathy/Sß+ thalassemia. TCD was performed and reviewed by a single investigator using Doppler ultrasonography with a 2 MHz transducer, in accordance with the Stroke Prevention Trial in Sickle Cell Anemia (STOP protocol. The hematological parameters evaluated were: hematocrit, hemoglobin, reticulocytes, leukocytes, platelets and fetal hemoglobin. Univariate analysis was performed and Pearson's coefficient was calculated for hematological parameters and TAMM velocities (P < 0.05. RESULTS: TAMM velocities were 137 ± 28 and 103 ± 19 cm/s in groups I and II, respectively, and correlated negatively with hematocrit and hemoglobin in group I. There was one abnormal result (1.6% and five conditional results (8.1% in group I. All results were normal in group II. Middle cerebral arteries were the only vessels affected. CONCLUSION: There was a low prevalence of abnormal Doppler results in patients with sickle-cell disease. Time-average maximum mean velocity was significantly different between the genotypes and correlated with hematological characteristics.
U.S. Department of Health & Human Services — A list of a variety of averages for each state or territory as well as the national average, including each quality measure, staffing, fine amount and number of...
8 Hour Ozone Design Value for 1998-2000
U.S. Environmental Protection Agency — The Ozone design value is based on the average of the annual 4th highest daily 8-hour maximum over a 3-year period (1998-2000) in this case. This is a human health...
Carré, Françoise; Tilly, Chris
2012-01-01
In settings where most workers have full-time schedules, hourly wages are appropriate primary indicators of job quality and worker outcomes. However, in sectors where full-time schedules do not dominate - primarily service-producing activities - total hours matter, in addition to hourly wages, for job quality and worker outcomes. In this paper we employ a sector-focused, comparative framework to further examine hours levels - measured as average weekly hours - and trends in Canada, the United...
A simple method to downscale daily wind statistics to hourly wind data
Guo, Zhongling
2013-01-01
Wind is the principal driver in the wind erosion models. The hourly wind speed data were generally required for precisely wind erosion modeling. In this study, a simple method to generate hourly wind speed data from daily wind statistics (daily average and maximum wind speeds together or daily average wind speed only) was established. A typical windy location with 3285 days (9 years) measured hourly wind speed data were used to validate the downscaling method. The results showed that the overall agreement between observed and simulated cumulative wind speed probability distributions appears excellent, especially for the wind speeds greater than 5 m s-1 range (erosive wind speed). The results further revealed that the values of daily average erosive wind power density (AWPD) calculated from generated wind speeds fit the counterparts computed from measured wind speeds well with high models' efficiency (Nash-Sutcliffe coefficient). So that the hourly wind speed data can be predicted from daily average and maximu...
29 CFR 778.101 - Maximum nonovertime hours.
2010-07-01
... POLICY OR INTERPRETATION NOT DIRECTLY RELATED TO REGULATIONS OVERTIME COMPENSATION The Overtime Pay... in any workweek without receiving additional compensation at not less than the statutory rate...
Kinkhabwala, Ali
2013-01-01
The most fundamental problem in statistics is the inference of an unknown probability distribution from a finite number of samples. For a specific observed data set, answers to the following questions would be desirable: (1) Estimation: Which candidate distribution provides the best fit to the observed data?, (2) Goodness-of-fit: How concordant is this distribution with the observed data?, and (3) Uncertainty: How concordant are other candidate distributions with the observed data? A simple unified approach for univariate data that addresses these traditionally distinct statistical notions is presented called "maximum fidelity". Maximum fidelity is a strict frequentist approach that is fundamentally based on model concordance with the observed data. The fidelity statistic is a general information measure based on the coordinate-independent cumulative distribution and critical yet previously neglected symmetry considerations. An approximation for the null distribution of the fidelity allows its direct conversi...
[Travel time and distances to Norwegian out-of-hours casualty clinics].
Raknes, Guttorm; Morken, Tone; Hunskår, Steinar
2014-11-01
Geographical factors have an impact on the utilisation of out-of-hours services. In this study we have investigated the travel distance to out-of-hours casualty clinics in Norwegian municipalities in 2011 and the number of municipalities covered by the proposed recommendations for secondary on-call arrangements due to long distances. We estimated the average maximum travel times and distances in Norwegian municipalities using a postcode-based method. Separate analyses were performed for municipalities with a single, permanently located casualty clinic. Altogether 417 out of 430 municipalities were included. We present the median value of the maximum travel times and distances for the included municipalities. The median maximum average travel distance for the municipalities was 19 km. The median maximum average travel time was 22 minutes. In 40 of the municipalities (10 %) the median maximum average travel time exceeded 60 minutes, and in 97 municipalities (23 %) the median maximum average travel time exceeded 40 minutes. The population of these groups comprised 2 % and 5 % of the country's total population respectively. For municipalities with permanent emergency facilities(N = 316), the median average flight time 16 minutes and median average distance 13 km.. In many municipalities, the inhabitants have a long average journey to out-of-hours emergency health services, but seen as a whole, the inhabitants of these municipalities account for a very small proportion of the Norwegian population. The results indicate that the proposed recommendations for secondary on-call duty based on long distances apply to only a small number of inhabitants. The recommendations should therefore be adjusted and reformulated to become more relevant.
Surface Weather Observations Hourly
National Oceanic and Atmospheric Administration, Department of Commerce — Standard hourly observations taken at Weather Bureau/National Weather Service offices and airports throughout the United States. Hourly observations began during the...
Long working hours and cancer risk
Heikkila, Katriina; Nyberg, Solja T.; Madsen, Ida E. H.
2016-01-01
Background: Working longer than the maximum recommended hours is associated with an increased risk of cardiovascular disease, but the relationship of excess working hours with incident cancer is unclear. Methods: This multi-cohort study examined the association between working hours and cancer risk...... in 116 462 men and women who were free of cancer at baseline. Incident cancers were ascertained from national cancer, hospitalisation and death registers; weekly working hours were self-reported. Results: During median follow-up of 10.8 years, 4371 participants developed cancer (n colorectal cancer: 393......; n lung cancer: 247; n breast cancer: 833; and n prostate cancer: 534). We found no clear evidence for an association between working hours and the overall cancer risk. Working hours were also unrelated the risk of incident colorectal, lung or prostate cancers. Working greater than or equal to55 h...
A Century of Human Capital and Hours
Diego Restuccia; Guillaume Vandenbroucke
2012-01-01
An average person born in the United States in the second half of the nineteenth century completed 7 years of schooling and spent 58 hours a week working in the market. By contrast, an average person born at the end of the twentieth century completed 14 years of schooling and spent 40 hours a week working. In the span of 100 years, completed years of schooling doubled and working hours decreased by 30 percent. What explains these trends? We consider a model of human capital and labor supply t...
Intensity contrast of the average supergranule
Langfellner, J; Gizon, L
2016-01-01
While the velocity fluctuations of supergranulation dominate the spectrum of solar convection at the solar surface, very little is known about the fluctuations in other physical quantities like temperature or density at supergranulation scale. Using SDO/HMI observations, we characterize the intensity contrast of solar supergranulation at the solar surface. We identify the positions of ${\\sim}10^4$ outflow and inflow regions at supergranulation scales, from which we construct average flow maps and co-aligned intensity and magnetic field maps. In the average outflow center, the maximum intensity contrast is $(7.8\\pm0.6)\\times10^{-4}$ (there is no corresponding feature in the line-of-sight magnetic field). This corresponds to a temperature perturbation of about $1.1\\pm0.1$ K, in agreement with previous studies. We discover an east-west anisotropy, with a slightly deeper intensity minimum east of the outflow center. The evolution is asymmetric in time: the intensity excess is larger 8 hours before the reference t...
GS Department
2009-01-01
Please note the new opening hours of the gates as well as the intersites tunnel from the 19 May 2009: GATE A 7h - 19h GATE B 24h/24 GATE C 7h - 9h\t17h - 19h GATE D 8h - 12h\t13h - 16h GATE E 7h - 9h\t17h - 19h Prévessin 24h/24 The intersites tunnel will be opened from 7h30 to 18h non stop. GS-SEM Group Infrastructure and General Services Department
Dubrawsky, Ido
2009-01-01
This book will focus on just the essentials needed to pass the Security+ certification exam. It will be filled with critical information in a way that will be easy to remember and use for your quickly approaching exam. It will focus on the main objectives of the exam and include the following pedagogy for ease of use in those final hours. The book will include:. •Exam Objectives – Fast Track Review. •Key words/definitions. •Five Toughest questions and their answers. •Exam Warnings – What to pay attention to
Speake, Graham; Happel, Chris; Lillard, Terrence V
2009-01-01
The 11th Hour Linux+ Study Guide is keyed to the XK0-003 revision of the CompTIA Linux+exam. This book is streamlined to include only core certification information and is presented for ease of last-minute studying. Main objectives of the exam are covered with key concepts highlighted. ..: ..; Fast Facts quickly review fundamentals ..; Exam Warnings highlight particularly tough sections of the exam ..; Crunch Time sidebars point out key concepts to remember ..; Did You Know? sidebars cover sometimes forgotten details ..; Top Five Toughest Questions and answers help you to prepare .. The 11th H
2003-01-01
The 18th edition of the Geneva 24 hours swim competition will take place at the Vernets Swimming Pool on the 4th and 5th of October. More information and the results of previous years are given at: http://www.carouge-natation.com/24_heures/home_24_heures.htm Last year, CERN obtained first position in the inter-company category with a total of 152.3 kms swam by 45 participants. We are counting on your support to repeat this excellent performance this year. For those who would like to train, the Livron swimming pool in Meyrin is open as from Monday the 8th September. For further information please do not hesitate to contact us. Gino de Bilio and Catherine Delamare
2003-01-01
The 18th edition of the Geneva 24 hours swim competition will take place at the Vernets Swimming Pool on the 4th and 5th of October. More information and the results of previous years are given at: http://www.carouge-natation.com/24_heures/home_24_heures.htm Last year, CERN obtained first position in the inter-company category with a total of 152.3 kms swam by 45 participants. We are counting on your support to repeat this excellent performance this year. For those who would like to train, the Livron swimming pool in Meyrin is open as from Monday the 8th September. For further information please do not hesitate to contact us. Gino de Bilio and Catherine Delamare
NingQuanxin; QiuLizhong
2004-01-01
The goal of this study is to improve instructional classes for “Extensive Reading”. We compare two approaches. The traditional method is based on the original procedure; the method we try is a “task-placemented” one. inspired by the method we try is a “task-placemented” one. inspired by the theory of “Tasks of Ascending Difficulty”.This paper provides a comparative evaluation of our experiments on Extensive Reading. Our aim is to resolve the logistical problem of heavy teaching loads and limited class hours, which often prevents English learners from fulfilling their study targets and our intended learning outcomes. In view of the experimental data,we have managed to establish a “feedback” to test the two hypotheses, and demonstrate that the new approach contributes to efficient acquisition of the information in the presented reading materials. Our study concludes with an interpretation of the function and merits of this approach.
Resident Duty Hours: Enhancing Sleep, Supervision, and Safety
Ulmer, Cheryl, Ed.; Wolman, Dianne Miller, Ed.; Johns, Michael M. E., Ed.
2009-01-01
Medical residents in hospitals are often required to be on duty for long hours. In 2003 the organization overseeing graduate medical education adopted common program requirements to restrict resident workweeks, including limits to an average of 80 hours over 4 weeks and the longest consecutive period of work to 30 hours in order to protect…
Lake Basin Fetch and Maximum Length/Width
Minnesota Department of Natural Resources — Linear features representing the Fetch, Maximum Length and Maximum Width of a lake basin. Fetch, maximum length and average width are calcuated from the lake polygon...
Long working hours and alcohol use
Virtanen, Marianna; Jokela, Markus; Nyberg, Solja T
2015-01-01
OBJECTIVE: To quantify the association between long working hours and alcohol use. DESIGN: Systematic review and meta-analysis of published studies and unpublished individual participant data. DATA SOURCES: A systematic search of PubMed and Embase databases in April 2014 for published studies......, supplemented with manual searches. Unpublished individual participant data were obtained from 27 additional studies. REVIEW METHODS: The search strategy was designed to retrieve cross sectional and prospective studies of the association between long working hours and alcohol use. Summary estimates were...... countries. The pooled maximum adjusted odds ratio for the association between long working hours and alcohol use was 1.11 (95% confidence interval 1.05 to 1.18) in the cross sectional analysis of published and unpublished data. Odds ratio of new onset risky alcohol use was 1.12 (1.04 to 1...
Siegel, Irving H.
The arithmetic processes of aggregation and averaging are basic to quantitative investigations of employment, unemployment, and related concepts. In explaining these concepts, this report stresses need for accuracy and consistency in measurements, and describes tools for analyzing alternative measures. (BH)
Gramkow, Claus
1999-01-01
In this article two common approaches to averaging rotations are compared to a more advanced approach based on a Riemannian metric. Very offten the barycenter of the quaternions or matrices that represent the rotations are used as an estimate of the mean. These methods neglect that rotations belong...... approximations to the Riemannian metric, and that the subsequent corrections are inherient in the least squares estimation. Keywords: averaging rotations, Riemannian metric, matrix, quaternion...
Hourly Precipitation Data (HPD) Publication
National Oceanic and Atmospheric Administration, Department of Commerce — Hourly Precipitation Data (HPD) Publication is archived and available from the National Climatic Data Center (NCDC). This publication contains hourly precipitation...
US Naval Observatory Hourly Observations
National Oceanic and Atmospheric Administration, Department of Commerce — Hourly observations journal from the National Observatory in Washington DC. The observatory is the first station in the United States to produce hourly observations...
Automatization of physicians' phone-in hours.
Turunen, J; Loula, P; Soininen, H
1999-01-01
In this paper we present a computer-based system that is designed to reduce the number of patient calls in physicians' phone-in hours. Simultaneously it will offer a parallel alternative channel for the patients to listen to their laboratory results. The system requirements were reliability, security and the low costs. The results showed after a two-month test period that the tailored system proved extremely successful. The average time that the physician saved by using this system was over 4 minutes and doctors can leave the messages to the server also outside the phone-in hours.
Maximum Autocorrelation Factorial Kriging
Nielsen, Allan Aasbjerg; Conradsen, Knut; Pedersen, John L.
2000-01-01
This paper describes maximum autocorrelation factor (MAF) analysis, maximum autocorrelation factorial kriging, and its application to irregularly sampled stream sediment geochemical data from South Greenland. Kriged MAF images are compared with kriged images of varimax rotated factors from...
Interns shall not sleep: the duty hours boomerang
Quan SF
2017-04-01
Full Text Available No abstract available. Article truncated after 150 words. On March 10, 2017, the Accreditation Council for Graduate Medical Education (ACGME announced revisions to its common program requirements related to duty hours (1. Effective on July 1, 2017, the most important change will be an increase in the maximum consecutive hours that an intern may work. Interns will now be able to continuously perform patient care work up to a maximum of 24 hours with an additional 4 hours for managing care transitions. This reverses the controversial reduction to 16 hours that occurred in 2011 (2. The regulation of house staff duty hours formally began in the late 1980s. It was precipitated largely because of the publicity resulting from the 1984 death of Libby Zion in a New York teaching hospital that was attributed partly to poor decisions made by fatigued and overworked house staff (3. Consequently, the state of New York in 1989 passed laws restricting the …
Young, Vershawn Ashanti
2004-01-01
"Your Average Nigga" contends that just as exaggerating the differences between black and white language leaves some black speakers, especially those from the ghetto, at an impasse, so exaggerating and reifying the differences between the races leaves blacks in the impossible position of either having to try to be white or forever struggling to…
Gramkow, Claus
2001-01-01
In this paper two common approaches to averaging rotations are compared to a more advanced approach based on a Riemannian metric. Very often the barycenter of the quaternions or matrices that represent the rotations are used as an estimate of the mean. These methods neglect that rotations belong...
Covariant approximation averaging
Shintani, Eigo; Blum, Thomas; Izubuchi, Taku; Jung, Chulwoo; Lehner, Christoph
2014-01-01
We present a new class of statistical error reduction techniques for Monte-Carlo simulations. Using covariant symmetries, we show that correlation functions can be constructed from inexpensive approximations without introducing any systematic bias in the final result. We introduce a new class of covariant approximation averaging techniques, known as all-mode averaging (AMA), in which the approximation takes account of contributions of all eigenmodes through the inverse of the Dirac operator computed from the conjugate gradient method with a relaxed stopping condition. In this paper we compare the performance and computational cost of our new method with traditional methods using correlation functions and masses of the pion, nucleon, and vector meson in $N_f=2+1$ lattice QCD using domain-wall fermions. This comparison indicates that AMA significantly reduces statistical errors in Monte-Carlo calculations over conventional methods for the same cost.
The maximum rotation of a galactic disc
Bottema, R
1997-01-01
The observed stellar velocity dispersions of galactic discs show that the maximum rotation of a disc is on average 63% of the observed maximum rotation. This criterion can, however, not be applied to small or low surface brightness (LSB) galaxies because such systems show, in general, a continuously
Hour-glass ceilings: Work-hour thresholds, gendered health inequities.
Dinh, Huong; Strazdins, Lyndall; Welsh, Jennifer
2017-03-01
Long workhours erode health, which the setting of maximum weekly hours aims to avert. This 48-h limit, and the evidence base to support it, has evolved from a workforce that was largely male, whose time in the labour force was enabled by women's domestic work and care giving. The gender composition of the workforce has now changed, and many women (as well as some men) combine care-giving with paid work, a change viewed as fundamental for gender equality. However, it raises questions on the suitability of the work time limit and the extent it is protective of health. We estimate workhour-mental health thresholds, testing if they vary for men and women due to gendered workloads and constraints on and off the job. Using six waves of data from a nationally representative sample of Australian adults (24-65 years), surveyed in the Household Income Labour Dynamics of Australia Survey (N = 3828 men; 4062 women), our study uses a longitudinal, simultaneous equation approach to address endogeneity. Averaging over the sample, we find an overall threshold of 39 h per week beyond which mental health declines. Separate curves then estimate thresholds for men and women, by high or low care and domestic time constraints, using stratified and pooled samples. We find gendered workhour-health limits (43.5 for men, 38 for women) which widen further once differences in resources on and off the job are considered. Only when time is 'unencumbered' and similar time constraints and contexts are assumed, do gender gaps narrow and thresholds approximate the 48-h limit. Our study reveals limits to contemporary workhour regulation which may be systematically disadvantaging women's health.
U.S. Hourly Climate Normals (1981-2010)
National Oceanic and Atmospheric Administration, Department of Commerce — The U.S. Hourly Climate Normals for 1981 to 2010 are 30-year averages of meteorological parameters for thousands of U.S. stations located across the 50 states, as...
Increased instruction hours and the widening gap in student performance
Huebener, Mathias; Kuger, Susanne; Marcus, Jan
2016-01-01
Do increased instruction hours improve the performance of all students? Using PISA scores of students in ninth grade, we analyse the effect of a German education reform that increased weekly instruction hours by two hours (6.5 percent) overalmost five years. In the additional time, students are taught new learning content. On average, the reform improves student performance. However, treatment effects are small and differ across the student performance distribution. While low-performing stude...
Long working hours and alcohol use
Virtanen, Marianna; Jokela, Markus; Nyberg, Solja T
2015-01-01
, supplemented with manual searches. Unpublished individual participant data were obtained from 27 additional studies. REVIEW METHODS: The search strategy was designed to retrieve cross sectional and prospective studies of the association between long working hours and alcohol use. Summary estimates were...... obtained with random effects meta-analysis. Sources of heterogeneity were examined with meta-regression. RESULTS: Cross sectional analysis was based on 61 studies representing 333,693 participants from 14 countries. Prospective analysis was based on 20 studies representing 100,602 participants from nine...... countries. The pooled maximum adjusted odds ratio for the association between long working hours and alcohol use was 1.11 (95% confidence interval 1.05 to 1.18) in the cross sectional analysis of published and unpublished data. Odds ratio of new onset risky alcohol use was 1.12 (1.04 to 1...
Negative Average Preference Utilitarianism
Roger Chao
2012-03-01
Full Text Available For many philosophers working in the area of Population Ethics, it seems that either they have to confront the Repugnant Conclusion (where they are forced to the conclusion of creating massive amounts of lives barely worth living, or they have to confront the Non-Identity Problem (where no one is seemingly harmed as their existence is dependent on the “harmful” event that took place. To them it seems there is no escape, they either have to face one problem or the other. However, there is a way around this, allowing us to escape the Repugnant Conclusion, by using what I will call Negative Average Preference Utilitarianism (NAPU – which though similar to anti-frustrationism, has some important differences in practice. Current “positive” forms of utilitarianism have struggled to deal with the Repugnant Conclusion, as their theory actually entails this conclusion; however, it seems that a form of Negative Average Preference Utilitarianism (NAPU easily escapes this dilemma (it never even arises within it.
Constructing gridded hourly temperature data sets
Dumitrescu, Alexandru
2013-04-01
Air temperature is the main climatological element, with major impact on the earth-atmosphere energy balance. The characteristics of the surface air temperature in locations without meteorological measurements are obtained using spatio-temporal interpolation techniques. Gridded surface meteorological data are essential for evaluating the performance of regional climate models (RCMs), for applying Statistical Downscaling (SD) methods and as input data for hydrological models. In this study we proposed a methodolgy for interpolating hourly surface temperatures. Three gridding methods are compared. A two-step multivariate gridding approach was used. First we interpolated the hourly normal maps, considered as multiannual average (1961-2010), of air temperature for each hour (4 meteorological terms) of a standard year (366 days). In this step, the Residual Kriging method was used with potential predictors derived from DEM and Landcover Corinne. For interpolating the residuals of the regression model we tested 3 gridding methods: Multiquadratic (MQ), Ordinary Kriging (OK) and 3D Kriging (using time as a third dimension). In the second step, we calculated the anomalies of each hour, day, year for the period 1961-2010. The anomalies were interpolated using the same methods applied for gridding regression residuals. The final hourly surface air temperature maps were obtained by summing the maps from first step with the anomlies map. The main data used in this work were the hourly air temperatures of the 4 observation terms (01, 07, 13, 19), measured between 1961-2010 at the weather stations of the Romanian Meteorological Administration. The predictors were derived from SRTM (Shuttle Radar Topography Mission) DEM and from CORINE Land Cover 2000 product. The gridding was performed in a Romanian National Grid (Stereo 70), at 1 km2 spatial resolution, using R language. The study has been financed by the research project Changes in climate extremes and associated impact in
Ensemble Averaged Gravity Theory
Khosravi, Nima
2016-01-01
We put forward the idea that all the theoretically consistent models of gravity have a contribution to the observed gravity interaction. In this formulation each model comes with its own Euclidean path integral weight where general relativity (GR) automatically has the maximum weight in high-curvature regions. We employ this idea in the framework of Lovelock models and show that in four dimensions the result is a specific form of $f(R,G)$ model. This specific $f(R,G)$ satisfies the stability conditions and has self-accelerating solution. Our model is consistent with the local tests of gravity since its behavior is same as GR for high-curvature regimes. In low-curvature regime the gravity force is weaker than GR which can interpret as existence of a repulsive fifth force for very large scales. Interestingly there is an intermediate-curvature regime where the gravity force is stronger in our model than GR. The different behavior of our model in comparison with GR in both low- and intermediate-curvature regimes ...
Independence, Odd Girth, and Average Degree
Löwenstein, Christian; Pedersen, Anders Sune; Rautenbach, Dieter;
2011-01-01
We prove several tight lower bounds in terms of the order and the average degree for the independence number of graphs that are connected and/or satisfy some odd girth condition. Our main result is the extension of a lower bound for the independence number of triangle-free graphs of maximum...
Verification of hourly forecasts of wind turbine power output
Wegley, H.L.
1984-08-01
A verification of hourly average wind speed forecasts in terms of hourly average power output of a MOD-2 was performed for four sites. Site-specific probabilistic transformation models were developed to transform the forecast and observed hourly average speeds to the percent probability of exceedance of an hourly average power output. (This transformation model also appears to have value in predicting annual energy production for use in wind energy feasibility studies.) The transformed forecasts were verified in a deterministic sense (i.e., as continuous values) and in a probabilistic sense (based upon the probability of power output falling in a specified category). Since the smoothing effects of time averaging are very pronounced, the 90% probability of exceedance was built into the transformation models. Semiobjective and objective (model output statistics) forecasts were made compared for the four sites. The verification results indicate that the correct category can be forecast an average of 75% of the time over a 24-hour period. Accuracy generally decreases with projection time out to approx. 18 hours and then may increase due to the fairly regular diurnal wind patterns that occur at many sites. The ability to forecast the correct power output category increases with increasing power output because occurrences of high hourly average power output (near rated) are relatively rare and are generally not forecast. The semiobjective forecasts proved superior to model output statistics in forecasting high values of power output and in the shorter time frames (1 to 6 hours). However, model output statistics were slightly more accurate at other power output levels and times. Noticeable differences were observed between deterministic and probabilistic (categorical) forecast verification results.
Ozone Nonattainment Areas - 1 Hour
U.S. Environmental Protection Agency — This data layer identifies areas in the U.S. where air pollution levels have not met the National Ambient Air Quality Standards (NAAQS) for Ozone - 1hour (Legacy...
Maximum Autocorrelation Factorial Kriging
Nielsen, Allan Aasbjerg; Conradsen, Knut; Pedersen, John L.; Steenfelt, Agnete
2000-01-01
This paper describes maximum autocorrelation factor (MAF) analysis, maximum autocorrelation factorial kriging, and its application to irregularly sampled stream sediment geochemical data from South Greenland. Kriged MAF images are compared with kriged images of varimax rotated factors from an ordinary non-spatial factor analysis, and they are interpreted in a geological context. It is demonstrated that MAF analysis contrary to ordinary non-spatial factor analysis gives an objective discrimina...
Remarks on the Lower Bounds for the Average Genus
Yi-chao Chen
2011-01-01
Let G be a graph of maximum degree at most four. By using the overlap matrix method which is introduced by B. Mohar, we show that the average genus of G is not less than 1/3 of its maximum genus, and the bound is best possible. Also, a new lower bound of average genus in terms of girth is derived.
SORCE Level 3 Total Solar Irradiance 6-Hour Average V016
National Aeronautics and Space Administration — The Total Solar Irradiance (TSI) data set SOR3TSI6 contains the total solar irradiance (a.k.a solar constant) data collected by the Total Irradiance Monitor (TIM)...
A novel diversion protocol dramatically reduces diversion hours.
Asamoah, Osei Kwame; Weiss, Steven J; Ernst, Amy A; Richards, Michael; Sklar, David P
2008-07-01
Ambulance diversion is a problem in many communities. When patients are diverted prompt and appropriate medical care may be delayed. Compare diversion hours and drop-off times before and after a dramatic change in diversion policy restricting each hospital to 1 hour out of every 8. This study was a retrospective study in a county of 600,000 people and 10 hospitals from September 2004 to February 2006. A countywide diversion protocol was implemented in March 2005 that limited diversion hours to 1 hour out of every 8 (maximum of 90 h/mo). No other changes were implemented during the study period. Pretrial (9/04-2/05), interim (3/05-8/05), and posttrial (9/05-2/06) periods were compared. The main outcome measures were ambulance diversion hours and emergency medical service (EMS) drop-off times. Results were compared using analysis of variance and a Tukey post hoc analysis. P ambulance diversion hours (difference, 251 hours; 95% CI, 136-368) and significant increase in additional time that EMS crews required to transport patients (drop-off times) (difference, 178 hours; 95% CI, 74-283) were observed. Posttrial diversion hours decreased to 18% of the pretrial values (from 305 to 54). This novel ambulance diversion protocol dramatically reduced diversion hours at the cost of increasing EMS drop-off times in a large community.
Use of hourly sequences of global radiation for simulation of photovoltaic systems grid connected
Sidrach de Cardona, M. [Dept. Fisica Aplicada 2, E.T.S.I., Univ. de Malaga (Spain); Mora Lopez, Ll. [Dept. Lenguajes y Ciencias de la Computacion, E.T.S.I., Univ. de Malaga (Spain)
1996-12-31
In this paper we analyse whether it is suitable to use hourly sequences of global radiation to study the performance of photovoltaic systems grid connected. The purpose of this paper is to quantify the error which is made on predicting the energy which a photovoltaic system connected to the grid, when the global radiation interval considered (which is an input value in our computations) varies. Using a simulation model we have studied how much energy is produced. The input values we considered are: values of daily global radiation, values of hourly global radiation and global radiation intervals of 10 minutes. These data have been recorded by the metereological station in Malaga of Andalusian Institute for renewable energies. In each interval, the simulation model takes into account the variations in module temperature as a function of ambient temperature. Therefore, when using daily values the model takes into account the average value of ambient temperature during sun hours and when using other values the model takes into account the average of this parameter in the interval under consideration. The nominal power considered is 2 kWp. The model we use assumes that the working point of the installation is always the maximum power point of modules and that all produced energy is injected in the grid. We also study the influence on this error when different set point of minimum input power for inverter starting (threshold value) are used. (orig.)
Estimation of hourly clear-sky solar radiation for P. D. R. Yemen
Abdul Aziz, G.M. (Aden Univ., (P.D.R. Yemen) Faculty of Engineering)
1990-01-01
Monthly average hourly clear sky solar radiation for some principal stations in P.D.R. Yemen are evaluated. Hottel's model is used to predict the hourly clear sky beam and horizontal solar radiation and that of Liu and Jordan for the estimation of clear sky diffuse radiation. The two methods are combined to obtain the clear sky total radiation. Maximum and minimum clear sky total solar radiation are evaluated. The maxima are situated between 21.0 MJ/m{sup 2} and 26.6 MJ/m{sup 2}, minima between 15.0 MJ/m{sup 2} and 19.1 MJ/m{sup 2} for Aden and Mukairas respectively. Mean hourly clear-sky solar radiation are computed and discussed. 75% of the estimated radiation on a horizontal surface is obtained between 9 a.m.-3 p.m. The average intensity is about 700 W.m{sup 2} for Aden and 900 W/m{sup 2} for Mukairas. (author).
Urtis, Tom
2015-01-01
Master VBA automation quickly and easily to get more out of Excel Excel VBA 24-Hour Trainer, 2nd Edition is the quick-start guide to getting more out of Excel, using Visual Basic for Applications. This unique book/video package has been updated with fifteen new advanced video lessons, providing a total of eleven hours of video training and 45 total lessons to teach you the basics and beyond. This self-paced tutorial explains Excel VBA from the ground up, demonstrating with each advancing lesson how you can increase your productivity. Clear, concise, step-by-step instructions are combined wit
Maximum likely scale estimation
Loog, Marco; Pedersen, Kim Steenstrup; Markussen, Bo
2005-01-01
A maximum likelihood local scale estimation principle is presented. An actual implementation of the estimation principle uses second order moments of multiple measurements at a fixed location in the image. These measurements consist of Gaussian derivatives possibly taken at several scales and/or ...
Resolving issues concerning Eskdalemuir geomagnetic hourly values
S. Macmillan
2011-02-01
Full Text Available The hourly values of the geomagnetic field from 1911 to 1931 derived from measurements made at Eskdalemuir observatory in the UK, and available online from the World Data Centre for Geomagnetism at http://www.wdc.bgs.ac.uk/, have now been corrected. Previously they were 2-point averaged and transformed from the original north, east and vertical down values in the tables in the observatory yearbooks. This paper documents the course of events from discovering the post-processing done to the data to the final resolution of the problem. As it was through the development of a new index, the Inter-Hour Variability index, that this post-processing came to light, we provide a revised series of this index for Eskdalemuir and compare it with that from another European observatory. Conclusions of studies concerning long-term magnetic field variability and inferred solar variability, whilst not necessarily consistent with one another, are not obviously invalidated by the incorrect hourly values from Eskdalemuir. This series of events illustrates the challenges that lie ahead in removing any remaining errors and inconsistencies in the data holdings of different World Data Centres.
Maximum information photoelectron metrology
Hockett, P; Wollenhaupt, M; Baumert, T
2015-01-01
Photoelectron interferograms, manifested in photoelectron angular distributions (PADs), are a high-information, coherent observable. In order to obtain the maximum information from angle-resolved photoionization experiments it is desirable to record the full, 3D, photoelectron momentum distribution. Here we apply tomographic reconstruction techniques to obtain such 3D distributions from multiphoton ionization of potassium atoms, and fully analyse the energy and angular content of the 3D data. The PADs obtained as a function of energy indicate good agreement with previous 2D data and detailed analysis [Hockett et. al., Phys. Rev. Lett. 112, 223001 (2014)] over the main spectral features, but also indicate unexpected symmetry-breaking in certain regions of momentum space, thus revealing additional continuum interferences which cannot otherwise be observed. These observations reflect the presence of additional ionization pathways and, most generally, illustrate the power of maximum information measurements of th...
Hourly temporal distribution of wind
Deligiannis, Ilias; Dimitriadis, Panayiotis; Koutsoyiannis, Demetris
2016-04-01
The wind process is essential for hydrometeorology and additionally, is one of the basic renewable energy resources. Most stochastic forecast models are limited up to daily scales disregarding the hourly scale which is significant for renewable energy management. Here, we analyze hourly wind timeseries giving emphasis on the temporal distribution of wind within the day. We finally present a periodic model based on statistical as well as hydrometeorological reasoning that shows good agreement with data. Acknowledgement: This research is conducted within the frame of the undergraduate course "Stochastic Methods in Water Resources" of the National Technical University of Athens (NTUA). The School of Civil Engineering of NTUA provided moral support for the participation of the students in the Assembly.
OPENING HOURS FOR CARDS OFFICE
Human Resources Division
2001-01-01
Due to the extra workload generated by the global renewal of French cards and in order to preserve the level of service offered by the cards office, please note that this office will in future be open every morning from 8.30 a.m. to 12.30 p.m. until further notice. The service can be contacted by telephone during the same hours. Thank you for your understanding.
Maximum Likelihood Associative Memories
Gripon, Vincent; Rabbat, Michael
2013-01-01
Associative memories are structures that store data in such a way that it can later be retrieved given only a part of its content -- a sort-of error/erasure-resilience property. They are used in applications ranging from caches and memory management in CPUs to database engines. In this work we study associative memories built on the maximum likelihood principle. We derive minimum residual error rates when the data stored comes from a uniform binary source. Second, we determine the minimum amo...
Maximum likely scale estimation
Loog, Marco; Pedersen, Kim Steenstrup; Markussen, Bo
2005-01-01
A maximum likelihood local scale estimation principle is presented. An actual implementation of the estimation principle uses second order moments of multiple measurements at a fixed location in the image. These measurements consist of Gaussian derivatives possibly taken at several scales and....../or having different derivative orders. Although the principle is applicable to a wide variety of image models, the main focus here is on the Brownian model and its use for scale selection in natural images. Furthermore, in the examples provided, the simplifying assumption is made that the behavior...... of the measurements is completely characterized by all moments up to second order....
Maximizing utilization of sport halls during peak hours
Iversen, Evald Bundgård; Forsberg, Peter
be increased during peak hours. DATA AND METHODOLOGYData is collected by observation of activities during two weeks on for example whether halls are used or not; the amount of playing field used; and number of participants (Iversen, 2012). Data on 1.331 activities in 36 sport halls across 4 municipalities have...... been collected. RESULTS The number of participants per activity is higher during peak hours, which is expected when demand is high. However, the usage of sport floor only differs slightly between peak and low hours. Both during peak and low hours on average 80-100 per cent of floor space is used......BACKGROUNDDuring peak hours (4.30pm-8pm) demand for timeslots in sport halls in Denmark are high and there are few timeslots available. Further, focus on how public resources are spent most efficient is increasing (Iversen, 2013). This makes it interesting to analyse how utilization could...
ACGME proposes dropping the 16 hour resident shift limit
Robbins RA
2016-11-01
Full Text Available No abstract available. Article truncated after 150 words. The Accreditation Council for Graduate Medical Education (ACGME is proposing that first-year residents would no longer be limited to 16-hour shifts during the 2017-2018 academic year under a controversial proposal released today (1. Instead, individual residency programs could assign first-year trainees to shifts as long as 28 hours, the current limit for all other residents. The 28-hour maximum includes 4 transitional hours that's designed in part to help residents improve continuity of care. The plan to revise training requirements does not change other rules designed to protect all residents from overwork. including the maximum80 hours per week. The ACGME capped the shifts of first-year residents at 16 hours in 2011 as a part of an ongoing effort to make trainee schedules more humane and avoid clinical errors caused by sleep deprivation. ACGME CEO Thomas Nasca, MD, told Medscape Medical News that the problem arises largely from first-year residents not being ...
Physical Theories with Average Symmetry
Alamino, Roberto C.
2013-01-01
This Letter probes the existence of physical laws invariant only in average when subjected to some transformation. The concept of a symmetry transformation is broadened to include corruption by random noise and average symmetry is introduced by considering functions which are invariant only in average under these transformations. It is then shown that actions with average symmetry obey a modified version of Noether's Theorem with dissipative currents. The relation of this with possible violat...
49 CFR 174.86 - Maximum allowable operating speed.
2010-10-01
... 49 Transportation 2 2010-10-01 2010-10-01 false Maximum allowable operating speed. 174.86 Section... operating speed. (a) For molten metals and molten glass shipped in packagings other than those prescribed in § 173.247 of this subchapter, the maximum allowable operating speed may not exceed 24 km/hour (15...
Average Convexity in Communication Situations
Slikker, M.
1998-01-01
In this paper we study inheritance properties of average convexity in communication situations. We show that the underlying graph ensures that the graphrestricted game originating from an average convex game is average convex if and only if every subgraph associated with a component of the underlyin
Palesa; Temaswati; Mthethwa
2011-01-01
VEGETARIAN for a week! That was the plan.But it’s funny how side effects start kicking in so early when trying to let go of an everyday habit.The first morning of my experiment,I felt like I hadn’t eaten poultry in months(I had devoured a bucket of KFC the night before).Lunch was a disaster.I didn’t know whether to cook or spend an hour searching for a restaurant. That afternoon I received the worst phone call a first-time vegetarian can get:
Lowery, Joseph W
2011-01-01
Comprehensive written and interactive instruction for learning HTML5 HTML is the core technology for building websites. Today, with HTML5 opening the Internet to new levels of rich content and dynamic interactivity, developers are looking for information to learn and utilize HTML5. HTML5 24-Hour Trainer provides that information, giving new and aspiring web developers the knowledge they need to achieve early success when building websites. Covers the most basic aspects of a web page, including a brief introduction to Cascading Style Sheets (CSS) Provides lessons that are backed up by prof
Java programming 24-hour trainer
Fain, Yakov
2015-01-01
Quick and painless Java programming with expert multimedia instruction Java Programming 24-Hour Trainer, 2nd Edition is your complete beginner's guide to the Java programming language, with easy-to-follow lessons and supplemental exercises that help you get up and running quickly. Step-by-step instruction walks you through the basics of object-oriented programming, syntax, interfaces, and more, before building upon your skills to develop games, web apps, networks, and automations. This second edition has been updated to align with Java SE 8 and Java EE 7, and includes new information on GUI b
F. TopsÃƒÂ¸e
2001-09-01
Full Text Available Abstract: In its modern formulation, the Maximum Entropy Principle was promoted by E.T. Jaynes, starting in the mid-fifties. The principle dictates that one should look for a distribution, consistent with available information, which maximizes the entropy. However, this principle focuses only on distributions and it appears advantageous to bring information theoretical thinking more prominently into play by also focusing on the "observer" and on coding. This view was brought forward by the second named author in the late seventies and is the view we will follow-up on here. It leads to the consideration of a certain game, the Code Length Game and, via standard game theoretical thinking, to a principle of Game Theoretical Equilibrium. This principle is more basic than the Maximum Entropy Principle in the sense that the search for one type of optimal strategies in the Code Length Game translates directly into the search for distributions with maximum entropy. In the present paper we offer a self-contained and comprehensive treatment of fundamentals of both principles mentioned, based on a study of the Code Length Game. Though new concepts and results are presented, the reading should be instructional and accessible to a rather wide audience, at least if certain mathematical details are left aside at a rst reading. The most frequently studied instance of entropy maximization pertains to the Mean Energy Model which involves a moment constraint related to a given function, here taken to represent "energy". This type of application is very well known from the literature with hundreds of applications pertaining to several different elds and will also here serve as important illustration of the theory. But our approach reaches further, especially regarding the study of continuity properties of the entropy function, and this leads to new results which allow a discussion of models with so-called entropy loss. These results have tempted us to speculate over
Regularized maximum correntropy machine
Wang, Jim Jing-Yan
2015-02-12
In this paper we investigate the usage of regularized correntropy framework for learning of classifiers from noisy labels. The class label predictors learned by minimizing transitional loss functions are sensitive to the noisy and outlying labels of training samples, because the transitional loss functions are equally applied to all the samples. To solve this problem, we propose to learn the class label predictors by maximizing the correntropy between the predicted labels and the true labels of the training samples, under the regularized Maximum Correntropy Criteria (MCC) framework. Moreover, we regularize the predictor parameter to control the complexity of the predictor. The learning problem is formulated by an objective function considering the parameter regularization and MCC simultaneously. By optimizing the objective function alternately, we develop a novel predictor learning algorithm. The experiments on two challenging pattern classification tasks show that it significantly outperforms the machines with transitional loss functions.
Sampling Based Average Classifier Fusion
Jian Hou
2014-01-01
fusion algorithms have been proposed in literature, average fusion is almost always selected as the baseline for comparison. Little is done on exploring the potential of average fusion and proposing a better baseline. In this paper we empirically investigate the behavior of soft labels and classifiers in average fusion. As a result, we find that; by proper sampling of soft labels and classifiers, the average fusion performance can be evidently improved. This result presents sampling based average fusion as a better baseline; that is, a newly proposed classifier fusion algorithm should at least perform better than this baseline in order to demonstrate its effectiveness.
Jessica Ackerman
2013-07-01
Full Text Available 7 positions in 2 hours (2013 is a drawing that documents the process of making the short film Role Reversal Rehearsal. It became quickly apparent that the process of making the work was more dynamic and interesting than the finished piece itself. Relationships between the childcare arrangements of the participants and the collective working process brought about the necessity of collaboration for parent artists. Each participant gave their time, energy and creative insight towards filming a series of birthing positions with roles reversed. The male performer became the central figure in an attempt to prompt empathy, humour, and to embody the importance of the male role in childbirth. There were two hours to choreograph, rehearse, and film the sequence. The drawing by Ackerman encapsulates the 'rhizomatic' approach to producing creative work under the constraints of parenthood. The 'arborescent' structure of hierarchy encouraged in industrial filmmaking is subsumed in favour of a horizontal structure. This new structure allows for the creative input, and flow of collaboration between all people involved - including the 3 and 5 year olds, who contributed ideas for camera and soundtrack in situ.
Equalized near maximum likelihood detector
2012-01-01
This paper presents new detector that is used to mitigate intersymbol interference introduced by bandlimited channels. This detector is named equalized near maximum likelihood detector which combines nonlinear equalizer and near maximum likelihood detector. Simulation results show that the performance of equalized near maximum likelihood detector is better than the performance of nonlinear equalizer but worse than near maximum likelihood detector.
Cheeseman, Peter; Stutz, John
2005-01-01
A long standing mystery in using Maximum Entropy (MaxEnt) is how to deal with constraints whose values are uncertain. This situation arises when constraint values are estimated from data, because of finite sample sizes. One approach to this problem, advocated by E.T. Jaynes [1], is to ignore this uncertainty, and treat the empirically observed values as exact. We refer to this as the classic MaxEnt approach. Classic MaxEnt gives point probabilities (subject to the given constraints), rather than probability densities. We develop an alternative approach that assumes that the uncertain constraint values are represented by a probability density {e.g: a Gaussian), and this uncertainty yields a MaxEnt posterior probability density. That is, the classic MaxEnt point probabilities are regarded as a multidimensional function of the given constraint values, and uncertainty on these values is transmitted through the MaxEnt function to give uncertainty over the MaXEnt probabilities. We illustrate this approach by explicitly calculating the generalized MaxEnt density for a simple but common case, then show how this can be extended numerically to the general case. This paper expands the generalized MaxEnt concept introduced in a previous paper [3].
Physical Theories with Average Symmetry
Alamino, Roberto C
2013-01-01
This Letter probes the existence of physical laws invariant only in average when subjected to some transformation. The concept of a symmetry transformation is broadened to include corruption by random noise and average symmetry is introduced by considering functions which are invariant only in average under these transformations. It is then shown that actions with average symmetry obey a modified version of Noether's Theorem with dissipative currents. The relation of this with possible violations of physical symmetries, as for instance Lorentz invariance in some quantum gravity theories, is briefly commented.
Multiplicative ARMA models to generate hourly series of global irradiation
Mora-Lopez, L. [Universidad de Malaga (Spain). Dpto. Lenguajes y C. Computacion; Sidrach-de-Cardona, M. [Universidad de Malaga (Spain). Dpto. Fisica Aplicada
1998-11-01
A methodology to generate hourly series of global irradiation is proposed. The only input parameter which is required is the monthly mean value of daily global irradiation, which is available for most locations. The procedure to obtain new series is based on the use of a multiplicative autoregressive moving-average statistical model for time series with regular and seasonal components. The multiplicative nature of these models enables capture of the two types of relationships observed in recorded hourly series of global irradiation: on the one hand, the relationship between the value at one hour and the value at the previous hour; and on the other hand, the relationship between the value at one hour in one day and the value at the same hour in the previous day. In this paper the main drawback which arises when using these models to generate new series is solved: namely, the need for available recorded series in order to obtain the three parameters contained in the statistical ARMA model which is proposed (autoregressive coefficient, moving-average coefficient and variance of the error term). Specifically, expressions which enable estimation of these parameters using only monthly mean values of daily global irradiation are proposed in this paper. (author)
Maximum phonation time: variability and reliability.
Speyer, Renée; Bogaardt, Hans C A; Passos, Valéria Lima; Roodenburg, Nel P H D; Zumach, Anne; Heijnen, Mariëlle A M; Baijens, Laura W J; Fleskens, Stijn J H M; Brunings, Jan W
2010-05-01
The objective of the study was to determine maximum phonation time reliability as a function of the number of trials, days, and raters in dysphonic and control subjects. Two groups of adult subjects participated in this reliability study: a group of outpatients with functional or organic dysphonia versus a group of healthy control subjects matched by age and gender. Over a period of maximally 6 weeks, three video recordings were made of five subjects' maximum phonation time trials. A panel of five experts were responsible for all measurements, including a repeated measurement of the subjects' first recordings. Patients showed significantly shorter maximum phonation times compared with healthy controls (on average, 6.6 seconds shorter). The averaged interclass correlation coefficient (ICC) over all raters per trial for the first day was 0.998. The averaged reliability coefficient per rater and per trial for repeated measurements of the first day's data was 0.997, indicating high intrarater reliability. The mean reliability coefficient per day for one trial was 0.939. When using five trials, the reliability increased to 0.987. The reliability over five trials for a single day was 0.836; for 2 days, 0.911; and for 3 days, 0.935. To conclude, the maximum phonation time has proven to be a highly reliable measure in voice assessment. A single rater is sufficient to provide highly reliable measurements.
1990-11-01
findings contained in this report are thosE Df the author(s) and should not he construed as an official Department Df the Army position, policy , or...Marquardt methods" to perform linear and nonlinear estimations. One idea in this area by Box and Jenkins (1976) was the " backcasting " procedure to evaluate
Tehsin, Sara; Rehman, Saad; Awan, Ahmad B.; Chaudry, Qaiser; Abbas, Muhammad; Young, Rupert; Asif, Afia
2016-04-01
Sensitivity to the variations in the reference image is a major concern when recognizing target objects. A combinational framework of correlation filters and logarithmic transformation has been previously reported to resolve this issue alongside catering for scale and rotation changes of the object in the presence of distortion and noise. In this paper, we have extended the work to include the influence of different logarithmic bases on the resultant correlation plane. The meaningful changes in correlation parameters along with contraction/expansion in the correlation plane peak have been identified under different scenarios. Based on our research, we propose some specific log bases to be used in logarithmically transformed correlation filters for achieving suitable tolerance to different variations. The study is based upon testing a range of logarithmic bases for different situations and finding an optimal logarithmic base for each particular set of distortions. Our results show improved correlation and target detection accuracies.
Morales-Casique, E.; Neuman, S.P.; Vesselinov, V.V.
2010-01-01
We use log permeability and porosity data obtained from single-hole pneumatic packer tests in six boreholes drilled into unsaturated fractured tuff near Superior, Arizona, to postulate, calibrate and compare five alternative variogram models (exponential, exponential with linear drift, power, trunca
Quantized average consensus with delay
Jafarian, Matin; De Persis, Claudio
2012-01-01
Average consensus problem is a special case of cooperative control in which the agents of the network asymptotically converge to the average state (i.e., position) of the network by transferring information via a communication topology. One of the issues of the large scale networks is the cost of co
24-Hour Relativistic Bit Commitment
Verbanis, Ephanielle; Martin, Anthony; Houlmann, Raphaël; Boso, Gianluca; Bussières, Félix; Zbinden, Hugo
2016-09-01
Bit commitment is a fundamental cryptographic primitive in which a party wishes to commit a secret bit to another party. Perfect security between mistrustful parties is unfortunately impossible to achieve through the asynchronous exchange of classical and quantum messages. Perfect security can nonetheless be achieved if each party splits into two agents exchanging classical information at times and locations satisfying strict relativistic constraints. A relativistic multiround protocol to achieve this was previously proposed and used to implement a 2-millisecond commitment time. Much longer durations were initially thought to be insecure, but recent theoretical progress showed that this is not so. In this Letter, we report on the implementation of a 24-hour bit commitment solely based on timed high-speed optical communication and fast data processing, with all agents located within the city of Geneva. This duration is more than 6 orders of magnitude longer than before, and we argue that it could be extended to one year and allow much more flexibility on the locations of the agents. Our implementation offers a practical and viable solution for use in applications such as digital signatures, secure voting and honesty-preserving auctions.
Trajectory averaging for stochastic approximation MCMC algorithms
Liang, Faming
2010-01-01
The subject of stochastic approximation was founded by Robbins and Monro [Ann. Math. Statist. 22 (1951) 400--407]. After five decades of continual development, it has developed into an important area in systems control and optimization, and it has also served as a prototype for the development of adaptive algorithms for on-line estimation and control of stochastic systems. Recently, it has been used in statistics with Markov chain Monte Carlo for solving maximum likelihood estimation problems and for general simulation and optimizations. In this paper, we first show that the trajectory averaging estimator is asymptotically efficient for the stochastic approximation MCMC (SAMCMC) algorithm under mild conditions, and then apply this result to the stochastic approximation Monte Carlo algorithm [Liang, Liu and Carroll J. Amer. Statist. Assoc. 102 (2007) 305--320]. The application of the trajectory averaging estimator to other stochastic approximation MCMC algorithms, for example, a stochastic approximation MLE al...
Bravo, J. L [Instituto de Geofisica, UNAM, Mexico, D.F. (Mexico); Nava, M. M [Instituto Mexicano del Petroleo, Mexico, D.F. (Mexico); Gay, C [Centro de Ciencias de la Atmosfera, UNAM, Mexico, D.F. (Mexico)
2001-07-01
We developed a procedure to forecast, with 2 or 3 hours, the daily maximum of surface ozone concentrations. It involves the adjustment of Autoregressive Integrated and Moving Average (ARIMA) models to daily ozone maximum concentrations at 10 monitoring atmospheric stations in Mexico City during one-year period. A one-day forecast is made and it is adjusted with the meteorological and solar radiation information acquired during the first 3 hours before the occurrence of the maximum value. The relative importance for forecasting of the history of the process and of meteorological conditions is evaluated. Finally an estimate of the daily probability of exceeding a given ozone level is made. [Spanish] Se aplica un procedimiento basado en la metodologia conocida como ARIMA, para predecir, con 2 o 3 horas de anticipacion, el valor maximo de la concentracion diaria de ozono. Esta basado en el calculo de autorregresiones y promedios moviles aplicados a los valores maximos de ozono superficial provenientes de 10 estaciones de monitoreo atmosferico en la Ciudad de Mexico y obtenidos durante un ano de muestreo. El pronostico para un dia se ajusta con la informacion meteorologica y de radiacion solar correspondiente a un periodo que antecede con al menos tres horas la ocurrencia esperada del valor maximo. Se compara la importancia relativa de la historia del proceso y de las condiciones meteorologicas previas para el pronostico. Finalmente se estima la probabilidad diaria de que un nivel normativo o preestablecido para contingencias de ozono sea rebasado.
Gaussian moving averages and semimartingales
Basse-O'Connor, Andreas
2008-01-01
In the present paper we study moving averages (also known as stochastic convolutions) driven by a Wiener process and with a deterministic kernel. Necessary and sufficient conditions on the kernel are provided for the moving average to be a semimartingale in its natural filtration. Our results...... are constructive - meaning that they provide a simple method to obtain kernels for which the moving average is a semimartingale or a Wiener process. Several examples are considered. In the last part of the paper we study general Gaussian processes with stationary increments. We provide necessary and sufficient...
Effects of renal sympathetic denervation on 24-hour blood pressure variability
Christine Stefanie Zuern
2012-05-01
Full Text Available Background: In patients with arterial hypertension, increased blood pressure (BP variability contributes to end organ damage independently from mean levels of arterial BP. Increased BP variability has been linked to alterations in autonomic function including sympathetic overdrive. We hypothesized that catheter-based renal sympathetic denervation (RDN confers beneficial effects on BPV. Methods and Results: Eleven consecutive patients with therapy-refractory arterial hypertension (age 68.9±7.0 years; baseline systolic BP 189±23mmHg despite medication with 5.6±2.1 antihypertensive drugs underwent bilateral RDN. Twenty-four hour ambulatory blood pressure monitoring (ABPM was performed before RDN and six months thereafter. BPV was primarily assessed by means of standard deviation of 24-hour systolic arterial blood pressures (SDsys. Secondary measures of BPV were maximum systolic blood pressure (MAXsys and maximum difference between two consecutive readings of systolic BP (deltamaxsys over 24 hours. Six months after RDN, SDsys, MAXsys and deltamaxsys were significantly reduced from 16.9±4.6mmHg to 13.5±2.5mmHg (p=0.003, from 190±22mmHg to 172±20mmHg (p<0.001 and from 40±15mmHg to 28±7mmHg (p=0.006, respectively, without changes in concomitant antihypertensive therapy. Reductions of SDsys, MAXsys and deltamaxsys were observed in 10/11 (90.9%, 11/11 (100% and 9/11 (81.8% patients, respectively. Although we noted a significant reduction of systolic office blood pressure by 30.4±27.7mmHg (p=0.007, there was only a trend in reduction of average systolic BP assessed from ABPM (149±19mmHg to 142±18mmHg; p=0.086.Conclusions: In patients with therapy-refractory arterial hypertension, RDN leads to significant reductions of BP variability. Effects of RDN on BPV over 24 hours were more pronounced than on average levels of BP.
How Hours of Work Affect Occupational Earnings.
Hecker, David
1998-01-01
In most managerial, management-related, sales, production, and transportation occupations, workers with longer hours earned a high hourly rate. The reverse was true for some jobs, including computer specialists, engineers, schoolteachers, and construction workers. (JOW)
CERN restaurants: opening hours during summer
2016-01-01
In the summer, the three CERN restaurants remain open during their usual hours. On Monday 1st August and Thursday 8 September, the Restaurant 1 will be open from 7:00 a.m. to 10:00 p.m. The satellites will be open as follows: Building 6: normal hours Building 13: normal hours Building 30: normal hours Building 40: closing at 4:30 p.m. instead of 5:00 pm Building 54: normal hours in July, closed in August Building 864: normal hours Building 865: normal hours Building 774: normal hours
Vocal attractiveness increases by averaging.
Bruckert, Laetitia; Bestelmeyer, Patricia; Latinus, Marianne; Rouger, Julien; Charest, Ian; Rousselet, Guillaume A; Kawahara, Hideki; Belin, Pascal
2010-01-26
Vocal attractiveness has a profound influence on listeners-a bias known as the "what sounds beautiful is good" vocal attractiveness stereotype [1]-with tangible impact on a voice owner's success at mating, job applications, and/or elections. The prevailing view holds that attractive voices are those that signal desirable attributes in a potential mate [2-4]-e.g., lower pitch in male voices. However, this account does not explain our preferences in more general social contexts in which voices of both genders are evaluated. Here we show that averaging voices via auditory morphing [5] results in more attractive voices, irrespective of the speaker's or listener's gender. Moreover, we show that this phenomenon is largely explained by two independent by-products of averaging: a smoother voice texture (reduced aperiodicities) and a greater similarity in pitch and timbre with the average of all voices (reduced "distance to mean"). These results provide the first evidence for a phenomenon of vocal attractiveness increases by averaging, analogous to a well-established effect of facial averaging [6, 7]. They highlight prototype-based coding [8] as a central feature of voice perception, emphasizing the similarity in the mechanisms of face and voice perception.
Work Hours Constraints: Impacts and Policy Implications
Constant, Amelie F.; Otterbach, Steffen
2011-01-01
If individuals reveal their preference as consumers, then they are taken seriously. What happens if individuals, as employees, reveal their preferences in working hours? And what happens if there is a misalignment between actual hours worked and preferred hours, the so-called work hours constraints? How does this affect the productivity of workers, their health, and overall life satisfaction? Labor supply and corresponding demand are fundamental to production. Labor economists know for long t...
Zheng, Yongguang; Xue, Ming; Li, Bo; Chen, Jiong; Tao, Zuyu
2016-11-01
Hourly rainfall measurements of 1919 national-level meteorological stations from 1981 through 2012 are used to document, for the first time, the climatology of extreme rainfall in hourly through 24-h accumulation periods in China. Rainfall amounts for 3-, 6-, 12- and 24-h periods at each station are constructed through running accumulation from hourly rainfall data that have been screened by proper quality control procedures. For each station and for each accumulation period, the historical maximum is found, and the corresponding 50-year return values are estimated using generalized extreme value theory. Based on the percentiles of the two types of extreme rainfall values among all the stations, standard thresholds separating Grade I, Grade II and Grade III extreme rainfall are established, which roughly correspond to the 70th and 90th percentiles for each of the accumulation periods. The spatial characteristics of the two types of extreme rainfall are then examined for different accumulation periods. The spatial distributions of extreme rainfall in hourly through 6-h periods are more similar than those of 12- and 24-h periods. Grade III rainfall is mostly found over South China, the western Sichuan Basin, along the southern and eastern coastlines, and in the large river basins and plains. There are similar numbers of stations with Grade III extreme hourly rainfall north and south of 30°N, but the percentage increases to about 70% south of 30°N as the accumulation period increases to 24 hours, reflecting richer moisture and more prolonged rain events in southern China. Potential applications of the extreme rainfall climatology and classification standards are suggested at the end.
The maximum rotation of a galactic disc
Bottema, R
1997-01-01
The observed stellar velocity dispersions of galactic discs show that the maximum rotation of a disc is on average 63% of the observed maximum rotation. This criterion can, however, not be applied to small or low surface brightness (LSB) galaxies because such systems show, in general, a continuously rising rotation curve until the outermost measured radial position. That is why a general relation has been derived, giving the maximum rotation for a disc depending on the luminosity, surface brightness, and colour of the disc. As a physical basis of this relation serves an adopted fixed mass-to-light ratio as a function of colour. That functionality is consistent with results from population synthesis models and its absolute value is determined from the observed stellar velocity dispersions. The derived maximum disc rotation is compared with a number of observed maximum rotations, clearly demonstrating the need for appreciable amounts of dark matter in the disc region and even more so for LSB galaxies. Matters h...
Cycle Average Peak Fuel Temperature Prediction Using CAPP/GAMMA+
Tak, Nam-il; Lee, Hyun Chul; Lim, Hong Sik [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)
2015-10-15
In order to obtain a cycle average maximum fuel temperature without rigorous efforts, a neutronics/thermo-fluid coupled calculation is needed with depletion capability. Recently, a CAPP/GAMMA+ coupled code system has been developed and the initial core of PMR200 was analyzed using the CAPP/GAMMA+ code system. The GAMMA+ code is a system thermo-fluid analysis code and the CAPP code is a neutronics code. The General Atomics proposed that the design limit of the fuel temperature under normal operating conditions should be a cycle-averaged maximum value. Nonetheless, the existing works of Korea Atomic Energy Research Institute (KAERI) only calculated the maximum fuel temperature at a fixed time point, e.g., the beginning of cycle (BOC) just because the calculation capability was not ready for a cycle average value. In this work, a cycle average maximum fuel temperature has been calculated using CAPP/GAMMA+ code system for the equilibrium core of PMR200. The CAPP/GAMMA+ coupled calculation was carried out for the equilibrium core of PMR 200 from BOC to EOC to obtain a cycle average peak fuel temperature. The peak fuel temperature was predicted to be 1372 .deg. C near MOC. However, the cycle average peak fuel temperature was calculated as 1181 .deg. C, which is below the design target of 1250 .deg. C.
Averaged Electroencephalic Audiometry in Infants
Lentz, William E.; McCandless, Geary A.
1971-01-01
Normal, preterm, and high-risk infants were tested at 1, 3, 6, and 12 months of age using averaged electroencephalic audiometry (AEA) to determine the usefulness of AEA as a measurement technique for assessing auditory acuity in infants, and to delineate some of the procedural and technical problems often encountered. (KW)
Ergodic averages via dominating processes
Møller, Jesper; Mengersen, Kerrie
2006-01-01
We show how the mean of a monotone function (defined on a state space equipped with a partial ordering) can be estimated, using ergodic averages calculated from upper and lower dominating processes of a stationary irreducible Markov chain. In particular, we do not need to simulate the stationary ...
29 CFR 778.319 - Paying for but not counting hours worked.
2010-07-01
... STATEMENTS OF GENERAL POLICY OR INTERPRETATION NOT DIRECTLY RELATED TO REGULATIONS OVERTIME COMPENSATION... more hours have been worked, the employee must be paid overtime compensation at not less than one and... of the applicable maximum hours standard, extra half-time compensation will be due regardless of...
High average power supercontinuum sources
J C Travers
2010-11-01
The physical mechanisms and basic experimental techniques for the creation of high average spectral power supercontinuum sources is briefly reviewed. We focus on the use of high-power ytterbium-doped fibre lasers as pump sources, and the use of highly nonlinear photonic crystal fibres as the nonlinear medium. The most common experimental arrangements are described, including both continuous wave fibre laser systems with over 100 W pump power, and picosecond mode-locked, master oscillator power fibre amplifier systems, with over 10 kW peak pump power. These systems can produce broadband supercontinua with over 50 and 1 mW/nm average spectral power, respectively. Techniques for numerical modelling of the supercontinuum sources are presented and used to illustrate some supercontinuum dynamics. Some recent experimental results are presented.
Dependability in Aggregation by Averaging
Jesus, Paulo; Almeida, Paulo Sérgio
2010-01-01
Aggregation is an important building block of modern distributed applications, allowing the determination of meaningful properties (e.g. network size, total storage capacity, average load, majorities, etc.) that are used to direct the execution of the system. However, the majority of the existing aggregation algorithms exhibit relevant dependability issues, when prospecting their use in real application environments. In this paper, we reveal some dependability issues of aggregation algorithms based on iterative averaging techniques, giving some directions to solve them. This class of algorithms is considered robust (when compared to common tree-based approaches), being independent from the used routing topology and providing an aggregation result at all nodes. However, their robustness is strongly challenged and their correctness often compromised, when changing the assumptions of their working environment to more realistic ones. The correctness of this class of algorithms relies on the maintenance of a funda...
Measuring Complexity through Average Symmetry
Alamino, Roberto C.
2015-01-01
This work introduces a complexity measure which addresses some conflicting issues between existing ones by using a new principle - measuring the average amount of symmetry broken by an object. It attributes low (although different) complexity to either deterministic or random homogeneous densities and higher complexity to the intermediate cases. This new measure is easily computable, breaks the coarse graining paradigm and can be straightforwardly generalised, including to continuous cases an...
Morrison, Glenn; Shaughnessy, Richard; Shu, Shi
2011-02-01
A Monte Carlo analysis of indoor ozone levels in four cities was applied to provide guidance to regulatory agencies on setting maximum ozone emission rates from consumer appliances. Measured distributions of air exchange rates, ozone decay rates and outdoor ozone levels at monitoring stations were combined with a steady-state indoor air quality model resulting in emission rate distributions (mg h -1) as a function of % of building hours protected from exceeding a target maximum indoor concentration of 20 ppb. Whole-year, summer and winter results for Elizabeth, NJ, Houston, TX, Windsor, ON, and Los Angeles, CA exhibited strong regional differences, primarily due to differences in air exchange rates. Infiltration of ambient ozone at higher average air exchange rates significantly reduces allowable emission rates, even though air exchange also dilutes emissions from appliances. For Houston, TX and Windsor, ON, which have lower average residential air exchange rates, emission rates ranged from -1.1 to 2.3 mg h -1 for scenarios that protect 80% or more of building hours from experiencing ozone concentrations greater than 20 ppb in summer. For Los Angeles, CA and Elizabeth, NJ, with higher air exchange rates, only negative emission rates were allowable to provide the same level of protection. For the 80th percentile residence, we estimate that an 8-h average limit concentration of 20 ppb would be exceeded, even in the absence of an indoor ozone source, 40 or more days per year in any of the cities analyzed. The negative emission rates emerging from the analysis suggest that only a zero-emission rate standard is prudent for Los Angeles, Elizabeth, NJ and other regions with higher summertime air exchange rates. For regions such as Houston with lower summertime air exchange rates, the higher emission rates would likely increase occupant exposure to the undesirable products of ozone reactions, thus reinforcing the need for zero-emission rate standard.
Jürgens, C; Antal, S; Henrici, K; Grossjohann, R; Tost, F H
2009-01-01
Elevated intraocular pressure (IOP) is a clinically relevant factor in glaucoma progression. As a dynamic parameter the IOP depends on various internal and exogenic influencing factors. Therefore, we analysed intraindividual IOD variations between ambulant care and 24-h home-monitoring using self-tonometry. This study is based on paper-based glaucoma cards of 25 patients with primary open angle glaucoma. Additionally, all patients participated in a telemedical home-monitoring study with self-measurements of IOP and blood pressure stored in an electronic patient record. The glaucoma cards contained a total number of 409 IOP values with documentation periods from 0.5 to 10 years. In the teletonometry project all 25 patients were observed for 6 months with 1490 recorded IOP values. Statistical analysis was performed with SPSS software. Average IOP values for all 25 glaucoma patients were 16.3 +/- 2.9 mmHg for both eyes in glaucoma card documentation, whereas the electronic patient records showed IOP averages of 18.9 +/- 4.7 mmHg for right eyes and 18.2 +/- 4.4 mmHg for left eyes. Corresponding to the practice opening hours the glaucoma cards contained no IOP records from 12:00 pm to 01:30 pm as well as between 06:00 pm and 07:15 am. In these time periods 17 % of all IOP values recorded in 24-hour teletonometry were higher than 20 mmHg. However, statistical analysis and clinical evaluation of device parameters and measurement characteristics revealed sporadic measuring errors. The additional involvement of self-tonometry in telemedical 24-h home-monitoring is a feasible method to record and detect intraday IOP fluctuations. Compared to single IOP measurements documented in common paper-based glaucoma cards, the 24-h electronic patient record showed more frequent circadian IOP variations. As a result, self-tonometry and home-monitoring can be a useful link to fill the gap between singular ambulant IOP measurement and hospitalisation with 24-hour IOP profiles.
Yasso, B; Li, Y; Alexander, A; Mel'nikova, N B; Mukhina, I V
2014-01-01
A comparison of the relative bioavailability and intensity of penetration of glucosamine sulfate in oral, injection and topical administration of the dosage form Hondroxid Maximum as a cream containing micellar system for transdermal delivery of glucosamine in the experiment by Sprague-Dawley rats was carried out. On the base on the pharmacokinetic profiles data of glucosamine in rat blood plasma with daily administration in 3 times a day for 1 week by cream Hondroxid Maximum 400 mg/kg and the single injection solution of 4% Glucosamine sulfate 400 mg/kg was found that the relative bioavailability was 61.6%. Calculated penetration rate of glucosamine in the plasma through the rats skin in 4 hours, equal to 26.9 μg/cm2 x h, and the penetration of glucosamine through the skin into the plasma after a single dose of cream in 4 hours was 4.12%. Comparative analysis of literature and experimental data and calculations based on them suggest that medicine Hondroxid Maximum, cream with transdermal glucosamine complex in the treatment in accordance with the instructions can provide an average concentration of glucosamine in the synovial fluid of an inflamed joint in the range (0.7 - 1.5) μg/ml, much higher than the concentration of endogenous glucosamine human synovial joint fluid (0.02 - 0.07 μg/ml). By theoretical calculations taking into account experimental data it is shown that the medicine Hondroxid Maximum can reach the bioavailability level of the modern injection forms and exceed the bioavailability level of modern oral forms of glucosamine up to 2 times.
Kim Oanh, N. T.; Martel, M.; Pongkiatkul, P.; Berkowicz, R.
Roadside air quality and vehicle emission are important and challenging issues in urban air quality management which need to be adequately characterized. This study involves designing a monitoring program that produces suitable data to determine the on-road hourly fleet emission and emission factors of individual vehicles in a street canyon. Simultaneous hourly monitoring of roadside gaseous pollutants (both windward and leeward sides), traffic volume and speed, and wind in a busy street of Bangkok was conducted in the rainy season when traffic emission was predominant in the city. Higher pollutant concentrations often occurred at midday (11:00 to 14:00h) when higher traffic density (3700-3800vehicles h - 1, weekdays) was observed. The levels of toluene and xylenes found in this study are higher than the roadside levels reported in other Asian cities. Hourly maximum concentrations reached 258ppb for toluene, 51ppb for m, p-xylenes, 15ppb for o-xylene, 526ppb for NO x, and 10.5ppm for CO. Hourly monitoring data during the periods when the street canyon effects were pronounced were selected for determination of the fleet hourly emission and vehicle emission factors by back calculation using a street canyon model (Operational Street Pollution Model). The average fleet hourly emission at daytime of NO x (6.2kg km - 1 h - 1), CO (54kg km - 1 h - 1), toluene (2.1kg km - 1 h - 1), m, p-xylenes (0.73kg km - 1 h - 1) and o-xylene (0.27kg km - 1 h - 1) did not vary much. However, the emission rates were substantially reduced at nighttime following the traffic pattern. The obtained pollutant emission factors varied within each group of vehicles with the average values agreed reasonably with the chassis dynamometer results for NO x but somewhat higher for CO and TX. The model estimated results are, however, considered to better represent the real driving conditions in the street at the average vehicle travel speed of around 20km h - 1. A statistical sampling design is proposed
Mirror averaging with sparsity priors
Dalalyan, Arnak
2010-01-01
We consider the problem of aggregating the elements of a (possibly infinite) dictionary for building a decision procedure, that aims at minimizing a given criterion. Along with the dictionary, an independent identically distributed training sample is available, on which the performance of a given procedure can be tested. In a fairly general set-up, we establish an oracle inequality for the Mirror Averaging aggregate based on any prior distribution. This oracle inequality is applied in the context of sparse coding for different problems of statistics and machine learning such as regression, density estimation and binary classification.
Zipf's law, power laws and maximum entropy
Visser, Matt
2013-04-01
Zipf's law, and power laws in general, have attracted and continue to attract considerable attention in a wide variety of disciplines—from astronomy to demographics to software structure to economics to linguistics to zoology, and even warfare. A recent model of random group formation (RGF) attempts a general explanation of such phenomena based on Jaynes' notion of maximum entropy applied to a particular choice of cost function. In the present paper I argue that the specific cost function used in the RGF model is in fact unnecessarily complicated, and that power laws can be obtained in a much simpler way by applying maximum entropy ideas directly to the Shannon entropy subject only to a single constraint: that the average of the logarithm of the observable quantity is specified.
Zipf's law, power laws, and maximum entropy
Visser, Matt
2012-01-01
Zipf's law, and power laws in general, have attracted and continue to attract considerable attention in a wide variety of disciplines - from astronomy to demographics to economics to linguistics to zoology, and even warfare. A recent model of random group formation [RGF] attempts a general explanation of such phenomena based on Jaynes' notion of maximum entropy applied to a particular choice of cost function. In the present article I argue that the cost function used in the RGF model is in fact unnecessarily complicated, and that power laws can be obtained in a much simpler way by applying maximum entropy ideas directly to the Shannon entropy subject only to a single constraint: that the average of the logarithm of the observable quantity is specified.
Trajectory averaging for stochastic approximation MCMC algorithms
Liang, Faming
2010-10-01
The subject of stochastic approximation was founded by Robbins and Monro [Ann. Math. Statist. 22 (1951) 400-407]. After five decades of continual development, it has developed into an important area in systems control and optimization, and it has also served as a prototype for the development of adaptive algorithms for on-line estimation and control of stochastic systems. Recently, it has been used in statistics with Markov chain Monte Carlo for solving maximum likelihood estimation problems and for general simulation and optimizations. In this paper, we first show that the trajectory averaging estimator is asymptotically efficient for the stochastic approximation MCMC (SAMCMC) algorithm under mild conditions, and then apply this result to the stochastic approximation Monte Carlo algorithm [Liang, Liu and Carroll J. Amer. Statist. Assoc. 102 (2007) 305-320]. The application of the trajectory averaging estimator to other stochastic approximationMCMC algorithms, for example, a stochastic approximation MLE algorithm for missing data problems, is also considered in the paper. © Institute of Mathematical Statistics, 2010.
MACHINE PROTECTION FOR HIGH AVERAGE CURRENT LINACS
Jordan, Kevin; Allison, Trent; Evans, Richard; Coleman, James; Grippo, Albert
2003-05-01
A fully integrated Machine Protection System (MPS) is critical to efficient commissioning and safe operation of all high current accelerators. The Jefferson Lab FEL [1,2] has multiple electron beam paths and many different types of diagnostic insertion devices. The MPS [3] needs to monitor both the status of these devices and the magnet settings which define the beam path. The matrix of these devices and beam paths are programmed into gate arrays, the output of the matrix is an allowable maximum average power limit. This power limit is enforced by the drive laser for the photocathode gun. The Beam Loss Monitors (BLMs), RF status, and laser safety system status are also inputs to the control matrix. There are 8 Machine Modes (electron path) and 8 Beam Modes (average power limits) that define the safe operating limits for the FEL. Combinations outside of this matrix are unsafe and the beam is inhibited. The power limits range from no beam to 2 megawatts of electron beam power.
Long term forecasting of hourly electricity consumption in local areas in Denmark
Møller Andersen, Frits; Larsen, Helge V.; Gaardestrup, R.B.
2013-01-01
Long term projections of hourly electricity consumption in local areas are important for planning of the transmission grid. In Denmark, at present the method used for grid planning is based on statistical analysis of the hour of maximum load and for each local area the maximum load is projected...... to change proportional to changes in the aggregated national electricity consumption. That is, specific local conditions are not considered. Yet, from measurements of local consumption we know that:. •consumption profiles differ between local areas,•consumption by categories of customers contribute....... The model describes the entire profile of hourly consumption and is a first step towards differentiated local predictions of electricity consumption.The model is based on metering of aggregated hourly consumption at transformer stations covering selected local areas and on national statistics of hourly...
OECD Maximum Residue Limit Calculator
With the goal of harmonizing the calculation of maximum residue limits (MRLs) across the Organisation for Economic Cooperation and Development, the OECD has developed an MRL Calculator. View the calculator.
24-Hour Academic Libraries: Adjusting to Change
Bowman, Adam C.
2013-01-01
The purpose of this study was to explore the adaptive measures that academic libraries perform when implementing and operating a 24-hour schedule. Five in-depth interviews were conducted with current managerial-level librarians at 24-hour academic libraries. The exploratory interviews revealed similar measures for security, budgeting, employee…
17 CFR 201.104 - Business hours.
2010-04-01
... 17 Commodity and Securities Exchanges 2 2010-04-01 2010-04-01 false Business hours. 201.104 Section 201.104 Commodity and Securities Exchanges SECURITIES AND EXCHANGE COMMISSION RULES OF PRACTICE Rules of Practice General Rules § 201.104 Business hours. The Headquarters office of the Commission, at...
20 CFR 801.304 - Business hours.
2010-04-01
... 20 Employees' Benefits 3 2010-04-01 2010-04-01 false Business hours. 801.304 Section 801.304 Employees' Benefits BENEFITS REVIEW BOARD, DEPARTMENT OF LABOR ESTABLISHMENT AND OPERATION OF THE BOARD Action by the Board § 801.304 Business hours. The office of the Clerk of the Board at Washington, DC...
Cost-efficient staffing under annualized hours
van der Veen, Egbert; Hans, Elias W.; Veltman, Bart; Berrevoets, Leo M.; Berden, Hubert J.J.M.
2012-01-01
We study how flexibility in workforce capacity can be used to efficiently match capacity and demand. Flexibility in workforce capacity is introduced by the annualized hours regime. Annualized hours allow organizations to measure working time per year, instead of per month or per week. An additional
Feliciangeli, M D
1997-03-01
The comparative hourly activity of Lutzomyia ovallesi (Ortiz) and L. gomezi (Nitzulescu), vectors of cutaneous leishmaniasis in Miranda State, Venezuela, was studied between November and March during 1991-1994 using a Shannon trap with a fluorescent light. Female abundance of L. ovallesi increased from 1800 to 2000 hours, plateaued from 2000 to 2400 hours, then decreased progressively. L. gomezi always exhibited maximum activity between 1900 and 2000 hours, then declined abruptly. The importance of these activity patterns in Leishmania transmission is discussed.
Ensemble average theory of gravity
Khosravi, Nima
2016-12-01
We put forward the idea that all the theoretically consistent models of gravity have contributions to the observed gravity interaction. In this formulation, each model comes with its own Euclidean path-integral weight where general relativity (GR) has automatically the maximum weight in high-curvature regions. We employ this idea in the framework of Lovelock models and show that in four dimensions the result is a specific form of the f (R ,G ) model. This specific f (R ,G ) satisfies the stability conditions and possesses self-accelerating solutions. Our model is consistent with the local tests of gravity since its behavior is the same as in GR for the high-curvature regime. In the low-curvature regime the gravitational force is weaker than in GR, which can be interpreted as the existence of a repulsive fifth force for very large scales. Interestingly, there is an intermediate-curvature regime where the gravitational force is stronger in our model compared to GR. The different behavior of our model in comparison with GR in both low- and intermediate-curvature regimes makes it observationally distinguishable from Λ CDM .
Preliminary analysis of the afforestation role in the maximum runoff in Valea Rece Catchment
Mihalcea Andreea
2017-06-01
Full Text Available The aim of this article is to demonstrate the afforestation role in maximum surface runoff. In this way, it was made a comparison of simulated flows in the current conditions of afforestation and the simulated flows in conditions of applying both afforestation and deforestation scenarios in Valea Rece catchment. Through HEC-HMS 4.1 hydrologic modeling software, using the method of unit hydrograph SCS Curve Number, were simulated flow of the river Valea Rece closing section of the basin, where precipitation amounts of 30,50,80,120 mm fallen in intervals of 1.3 to 6 hours on a soil with varying degrees of moisture: dry soil, average soil moisture and high humidity. This was done for the current degree of afforestation basin, for the results from a possible afforestation that would increase the afforestation degree to 80%, and for a possible deforestation that would lead to a degree of afforestation 15 %.
Are hourly precipitation extremes increasing faster than daily precipitation extremes?
Barbero, Renaud; Fowler, Hayley; Blenkinsop, Stephen; Lenderink, Geert
2016-04-01
Extreme precipitation events appear to be increasing with climate change in many regions of the world, including the United States. These extreme events have large societal impacts, as seen during the recent Texas-Oklahoma flooding in May 2015 which caused several billion in damages and left 47 deaths in its path. Better understanding of past changes in the characteristics of extreme rainfall events is thus critical for reliable projections of future changes. Although it has been documented in several studies that daily precipitation extremes are increasing across parts of the contiguous United States, very few studies have looked at hourly extremes. However, this is of primary importance as recent studies on the temperature scaling of extreme precipitation have shown that increases above the Clausius-Clapeyron (~ 7% °C-1) are possible for hourly precipitation. In this study, we used hourly precipitation data (HPD) from the National Climatic Data Center and extracted more than 1,000 stations across the US with more than 40 years of data spanning the period 1950-2010. As hourly measurements are often associated with a range of issues, the data underwent multiple quality control processes to exclude erroneous data. While no significant changes were found in annual maximum precipitation using both hourly and daily resolution datasets, significant increasing trends in terms of frequency of episodes exceeding present-day 95th percentiles of wet hourly/daily precipitation were observed across a significant portion of the US. The fraction of stations with significant increasing trends falls outside the confidence interval range during all seasons but the summer. While less than 12% of stations exhibit significant trends at the daily scale in the wintertime, more than 45% of stations, mostly clustered in central and Northern United States, show significant increasing trends at the hourly scale. This suggests that short-duration storms have increased faster than daily
Maximum margin Bayesian network classifiers.
Pernkopf, Franz; Wohlmayr, Michael; Tschiatschek, Sebastian
2012-03-01
We present a maximum margin parameter learning algorithm for Bayesian network classifiers using a conjugate gradient (CG) method for optimization. In contrast to previous approaches, we maintain the normalization constraints on the parameters of the Bayesian network during optimization, i.e., the probabilistic interpretation of the model is not lost. This enables us to handle missing features in discriminatively optimized Bayesian networks. In experiments, we compare the classification performance of maximum margin parameter learning to conditional likelihood and maximum likelihood learning approaches. Discriminative parameter learning significantly outperforms generative maximum likelihood estimation for naive Bayes and tree augmented naive Bayes structures on all considered data sets. Furthermore, maximizing the margin dominates the conditional likelihood approach in terms of classification performance in most cases. We provide results for a recently proposed maximum margin optimization approach based on convex relaxation. While the classification results are highly similar, our CG-based optimization is computationally up to orders of magnitude faster. Margin-optimized Bayesian network classifiers achieve classification performance comparable to support vector machines (SVMs) using fewer parameters. Moreover, we show that unanticipated missing feature values during classification can be easily processed by discriminatively optimized Bayesian network classifiers, a case where discriminative classifiers usually require mechanisms to complete unknown feature values in the data first.
Maximum Entropy in Drug Discovery
Chih-Yuan Tseng
2014-07-01
Full Text Available Drug discovery applies multidisciplinary approaches either experimentally, computationally or both ways to identify lead compounds to treat various diseases. While conventional approaches have yielded many US Food and Drug Administration (FDA-approved drugs, researchers continue investigating and designing better approaches to increase the success rate in the discovery process. In this article, we provide an overview of the current strategies and point out where and how the method of maximum entropy has been introduced in this area. The maximum entropy principle has its root in thermodynamics, yet since Jaynes’ pioneering work in the 1950s, the maximum entropy principle has not only been used as a physics law, but also as a reasoning tool that allows us to process information in hand with the least bias. Its applicability in various disciplines has been abundantly demonstrated. We give several examples of applications of maximum entropy in different stages of drug discovery. Finally, we discuss a promising new direction in drug discovery that is likely to hinge on the ways of utilizing maximum entropy.
Ozone Nonattainment Areas - 8 Hour (1997 Standard)
U.S. Environmental Protection Agency — This data layer identifies areas in the U.S. where air pollution levels have not met the National Ambient Air Quality Standards (NAAQS) for ozone over 8 hours and...
U.S. Hourly Precipitation Data Publication
National Oceanic and Atmospheric Administration, Department of Commerce — This publication contains hourly precipitation amounts obtained from recording rain gages located at National Weather Service, Federal Aviation Administration, and...
U.S. Hourly Precipitation Data
National Oceanic and Atmospheric Administration, Department of Commerce — Hourly Precipitation Data (HPD) is digital data set DSI-3240, archived at the National Climatic Data Center (NCDC). The primary source of data for this file is...
SO2 8 Hour Nonattainment Areas
U.S. Environmental Protection Agency — This data layer identifies areas in the U.S. where air pollution levels have not met the National Ambient Air Quality Standards (NAAQS) for Sulfur dioxide 8 hour...
Recurrence Within 24 Hours of Unprovoked Seizures
J Gordon Millichap
2006-09-01
Full Text Available The incidence and risk factors of acute recurrence of unprovoked seizures within 24 hours of admission to an Emergency Department (ED were analyzed by a retrospective chart review at Schneider Children’s Hospital, New York.
Long working hours and alcohol use
Virtanen, Marianna; Jokela, Markus; Nyberg, Solja T
2015-01-01
OBJECTIVE: To quantify the association between long working hours and alcohol use. DESIGN: Systematic review and meta-analysis of published studies and unpublished individual participant data. DATA SOURCES: A systematic search of PubMed and Embase databases in April 2014 for published studies......, supplemented with manual searches. Unpublished individual participant data were obtained from 27 additional studies. REVIEW METHODS: The search strategy was designed to retrieve cross sectional and prospective studies of the association between long working hours and alcohol use. Summary estimates were.......20) in the analysis of prospective published and unpublished data. In the 18 studies with individual participant data it was possible to assess the European Union Working Time Directive, which recommends an upper limit of 48 hours a week. Odds ratios of new onset risky alcohol use for those working 49-54 hours...
Longer opening hours, alcohol consumption and health
Green, Colin Peter; Hollingsworth, Bruce Philip; Navarro Paniagua, Maria
2015-01-01
Two related issues in public policy with respect to alcohol are how increased availability influences consumption and what effect excess consumption has on individual health outcomes. This paper examines one particular source of variation in availability, bar opening hours, and how this influences alcohol consumption, physical and mental health. We focus on the extension of opening hours in England and Wales that occurred in 2005. We demonstrate a marked increase in consumption, which appears...
Characterization and simulation of hourly exposure series of global radiation
Mora-Lopez, L. [Universidad de Malaga (Spain). Dpto. Lenguajes y C. Computacion; Sidrach-de-Cardona, M. [Universidad de Malaga (Spain). Dpto. Fisica Aplicada
1997-06-01
A statistical model which captures the main features of hourly exposure series of global radiation is proposed. This model is used to obtain a procedure to generate radiation series without imposing, a priori, any restriction on the form of the probability distribution function of the series. The statistical model was taken from the stationary stochastic processes theory. Data were obtained from ten different Spanish locations. As monthly hourly exposure series of global radiation are not stationary, they are modified in order to remove the observed trends. A multiplicative autoregressive moving average model with regular and seasonal components was used. It is statistically accepted that this is the true model which generates most of the analyzed sequences. However, the underlying parameters of the model vary from one location to another and from one month to another. Therefore, it is necessary to examine further the relationship between the parameters of the model and the available data from most locations. (author)
Ex Situ Perfusion of Human Limb Allografts for 24 Hours.
Werner, Nicole L; Alghanem, Fares; Rakestraw, Stephanie L; Sarver, Dylan C; Nicely, Bruce; Pietroski, Richard E; Lange, Paul; Rudich, Steven M; Mendias, Christopher L; Rojas-Pena, Alvaro; Magee, John C; Bartlett, Robert H; Ozer, Kagan
2017-03-01
Vascularized composite allografts, particularly hand and forearm, have limited ischemic tolerance after procurement. In bilateral hand transplantations, this demands a 2 team approach and expedited transfer of the allograft, limiting the recovery to a small geographic area. Ex situ perfusion may be an alternative allograft preservation method to extend allograft survival time. This is a short report of 5 human limbs maintained for 24 hours with ex situ perfusion. Upper limbs were procured from brain-dead organ donors. Following recovery, the brachial artery was cannulated and flushed with 10 000 U of heparin. The limb was then attached to a custom-made, near-normothermic (30-33°C) ex situ perfusion system composed of a pump, reservoir, and oxygenator. Perfusate was plasma-based with a hemoglobin concentration of 4 to 6 g/dL. Average warm ischemia time was 76 minutes. Perfusion was maintained at an average systolic pressure of 93 ± 2 mm Hg, flow 310 ± 20 mL/min, and vascular resistance 153 ± 16 mm Hg/L per minute. Average oxygen consumption was 1.1 ± 0.2 mL/kg per minute. Neuromuscular electrical stimulation continually displayed contraction until the end of perfusion, and histology showed no myocyte injury. Human limb allografts appeared viable after 24 hours of near-normothermic ex situ perfusion. Although these results are early and need validation with transplantation, this technology has promise for extending allograft storage times.
Yearly average performance of the principal solar collector types
Rabl, A.
1981-01-01
The results of hour-by-hour simulations for 26 meteorological stations are used to derive universal correlations for the yearly total energy that can be delivered by the principal solar collector types: flat plate, evacuated tubes, CPC, single- and dual-axis tracking collectors, and central receiver. The correlations are first- and second-order polynomials in yearly average insolation, latitude, and threshold (= heat loss/optical efficiency). With these correlations, the yearly collectible energy can be found by multiplying the coordinates of a single graph by the collector parameters, which reproduces the results of hour-by-hour simulations with an accuracy (rms error) of 2% for flat plates and 2% to 4% for concentrators. This method can be applied to collectors that operate year-around in such a way that no collected energy is discarded, including photovoltaic systems, solar-augmented industrial process heat systems, and solar thermal power systems. The method is also recommended for rating collectors of different type or manufacturer by yearly average performance, evaluating the effects of collector degradation, the benefits of collector cleaning, and the gains from collector improvements (due to enhanced optical efficiency or decreased heat loss per absorber surface). For most of these applications, the method is accurate enough to replace a system simulation.
Greenslade, Thomas B., Jr.
1985-01-01
Discusses a series of experiments performed by Thomas Hope in 1805 which show the temperature at which water has its maximum density. Early data cast into a modern form as well as guidelines and recent data collected from the author provide background for duplicating Hope's experiments in the classroom. (JN)
Abolishing the maximum tension principle
Dabrowski, Mariusz P
2015-01-01
We find the series of example theories for which the relativistic limit of maximum tension $F_{max} = c^2/4G$ represented by the entropic force can be abolished. Among them the varying constants theories, some generalized entropy models applied both for cosmological and black hole horizons as well as some generalized uncertainty principle models.
Abolishing the maximum tension principle
Mariusz P. Da̧browski
2015-09-01
Full Text Available We find the series of example theories for which the relativistic limit of maximum tension Fmax=c4/4G represented by the entropic force can be abolished. Among them the varying constants theories, some generalized entropy models applied both for cosmological and black hole horizons as well as some generalized uncertainty principle models.
Maximum Photovoltaic Penetration Levels on Typical Distribution Feeders: Preprint
Hoke, A.; Butler, R.; Hambrick, J.; Kroposki, B.
2012-07-01
This paper presents simulation results for a taxonomy of typical distribution feeders with various levels of photovoltaic (PV) penetration. For each of the 16 feeders simulated, the maximum PV penetration that did not result in steady-state voltage or current violation is presented for several PV location scenarios: clustered near the feeder source, clustered near the midpoint of the feeder, clustered near the end of the feeder, randomly located, and evenly distributed. In addition, the maximum level of PV is presented for single, large PV systems at each location. Maximum PV penetration was determined by requiring that feeder voltages stay within ANSI Range A and that feeder currents stay within the ranges determined by overcurrent protection devices. Simulations were run in GridLAB-D using hourly time steps over a year with randomized load profiles based on utility data and typical meteorological year weather data. For 86% of the cases simulated, maximum PV penetration was at least 30% of peak load.
On the maximum sufficient range of interstellar vessels
Cartin, Daniel
2011-01-01
This paper considers the likely maximum range of space vessels providing the basis of a mature interstellar transportation network. Using the principle of sufficiency, it is argued that this range will be less than three parsecs for the average interstellar vessel. This maximum range provides access from the Solar System to a large majority of nearby stellar systems, with total travel distances within the network not excessively greater than actual physical distance.
Global solar radiation estimation from relative sunshine hours in Yemen
Abdul-Aziz, J.; A-Nagi, A.; Zumailan, A.A.R. (Aden Univ. (Yemen, Republic of))
1993-01-01
The existing measurements of global solar radiation and sunshine duration for Yemen are examined. The errors of estimating global solar radiation from sunshine hour measurements using Angstrom's relation are evaluated. As a simple predictor for global solar radiation, an average Angstrom relation in the form H-bar/H[sub o] = 0.3518 + 0.3162 n/N for all stations is evaluated. Other Angstrom correlation relations are also proposed by classifying the stations under study into four groups. The estimated results are compared and seem to be satisfactory in the latter case. (Author)
How extreme is extreme hourly precipitation?
Papalexiou, Simon Michael; Dialynas, Yannis G.; Pappas, Christoforos
2016-04-01
The importance of accurate representation of precipitation at fine time scales (e.g., hourly), directly associated with flash flood events, is crucial in hydrological design and prediction. The upper part of a probability distribution, known as the distribution tail, determines the behavior of extreme events. In general, and loosely speaking, tails can be categorized in two families: the subexponential and the hyperexponential family, with the first generating more intense and more frequent extremes compared to the latter. In past studies, the focus has been mainly on daily precipitation, with the Gamma distribution being the most popular model. Here, we investigate the behaviour of tails of hourly precipitation by comparing the upper part of empirical distributions of thousands of records with three general types of tails corresponding to the Pareto, Lognormal, and Weibull distributions. Specifically, we use thousands of hourly rainfall records from all over the USA. The analysis indicates that heavier-tailed distributions describe better the observed hourly rainfall extremes in comparison to lighter tails. Traditional representations of the marginal distribution of hourly rainfall may significantly deviate from observed behaviours of extremes, with direct implications on hydroclimatic variables modelling and engineering design.
Maximum Genus of Strong Embeddings
Er-ling Wei; Yan-pei Liu; Han Ren
2003-01-01
The strong embedding conjecture states that any 2-connected graph has a strong embedding on some surface. It implies the circuit double cover conjecture: Any 2-connected graph has a circuit double cover.Conversely, it is not true. But for a 3-regular graph, the two conjectures are equivalent. In this paper, a characterization of graphs having a strong embedding with exactly 3 faces, which is the strong embedding of maximum genus, is given. In addition, some graphs with the property are provided. More generally, an upper bound of the maximum genus of strong embeddings of a graph is presented too. Lastly, it is shown that the interpolation theorem is true to planar Halin graph.
Rosati, F; M. Rossi
2001-01-01
We analyse the determinants of school attendance and hours worked by children in Pakistan and Nicaragua. On the basis of a theoretical model of children's labour supply, we simultaneously estimate the school attendance decision and the hours worked by Full Model Maximum Likelihood. We analyse the marginal effects of explanatory variables conditioning on the "latent" status of children in terms of schooling and work. We show that these effects are rather different, and discuss the policy impli...
Remizov, Ivan D
2009-01-01
In this note, we represent a subdifferential of a maximum functional defined on the space of all real-valued continuous functions on a given metric compact set. For a given argument, $f$ it coincides with the set of all probability measures on the set of points maximizing $f$ on the initial compact set. This complete characterization lies in the heart of several important identities in microeconomics, such as Roy's identity, Sheppard's lemma, as well as duality theory in production and linear programming.
Average monthly and annual climate maps for Bolivia
Vicente-Serrano, Sergio M.
2015-02-24
This study presents monthly and annual climate maps for relevant hydroclimatic variables in Bolivia. We used the most complete network of precipitation and temperature stations available in Bolivia, which passed a careful quality control and temporal homogenization procedure. Monthly average maps at the spatial resolution of 1 km were modeled by means of a regression-based approach using topographic and geographic variables as predictors. The monthly average maximum and minimum temperatures, precipitation and potential exoatmospheric solar radiation under clear sky conditions are used to estimate the monthly average atmospheric evaporative demand by means of the Hargreaves model. Finally, the average water balance is estimated on a monthly and annual scale for each 1 km cell by means of the difference between precipitation and atmospheric evaporative demand. The digital layers used to create the maps are available in the digital repository of the Spanish National Research Council.
The Testability of Maximum Magnitude
Clements, R.; Schorlemmer, D.; Gonzalez, A.; Zoeller, G.; Schneider, M.
2012-12-01
Recent disasters caused by earthquakes of unexpectedly large magnitude (such as Tohoku) illustrate the need for reliable assessments of the seismic hazard. Estimates of the maximum possible magnitude M at a given fault or in a particular zone are essential parameters in probabilistic seismic hazard assessment (PSHA), but their accuracy remains untested. In this study, we discuss the testability of long-term and short-term M estimates and the limitations that arise from testing such rare events. Of considerable importance is whether or not those limitations imply a lack of testability of a useful maximum magnitude estimate, and whether this should have any influence on current PSHA methodology. We use a simple extreme value theory approach to derive a probability distribution for the expected maximum magnitude in a future time interval, and we perform a sensitivity analysis on this distribution to determine if there is a reasonable avenue available for testing M estimates as they are commonly reported today: devoid of an appropriate probability distribution of their own and estimated only for infinite time (or relatively large untestable periods). Our results imply that any attempt at testing such estimates is futile, and that the distribution is highly sensitive to M estimates only under certain optimal conditions that are rarely observed in practice. In the future we suggest that PSHA modelers be brutally honest about the uncertainty of M estimates, or must find a way to decrease its influence on the estimated hazard.
Alternative Multiview Maximum Entropy Discrimination.
Chao, Guoqing; Sun, Shiliang
2016-07-01
Maximum entropy discrimination (MED) is a general framework for discriminative estimation based on maximum entropy and maximum margin principles, and can produce hard-margin support vector machines under some assumptions. Recently, the multiview version of MED multiview MED (MVMED) was proposed. In this paper, we try to explore a more natural MVMED framework by assuming two separate distributions p1( Θ1) over the first-view classifier parameter Θ1 and p2( Θ2) over the second-view classifier parameter Θ2 . We name the new MVMED framework as alternative MVMED (AMVMED), which enforces the posteriors of two view margins to be equal. The proposed AMVMED is more flexible than the existing MVMED, because compared with MVMED, which optimizes one relative entropy, AMVMED assigns one relative entropy term to each of the two views, thus incorporating a tradeoff between the two views. We give the detailed solving procedure, which can be divided into two steps. The first step is solving our optimization problem without considering the equal margin posteriors from two views, and then, in the second step, we consider the equal posteriors. Experimental results on multiple real-world data sets verify the effectiveness of the AMVMED, and comparisons with MVMED are also reported.
Longer Opening Hours for the Library
2001-01-01
The scientific information service. The CERN library is open 24 hours a day, 365 days a year. So how, you might be wondering, can they improve on that? The answer is in the detail. Although you can already use the library whenever you want, items can only be checked out when the front desk is staffed. A decision taken last week by the Scientific Information Policy Board now means that there will someone at the desk through out CERN's official working hours, with an extra 90 minutes at the end of the day so that people can check out material on their way home. In other words, the library will be open from 8:30 to 19:00, Monday to Friday. The library continues, of course, to be open 24 hours a day, all year round, and services provided via the digital library remain at your disposal day and night: http://library.cern.ch
Avoiding Lawsuits for Wage and Hour Violations.
Silberman, Cherie L
2016-01-01
Due to the highly technical language in the wage and hour laws and regulations, employers often find that they have unknowingly violated the Fair Labor Standards Act (FLSA). This can occur because employers have improperly classified an employee as exempt or because employers do not realize that certain time should be paid in full. Improperly classifying employees as exempt or failing to compensate nonexempt employees for all time worked can lead to costly lawsuits, audits, or enforcement actions by the Wage and Hour Division of the Department of Labor. This article discusses the most common FLSA exemptions and provides best practices to avoid liability under the FLSA.
Maximum entropy principle and texture formation
Arminjon, M; Arminjon, Mayeul; Imbault, Didier
2006-01-01
The macro-to-micro transition in a heterogeneous material is envisaged as the selection of a probability distribution by the Principle of Maximum Entropy (MAXENT). The material is made of constituents, e.g. given crystal orientations. Each constituent is itself made of a large number of elementary constituents. The relevant probability is the volume fraction of the elementary constituents that belong to a given constituent and undergo a given stimulus. Assuming only obvious constraints in MAXENT means describing a maximally disordered material. This is proved to have the same average stimulus in each constituent. By adding a constraint in MAXENT, a new model, potentially interesting e.g. for texture prediction, is obtained.
29 CFR 778.110 - Hourly rate employee.
2010-07-01
... number of hours worked in excess of 40 in the week. Thus a $6 hourly rate will bring, for an employee who....80 for 46 hours (46 hours at $6.20 plus 6 hours at $3.10, or 40 hours at $6.20 plus 6 hours at $9.30). ... Regulations Relating to Labor (Continued) WAGE AND HOUR DIVISION, DEPARTMENT OF LABOR STATEMENTS OF...
Measurement of the average lifetime of b hadrons
Adriani, O.; Aguilar-Benitez, M.; Ahlen, S.; Alcaraz, J.; Aloisio, A.; Alverson, G.; Alviggi, M. G.; Ambrosi, G.; An, Q.; Anderhub, H.; Anderson, A. L.; Andreev, V. P.; Angelescu, T.; Antonov, L.; Antreasyan, D.; Arce, P.; Arefiev, A.; Atamanchuk, A.; Azemoon, T.; Aziz, T.; Baba, P. V. K. S.; Bagnaia, P.; Bakken, J. A.; Ball, R. C.; Banerjee, S.; Bao, J.; Barillère, R.; Barone, L.; Baschirotto, A.; Battiston, R.; Bay, A.; Becattini, F.; Bechtluft, J.; Becker, R.; Becker, U.; Behner, F.; Behrens, J.; Bencze, Gy. L.; Berdugo, J.; Berges, P.; Bertucci, B.; Betev, B. L.; Biasini, M.; Biland, A.; Bilei, G. M.; Bizzarri, R.; Blaising, J. J.; Bobbink, G. J.; Bock, R.; Böhm, A.; Borgia, B.; Bosetti, M.; Bourilkov, D.; Bourquin, M.; Boutigny, D.; Bouwens, B.; Brambilla, E.; Branson, J. G.; Brock, I. C.; Brooks, M.; Bujak, A.; Burger, J. D.; Burger, W. J.; Busenitz, J.; Buytenhuijs, A.; Cai, X. D.; Capell, M.; Caria, M.; Carlino, G.; Cartacci, A. M.; Castello, R.; Cerrada, M.; Cesaroni, F.; Chang, Y. H.; Chaturvedi, U. K.; Chemarin, M.; Chen, A.; Chen, C.; Chen, G.; Chen, G. M.; Chen, H. F.; Chen, H. S.; Chen, M.; Chen, W. Y.; Chiefari, G.; Chien, C. Y.; Choi, M. T.; Chung, S.; Civinini, C.; Clare, I.; Clare, R.; Coan, T. E.; Cohn, H. O.; Coignet, G.; Colino, N.; Contin, A.; Costantini, S.; Cotorobai, F.; Cui, X. T.; Cui, X. Y.; Dai, T. S.; D'Alessandro, R.; de Asmundis, R.; Degré, A.; Deiters, K.; Dénes, E.; Denes, P.; DeNotaristefani, F.; Dhina, M.; DiBitonto, D.; Diemoz, M.; Dimitrov, H. R.; Dionisi, C.; Ditmarr, M.; Djambazov, L.; Dova, M. T.; Drago, E.; Duchesneau, D.; Duinker, P.; Duran, I.; Easo, S.; El Mamouni, H.; Engler, A.; Eppling, F. J.; Erné, F. C.; Extermann, P.; Fabbretti, R.; Fabre, M.; Falciano, S.; Fan, S. J.; Fackler, O.; Fay, J.; Felcini, M.; Ferguson, T.; Fernandez, D.; Fernandez, G.; Ferroni, F.; Fesefeldt, H.; Fiandrini, E.; Field, J. H.; Filthaut, F.; Fisher, P. H.; Forconi, G.; Fredj, L.; Freudenreich, K.; Friebel, W.; Fukushima, M.; Gailloud, M.; Galaktionov, Yu.; Gallo, E.; Ganguli, S. N.; Garcia-Abia, P.; Gele, D.; Gentile, S.; Gheordanescu, N.; Giagu, S.; Goldfarb, S.; Gong, Z. F.; Gonzalez, E.; Gougas, A.; Goujon, D.; Gratta, G.; Gruenewald, M.; Gu, C.; Guanziroli, M.; Guo, J. K.; Gupta, V. K.; Gurtu, A.; Gustafson, H. R.; Gutay, L. J.; Hangarter, K.; Hartmann, B.; Hasan, A.; Hauschildt, D.; He, C. F.; He, J. T.; Hebbeker, T.; Hebert, M.; Hervé, A.; Hilgers, K.; Hofer, H.; Hoorani, H.; Hu, G.; Hu, G. Q.; Ille, B.; Ilyas, M. M.; Innocente, V.; Janssen, H.; Jezequel, S.; Jin, B. N.; Jones, L. W.; Josa-Mutuberria, I.; Kasser, A.; Khan, R. A.; Kamyshkov, Yu.; Kapinos, P.; Kapustinsky, J. S.; Karyotakis, Y.; Kaur, M.; Khokhar, S.; Kienzle-Focacci, M. N.; Kim, J. K.; Kim, S. C.; Kim, Y. G.; Kinnison, W. W.; Kirkby, A.; Kirkby, D.; Kirsch, S.; Kittel, W.; Klimentov, A.; Klöckner, R.; König, A. C.; Koffeman, E.; Kornadt, O.; Koutsenko, V.; Koulbardis, A.; Kraemer, R. W.; Kramer, T.; Krastev, V. R.; Krenz, W.; Krivshich, A.; Kuijten, H.; Kumar, K. S.; Kunin, A.; Landi, G.; Lanske, D.; Lanzano, S.; Lebedev, A.; Lebrun, P.; Lecomte, P.; Lecoq, P.; Le Coultre, P.; Lee, D. M.; Lee, J. S.; Lee, K. Y.; Leedom, I.; Leggett, C.; Le Goff, J. M.; Leiste, R.; Lenti, M.; Leonardi, E.; Li, C.; Li, H. T.; Li, P. J.; Liao, J. Y.; Lin, W. T.; Lin, Z. Y.; Linde, F. L.; Lindemann, B.; Lista, L.; Liu, Y.; Lohmann, W.; Longo, E.; Lu, Y. S.; Lubbers, J. M.; Lübelsmeyer, K.; Luci, C.; Luckey, D.; Ludovici, L.; Luminari, L.; Lustermann, W.; Ma, J. M.; Ma, W. G.; MacDermott, M.; Malik, R.; Malinin, A.; Maña, C.; Maolinbay, M.; Marchesini, P.; Marion, F.; Marin, A.; Martin, J. P.; Martinez-Laso, L.; Marzano, F.; Massaro, G. G. G.; Mazumdar, K.; McBride, P.; McMahon, T.; McNally, D.; Merk, M.; Merola, L.; Meschini, M.; Metzger, W. J.; Mi, Y.; Mihul, A.; Mills, G. B.; Mir, Y.; Mirabelli, G.; Mnich, J.; Möller, M.; Monteleoni, B.; Morand, R.; Morganti, S.; Moulai, N. E.; Mount, R.; Müller, S.; Nadtochy, A.; Nagy, E.; Napolitano, M.; Nessi-Tedaldi, F.; Newman, H.; Neyer, C.; Niaz, M. A.; Nippe, A.; Nowak, H.; Organtini, G.; Pandoulas, D.; Paoletti, S.; Paolucci, P.; Pascale, G.; Passaleva, G.; Patricelli, S.; Paul, T.; Pauluzzi, M.; Paus, C.; Pauss, F.; Pei, Y. J.; Pensotti, S.; Perret-Gallix, D.; Perrier, J.; Pevsner, A.; Piccolo, D.; Pieri, M.; Piroué, P. A.; Plasil, F.; Plyaskin, V.; Pohl, M.; Pojidaev, V.; Postema, H.; Qi, Z. D.; Qian, J. M.; Qureshi, K. N.; Raghavan, R.; Rahal-Callot, G.; Rancoita, P. G.; Rattaggi, M.; Raven, G.; Razis, P.; Read, K.; Ren, D.; Ren, Z.; Rescigno, M.; Reucroft, S.; Ricker, A.; Riemann, S.; Riemers, B. C.; Riles, K.; Rind, O.; Rizvi, H. A.; Ro, S.; Rodriguez, F. J.; Roe, B. P.; Röhner, M.; Romero, L.; Rosier-Lees, S.; Rosmalen, R.; Rosselet, Ph.; van Rossum, W.; Roth, S.; Rubbia, A.; Rubio, J. A.; Rykaczewski, H.; Sachwitz, M.; Salicio, J.; Salicio, J. M.; Sanders, G. S.; Santocchia, A.; Sarakinos, M. S.; Sartorelli, G.; Sassowsky, M.; Sauvage, G.; Schegelsky, V.; Schmitz, D.; Schmitz, P.; Schneegans, M.; Schopper, H.; Schotanus, D. J.; Shotkin, S.; Schreiber, H. J.; Shukla, J.; Schulte, R.; Schulte, S.; Schultze, K.; Schwenke, J.; Schwering, G.; Sciacca, C.; Scott, I.; Sehgal, R.; Seiler, P. G.; Sens, J. C.; Servoli, L.; Sheer, I.; Shen, D. Z.; Shevchenko, S.; Shi, X. R.; Shumilov, E.; Shoutko, V.; Son, D.; Sopczak, A.; Soulimov, V.; Spartiotis, C.; Spickermann, T.; Spillantini, P.; Starosta, R.; Steuer, M.; Stickland, D. P.; Sticozzi, F.; Stone, H.; Strauch, K.; Stringfellow, B. C.; Sudhakar, K.; Sultanov, G.; Sun, L. Z.; Susinno, G. F.; Suter, H.; Swain, J. D.; Syed, A. A.; Tang, X. W.; Taylor, L.; Terzi, G.; Ting, Samuel C. C.; Ting, S. M.; Tonutti, M.; Tonwar, S. C.; Tóth, J.; Tsaregorodtsev, A.; Tsipolitis, G.; Tully, C.; Tung, K. L.; Ulbricht, J.; Urbán, L.; Uwer, U.; Valente, E.; Van de Walle, R. T.; Vetlitsky, I.; Viertel, G.; Vikas, P.; Vikas, U.; Vivargent, M.; Vogel, H.; Vogt, H.; Vorobiev, I.; Vorobyov, A. A.; Vuilleumier, L.; Wadhwa, M.; Wallraff, W.; Wang, C.; Wang, C. R.; Wang, X. L.; Wang, Y. F.; Wang, Z. M.; Warner, C.; Weber, A.; Weber, J.; Weill, R.; Wenaus, T. J.; Wenninger, J.; White, M.; Willmott, C.; Wittgenstein, F.; Wright, D.; Wu, S. X.; Wynhoff, S.; Wysłouch, B.; Xie, Y. Y.; Xu, J. G.; Xu, Z. Z.; Xue, Z. L.; Yan, D. S.; Yang, B. Z.; Yang, C. G.; Yang, G.; Ye, C. H.; Ye, J. B.; Ye, Q.; Yeh, S. C.; Yin, Z. W.; You, J. M.; Yunus, N.; Yzerman, M.; Zaccardelli, C.; Zaitsev, N.; Zemp, P.; Zeng, M.; Zeng, Y.; Zhang, D. H.; Zhang, Z. P.; Zhou, B.; Zhou, G. J.; Zhou, J. F.; Zhu, R. Y.; Zichichi, A.; van der Zwaan, B. C. C.; L3 Collaboration
1993-11-01
The average lifetime of b hadrons has been measured using the L3 detector at LEP, running at √ s ≈ MZ. A b-enriched sample was obtained from 432538 hadronic Z events collected in 1990 and 1991 by tagging electrons and muons from semileptonic b hadron decays. From maximum likelihood fits to the electron and muon impact parameter distributions, the average b hadron lifetime was measured to be τb = (1535 ± 35 ± 28) fs, where the first error is statistical and the second includes both the experimental and the theoretical systematic uncertainties.
2010-01-01
... 7 Agriculture 10 2010-01-01 2010-01-01 false On average. 1209.12 Section 1209.12 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (MARKETING AGREEMENTS....12 On average. On average means a rolling average of production or imports during the last two...
Ejiri, A.
The second questionnaire for scientists and engineers was carried out in 2007, and status of Japanese scientists and engineers were analyzed and reported. A part of the data was reanalyzed from the viewpoint of work life balance. In particular, office/laboratory staying hour and home working hour were analyzed and dependences on various factors were investigated. It was found that these hours depend on gender, marital status, number of child, employment status and age. In addition, the total hours tend to be kept constant regardless of various factors.
Installation Service - Changes in opening hours
GS Department
2010-01-01
For organizational matters, please note that, as from 15 March 2010, the Installation Service will have wen opening hours. The new schedule will be from 14:00 to 17:00 (Monday to Friday). Contact persons are: Martine Briant, Karine Robert and Claudia Bruggmann. The office address remains 73-3-014. Installation Service
Why Surface Nanobubbles Live for Hours
Weijs, Joost; Lohse, Detlef
2013-01-01
We present a theoretical model for the experimentally found but counterintuitive exceptionally long lifetime of surface nanobubbles. We can explain why, under normal experimental conditions, surface nanobubbles are stable for many hours or even up to days rather than the expected microseconds. The
2010-01-01
..., training, or pre-shift briefings for special evolutions; and holdovers for interviews needed for event... waivers under which work was performed to evaluate staffing adequacy for all jobs subject to the work hour...; and (4) Record, trend, and correct, under the licensee's corrective action program, any problems...
Electoral (Amendment) (Hours of Polling) Bill 2013
Butler, Graham; Hunt, Brian; Doyle, Andrew
2013-01-01
Private Members' Bill (legislation) introduced in Dáil Éireann (House of Deputies), Houses of the Oireachtas (Parliament of Ireland). An Act to set voting hours for Dáil Elections, Dáil Bye-Elections, Presidential Elections, European Parliament Elections, Local Government Elections and Referenda...
The 24-Hour Mathematical Modeling Challenge
Galluzzo, Benjamin J.; Wendt, Theodore J.
2015-01-01
Across the mathematics curriculum there is a renewed emphasis on applications of mathematics and on mathematical modeling. Providing students with modeling experiences beyond the ordinary classroom setting remains a challenge, however. In this article, we describe the 24-hour Mathematical Modeling Challenge, an extracurricular event that exposes…
Analysis of The Story of An Hour
曹玉琴
2014-01-01
By reading the short story The Story of An Hour written by Kate Chopin, we tried to probe the story better by analyz-ing five characters, plots of the story and rhetorical devices, such as irony and symbolism. With originality and profound mean-ings, the novel about the small theme of family life had a wide variety of social significance.
Estimator's electrical man-hour manual
Page, John S
1999-01-01
This manual's latest edition continues to be the best source available for making accurate, reliable man-hour estimates for electrical installation. This new edition is revised and expanded to include installation of electrical instrumentation, which is used in monitoring various process systems.
Thermodynamic hardness and the maximum hardness principle
Franco-Pérez, Marco; Gázquez, José L.; Ayers, Paul W.; Vela, Alberto
2017-08-01
An alternative definition of hardness (called the thermodynamic hardness) within the grand canonical ensemble formalism is proposed in terms of the partial derivative of the electronic chemical potential with respect to the thermodynamic chemical potential of the reservoir, keeping the temperature and the external potential constant. This temperature dependent definition may be interpreted as a measure of the propensity of a system to go through a charge transfer process when it interacts with other species, and thus it keeps the philosophy of the original definition. When the derivative is expressed in terms of the three-state ensemble model, in the regime of low temperatures and up to temperatures of chemical interest, one finds that for zero fractional charge, the thermodynamic hardness is proportional to T-1(I -A ) , where I is the first ionization potential, A is the electron affinity, and T is the temperature. However, the thermodynamic hardness is nearly zero when the fractional charge is different from zero. Thus, through the present definition, one avoids the presence of the Dirac delta function. We show that the chemical hardness defined in this way provides meaningful and discernible information about the hardness properties of a chemical species exhibiting integer or a fractional average number of electrons, and this analysis allowed us to establish a link between the maximum possible value of the hardness here defined, with the minimum softness principle, showing that both principles are related to minimum fractional charge and maximum stability conditions.
Maximum Likelihood Analysis in the PEN Experiment
Lehman, Martin
2013-10-01
The experimental determination of the π+ -->e+ ν (γ) decay branching ratio currently provides the most accurate test of lepton universality. The PEN experiment at PSI, Switzerland, aims to improve the present world average experimental precision of 3 . 3 ×10-3 to 5 ×10-4 using a stopped beam approach. During runs in 2008-10, PEN has acquired over 2 ×107 πe 2 events. The experiment includes active beam detectors (degrader, mini TPC, target), central MWPC tracking with plastic scintillator hodoscopes, and a spherical pure CsI electromagnetic shower calorimeter. The final branching ratio will be calculated using a maximum likelihood analysis. This analysis assigns each event a probability for 5 processes (π+ -->e+ ν , π+ -->μ+ ν , decay-in-flight, pile-up, and hadronic events) using Monte Carlo verified probability distribution functions of our observables (energies, times, etc). A progress report on the PEN maximum likelihood analysis will be presented. Work supported by NSF grant PHY-0970013.
Cacti with maximum Kirchhoff index
Wang, Wen-Rui; Pan, Xiang-Feng
2015-01-01
The concept of resistance distance was first proposed by Klein and Randi\\'c. The Kirchhoff index $Kf(G)$ of a graph $G$ is the sum of resistance distance between all pairs of vertices in $G$. A connected graph $G$ is called a cactus if each block of $G$ is either an edge or a cycle. Let $Cat(n;t)$ be the set of connected cacti possessing $n$ vertices and $t$ cycles, where $0\\leq t \\leq \\lfloor\\frac{n-1}{2}\\rfloor$. In this paper, the maximum kirchhoff index of cacti are characterized, as well...
Generic maximum likely scale selection
Pedersen, Kim Steenstrup; Loog, Marco; Markussen, Bo
2007-01-01
The fundamental problem of local scale selection is addressed by means of a novel principle, which is based on maximum likelihood estimation. The principle is generally applicable to a broad variety of image models and descriptors, and provides a generic scale estimation methodology. The focus...... on second order moments of multiple measurements outputs at a fixed location. These measurements, which reflect local image structure, consist in the cases considered here of Gaussian derivatives taken at several scales and/or having different derivative orders....
Hourly associations between heat and ambulance calls.
Guo, Yuming
2017-01-01
The response speed of ambulance calls is very crucial to rescue patients suffering immediately life threatening conditions. The serious health outcomes might be caused by exposing to extreme heat only several hours before. However, limited evidence is available on this topic. This study aims to examine the hourly association between heat and ambulance calls, to improve the ambulance services and to better protect health. Hourly data on ambulance calls for non-accidental causes, temperature and air pollutants (PM10, NO2, and O3) were collected from Brisbane, Australia, during 2001 and 2007. A time-stratified case-crossover design was used to examine the associations between hourly ambulance calls and temperature during warm season (Nov, Dec, Jan, Feb, and Mar), while adjusting for potential confounders. Stratified analyses were performed for sex and age groups. Ambulance calls peaked at 10am for all groups, except those aged heat-ambulance calls relationships were non-linear for all groups, with thresholds between 27 °C and 31 °C. The associations appeared immediately, and lasted for about 24 h. There were no significant modification effect by sex and age. The findings suggest that hot hourly temperatures (>27 °C) increase the demands of ambulance. This information is helpful to increase the efficiency of ambulance service then save lives, for example, preparing more ambulance before appearance of extremely hot temperature in combination with weather forecast. Also, people should better arrange their time for outdoor activities to avoid exposing to extreme hot temperatures. Copyright Â© 2016 Elsevier Ltd. All rights reserved.
Bounds on Average Time Complexity of Decision Trees
Chikalov, Igor
2011-01-01
In this chapter, bounds on the average depth and the average weighted depth of decision trees are considered. Similar problems are studied in search theory [1], coding theory [77], design and analysis of algorithms (e.g., sorting) [38]. For any diagnostic problem, the minimum average depth of decision tree is bounded from below by the entropy of probability distribution (with a multiplier 1/log2 k for a problem over a k-valued information system). Among diagnostic problems, the problems with a complete set of attributes have the lowest minimum average depth of decision trees (e.g, the problem of building optimal prefix code [1] and a blood test study in assumption that exactly one patient is ill [23]). For such problems, the minimum average depth of decision tree exceeds the lower bound by at most one. The minimum average depth reaches the maximum on the problems in which each attribute is "indispensable" [44] (e.g., a diagnostic problem with n attributes and kn pairwise different rows in the decision table and the problem of implementing the modulo 2 summation function). These problems have the minimum average depth of decision tree equal to the number of attributes in the problem description. © Springer-Verlag Berlin Heidelberg 2011.
Economics and Maximum Entropy Production
Lorenz, R. D.
2003-04-01
Price differentials, sales volume and profit can be seen as analogues of temperature difference, heat flow and work or entropy production in the climate system. One aspect in which economic systems exhibit more clarity than the climate is that the empirical and/or statistical mechanical tendency for systems to seek a maximum in production is very evident in economics, in that the profit motive is very clear. Noting the common link between 1/f noise, power laws and Self-Organized Criticality with Maximum Entropy Production, the power law fluctuations in security and commodity prices is not inconsistent with the analogy. There is an additional thermodynamic analogy, in that scarcity is valued. A commodity concentrated among a few traders is valued highly by the many who do not have it. The market therefore encourages via prices the spreading of those goods among a wider group, just as heat tends to diffuse, increasing entropy. I explore some empirical price-volume relationships of metals and meteorites in this context.
Effect of intraoperatory peritoneal bupivacaine irrigation on inmediate, 24 hour
Navarro Hernando
2004-09-01
Full Text Available Objective: To compare the response on inmediate, 24 hour abdominal pain and referred shoulder pain (omalgia, when an intraperitoneal irrigation of bupivacaine was used as opposed to saline irrigation, in a group of patients who has laparoscopy done in a medical center in Cali, between April and June of 2000. Material and methods: After previous authorization and consentment by all 100 patients, a placebo was applied to 50 patients and bupivacaine to the other 50 patients, without randomization. The drug came in unmarked bottles, some of which had the medication and some had only saline and nobody knew beforehand which one was given to the patient. Its effect on post surgical pain was evaluated throught the cromatic visual analog scale, which subjectively determines the intensity of the pain according to colors which range from white to dark red, with a numerical equivalent from 0 to 10, in the following manner: zero (0 no pain , 2 mild, 4 moderate, 6 strong, 8 sevre and 10 intolerable pain. The scale was measured in the postoperative room two hours after procedure. A telephonic follow up was done at 24 hours after surgery, applying the visual analog scale on abdominal and shoulder pain. The information was processed using Epi Info 2000. An univariante and bivariante analysis was done. A comparasion was established using the differences in averages with inmediate abdominal pain and omalgia, and at 24 hours according to the use or not bupivacaine. Results: The variable are described in a univariante manner in relation to the distribution of pain at 2 and 24 hours, as well as the shoulder pain, where the mean score of pain decreases from 3.8, to 2.1 to 1.1 respectively. The variables were also described in a bivariant way comparing the drug and its effect on the intensity of pain at 2 and 24 hours, and the omalgia, where no statistically significant difference was found between the drug, the saline and the scales of pain. The relationship between
Santos-Silva, Paulo Roberto; Fonseca, Alfredo José; Castro, Anita Weigand de; Greve, Júlia Maria D'Andréa; Hernandez, Arnaldo José
2007-08-01
To determine the degree of reproducibility of maximum oxygen consumption (VO2max) among soccer players, using a modified Heck protocol. 2 evaluations with an interval of 15 days between them were performed on 11 male soccer players. All the players were at a high performance level; they were training for an average of 10 hours per week, totaling 5 times a week. When they were evaluated, they were in the middle of the competitive season, playing 1 match per week. The soccer players were evaluated on an ergometric treadmill with velocity increments of 1.2 km.h-1 every 2 minutes and a fixed inclination of 3% during the test. VO2max was measured directly using a breath-by-breath metabolic gas analyzer. The maximum running speed and VO2max attained in the 2 tests were, respectively: (15.6 +/- 1.1 vs. 15.7 +/- 1.2 km.h-1; [P = .78]) and (54.5 +/- 3.9 vs. 55.2 +/- 4.4 ml.kg-1.min-1; [P = .88]). There was high and significant correlation of VO2max between the 2 tests with a 15-day interval between them [r = 0.97; P testing was insufficient to significantly modify the soccer players' VO2max values.
Stability, precision, and near-24-hour period of the human circadian pacemaker
Czeisler, C. A.; Duffy, J. F.; Shanahan, T. L.; Brown, E. N.; Mitchell, J. F.; Rimmer, D. W.; Ronda, J. M.; Silva, E. J.; Allan, J. S.; Emens, J. S.;
1999-01-01
Regulation of circadian period in humans was thought to differ from that of other species, with the period of the activity rhythm reported to range from 13 to 65 hours (median 25.2 hours) and the period of the body temperature rhythm reported to average 25 hours in adulthood, and to shorten with age. However, those observations were based on studies of humans exposed to light levels sufficient to confound circadian period estimation. Precise estimation of the periods of the endogenous circadian rhythms of melatonin, core body temperature, and cortisol in healthy young and older individuals living in carefully controlled lighting conditions has now revealed that the intrinsic period of the human circadian pacemaker averages 24.18 hours in both age groups, with a tight distribution consistent with other species. These findings have important implications for understanding the pathophysiology of disrupted sleep in older people.
On the errors associated with the use of hourly data in wind-driven rain calculations on building facades
Blocken, Bert; Carmeliet, Jan
Wind-driven rain (WDR) is the one of the main moisture sources for building facades. It is an important factor in the dry and wet deposition of pollutants, facade surface soiling and facade erosion. WDR calculations require data records of wind speed, wind direction and horizontal rainfall intensity as input. Most meteorological datasets contain at best arithmetically averaged hourly wind and rain data. Their use is common practice in WDR calculations. As an example, existing WDR standards request at best hourly data. This paper however demonstrates that the use of such data can yield (very) large errors in the calculated WDR amounts and intensities. The reason is that arithmetic averaging on an hourly basis generally causes an important loss of information about the co-occurrence of wind and rain. An improved data averaging technique for wind and rain data is proposed that respects this co-occurrence by applying appropriate weighting factors in the averaging procedure. The performance of this technique is evaluated by WDR calculations on buildings in three cities with different climates. While arithmetically averaged hourly data yield large underestimation errors (Eindhoven, The Netherlands: 11%, Bloomington, USA: 45%, Grahamstown, South Africa: 31%), the improved averaging technique provides very good results (errors: 0%, 4%, 3%, respectively). In conclusion, WDR calculations should not be performed with arithmetically averaged hourly data. Instead, either high-resolution data (e.g. 10-min data) or hourly data that have been obtained with the proposed weighted averaging technique should be used.
Objects of maximum electromagnetic chirality
Fernandez-Corbaton, Ivan
2015-01-01
We introduce a definition of the electromagnetic chirality of an object and show that it has an upper bound. The upper bound is attained if and only if the object is transparent for fields of one handedness (helicity). Additionally, electromagnetic duality symmetry, i.e. helicity preservation upon scattering, turns out to be a necessary condition for reciprocal scatterers to attain the upper bound. We use these results to provide requirements for the design of such extremal scatterers. The requirements can be formulated as constraints on the polarizability tensors for dipolar scatterers or as material constitutive relations. We also outline two applications for objects of maximum electromagnetic chirality: A twofold resonantly enhanced and background free circular dichroism measurement setup, and angle independent helicity filtering glasses.
Maximum mutual information regularized classification
Wang, Jim Jing-Yan
2014-09-07
In this paper, a novel pattern classification approach is proposed by regularizing the classifier learning to maximize mutual information between the classification response and the true class label. We argue that, with the learned classifier, the uncertainty of the true class label of a data sample should be reduced by knowing its classification response as much as possible. The reduced uncertainty is measured by the mutual information between the classification response and the true class label. To this end, when learning a linear classifier, we propose to maximize the mutual information between classification responses and true class labels of training samples, besides minimizing the classification error and reducing the classifier complexity. An objective function is constructed by modeling mutual information with entropy estimation, and it is optimized by a gradient descend method in an iterative algorithm. Experiments on two real world pattern classification problems show the significant improvements achieved by maximum mutual information regularization.
Why surface nanobubbles live for hours
Weijs, Joost H
2012-01-01
We present a theoretical model for the experimentally found but counter-intuitive exceptionally long lifetime of surface nanobubbles. We can explain why, under normal experimental conditions, surface nanobubbles are stable for many hours or even up to days rather than the expected microseconds. The limited gas diffusion through the water in the far field, the cooperative effect of nanobubble clusters, and the pinned contact line of the nanobubbles lead to the slow dissolution rate.
Optimum Arrangement of Taxi Drivers’ Working Hours
TANIZAKI, Takashi
2014-01-01
Part 2: Knowledge Discovery and Sharing; International audience; We propose optimum arrangement of taxi drivers’ working hours. In Japan, income of taxi vehicle is decreasing about 11 thousand yen in the past 15 years. Then some taxi companies are investing to gain more customers. But there are many small taxi companies that are difficult to invest with much money. Therefore we have been researching the other method to gain more customers by little investment for small companies. In this pape...
Hourly Spectral Variability of Mrk 421
Krennrich, F; Boyle, P J; Bradbury, S M; Buckley, J H
2003-01-01
Mrk 421 is the first TeV blazar found to exhibit significant spectral variability during strong flaring activity, showing hardening of the TeV spectrum in high emission states. Mrk 421 is also known to exhibit flux variability on time scales as short as 15 minutes. In this paper we present studies of hourly spectral variability of Mrk 421 in 2001 using data from the Whipple Observatory 10m gamma-ray telescope.
The strong maximum principle revisited
Pucci, Patrizia; Serrin, James
In this paper we first present the classical maximum principle due to E. Hopf, together with an extended commentary and discussion of Hopf's paper. We emphasize the comparison technique invented by Hopf to prove this principle, which has since become a main mathematical tool for the study of second order elliptic partial differential equations and has generated an enormous number of important applications. While Hopf's principle is generally understood to apply to linear equations, it is in fact also crucial in nonlinear theories, such as those under consideration here. In particular, we shall treat and discuss recent generalizations of the strong maximum principle, and also the compact support principle, for the case of singular quasilinear elliptic differential inequalities, under generally weak assumptions on the quasilinear operators and the nonlinearities involved. Our principal interest is in necessary and sufficient conditions for the validity of both principles; in exposing and simplifying earlier proofs of corresponding results; and in extending the conclusions to wider classes of singular operators than previously considered. The results have unexpected ramifications for other problems, as will develop from the exposition, e.g. two point boundary value problems for singular quasilinear ordinary differential equations (Sections 3 and 4); the exterior Dirichlet boundary value problem (Section 5); the existence of dead cores and compact support solutions, i.e. dead cores at infinity (Section 7); Euler-Lagrange inequalities on a Riemannian manifold (Section 9); comparison and uniqueness theorems for solutions of singular quasilinear differential inequalities (Section 10). The case of p-regular elliptic inequalities is briefly considered in Section 11.
Anaïs Vernède
2011-01-01
From 8.30 to 9.30 p.m. on Saturday, 26 March 2011, the Globe of Science and Innovation will be plunged into darkness to mark CERN's participation in Earth Hour. A growing number of countries and cities across the planet are involved in this global initiative against climate change, which was launched by the WWF in 2007. The lights on the Globe were switched off for the 2009 Earth Hour event. Along with individuals, companies and tourist attractions in thousands of towns and cities all over the world participating in the fourth annual Earth Hour event, CERN will turn off the lights of the Globe for 60 minutes at 8.30 p.m. on Saturday 26 March. CERN's participation in the initiative is one of several examples of its commitment to respect the environment and keep its ecological footprint to the minimum. A recent example under the green transport heading was the replacement of part of CERN's petrol vehicle fleet with cars running on natural gas with a view to reducing air pollution. Other examples...
Kamal, Lalarukh [Balochistan Univ., Dept. of Mathematics, Quetta (Pakistan); Jafri, Yasmin Zahra [Balochistan Univ., Dept. of Statistics, Quetta (Pakistan)
1999-07-01
Using hourly global radiation data at Quetta, Pakistan for 10 yr, an Autoregressive Moving Average (ARMA) process is fitted. Markov Transition Matrices have also been developed. These models are used for generating synthetic sequences for hourly radiations in MJ/m{sup 2} and that the generated sequences are compared with the observed data. We found the MTM approach relatively better as a simulator compared to ARMA modeling. (Author)
Level sets of multiple ergodic averages
Ai-Hua, Fan; Ma, Ji-Hua
2011-01-01
We propose to study multiple ergodic averages from multifractal analysis point of view. In some special cases in the symbolic dynamics, Hausdorff dimensions of the level sets of multiple ergodic average limit are determined by using Riesz products.
29 CFR 778.411 - Sixty-hour limit on pay guaranteed by contract.
2010-07-01
... LABOR STATEMENTS OF GENERAL POLICY OR INTERPRETATION NOT DIRECTLY RELATED TO REGULATIONS OVERTIME COMPENSATION Exceptions From the Regular Rate Principles Guaranteed Compensation Which Includes Overtime Pay... exceed compensation due at the specified regular rate for the applicable maximum hours standard and...
Accurate Switched-Voltage voltage averaging circuit
金光, 一幸; 松本, 寛樹
2006-01-01
Abstract ###This paper proposes an accurate Switched-Voltage (SV) voltage averaging circuit. It is presented ###to compensated for NMOS missmatch error at MOS differential type voltage averaging circuit. ###The proposed circuit consists of a voltage averaging and a SV sample/hold (S/H) circuit. It can ###operate using nonoverlapping three phase clocks. Performance of this circuit is verified by PSpice ###simulations.
Spectral averaging techniques for Jacobi matrices
del Rio, Rafael; Schulz-Baldes, Hermann
2008-01-01
Spectral averaging techniques for one-dimensional discrete Schroedinger operators are revisited and extended. In particular, simultaneous averaging over several parameters is discussed. Special focus is put on proving lower bounds on the density of the averaged spectral measures. These Wegner type estimates are used to analyze stability properties for the spectral types of Jacobi matrices under local perturbations.
无
2011-01-01
[Objective] The research aimed to analyze temporal and spatial variation characteristics of temperature in Shangqiu City during 1961-2010.[Method] Based on temperature data in eight meteorological stations of Shangqiu during 1961-2010,by using trend analysis method,the temporal and spatial evolution characteristics of annual average temperature,annual average maximum and minimum temperatures,annual extreme maximum and minimum temperatures,daily range of annual average temperature in Shangqiu City were analy...
Impact of duty hours restrictions on quality of care and clinical outcomes.
Bhavsar, Jignesh; Montgomery, Daniel; Li, Jin; Kline-Rogers, Eva; Saab, Fadi; Motivala, Apurva; Froehlich, James B; Parekh, Vikas; Del Valle, John; Eagle, Kim A
2007-11-01
In July 2003, the Accreditation Council for Graduate Medical Education instituted residency duty-hours requirements in response to growing concerns regarding clinician fatigue and the incidence of medical errors. These changes, which limited maximum continuous hours worked and total hours per week, often resulted in increased discontinuity of care. The objective of this study was to assess the impact of the duty-hours restrictions on quality of care and outcomes of patients with acute coronary syndrome. We performed a retrospective analysis of 1003 consecutive patients with acute coronary syndrome admitted to the University of Michigan Hospital between July 2002 and June 2004. Patients were stratified by hospital admission during academic year 2002-2003 (pre-duty-hours changes, n=572) and academic year 2003-2004 (post-duty-hours changes, n=431). Main outcome measures included differences in adherence to quality indicators, length of stay, and in-hospital and 6-month adverse events. Post-duty-hours changes, there was an increase in the usage of beta-blockers (85.8% vs 93.8%, P Accreditation Council for Graduate Medical Education residency duty-hours restrictions on an academic inpatient cardiology service was associated with improved quality of care and efficiency in patients admitted with acute coronary syndrome. In addition, improved efficiency did not adversely impact patient outcomes, including mortality.
Analysis on the Variation Characteristics of Sunshine Hours in Puyang City
无
2011-01-01
[Objective] The research aimed to analyze the variation characteristics of sunshine hours in Puyang City in recent 51 years.[Method] Based on the monthly sunshine hours in five meteorological stations of Puyang City during 1960-2010,the annual,seasonal and monthly variation trends of sunshine hours,mutation time node were analyzed by using the climate trend coefficient,climate tendency rate,sliding average method,SNR and so on.The objective fact and rule of sunshine variation in Puyang in recent 51 years we...
46 CFR 15.1111 - Work hours and rest periods.
2010-10-01
... 46 Shipping 1 2010-10-01 2010-10-01 false Work hours and rest periods. 15.1111 Section 15.1111... REQUIREMENTS Vessels Subject to Requirements of STCW § 15.1111 Work hours and rest periods. (a) Each person... shall receive a minimum of 10 hours of rest in any 24-hour period. (b) The hours of rest required under...
29 CFR 2530.200b-2 - Hour of service.
2010-07-01
... 80 hours of service for his vacation in computation period III (40 hours per week multiplied by 2... would be credited with 80 hours (2 weeks worked multiplied by 40 hours) plus such credit as may be... a payment made or due which is calculated on the basis of units of time, such as hours, days,...
Average-Time Games on Timed Automata
Jurdzinski, Marcin; Trivedi, Ashutosh
2009-01-01
An average-time game is played on the infinite graph of configurations of a finite timed automaton. The two players, Min and Max, construct an infinite run of the automaton by taking turns to perform a timed transition. Player Min wants to minimise the average time per transition and player Max wants to maximise it. A solution of average-time games is presented using a reduction to average-price game on a finite graph. A direct consequence is an elementary proof of determinacy for average-tim...
Concentration fluctuations and averaging time in vapor clouds
Wilson, David J
2010-01-01
This book contributes to more reliable and realistic predictions by focusing on sampling times from a few seconds to a few hours. Its objectives include developing clear definitions of statistical terms, such as plume sampling time, concentration averaging time, receptor exposure time, and other terms often confused with each other or incorrectly specified in hazard assessments; identifying and quantifying situations for which there is no adequate knowledge to predict concentration fluctuations in the near-field, close to sources, and far downwind where dispersion is dominated by atmospheric t
Maximum entropy production in daisyworld
Maunu, Haley A.; Knuth, Kevin H.
2012-05-01
Daisyworld was first introduced in 1983 by Watson and Lovelock as a model that illustrates how life can influence a planet's climate. These models typically involve modeling a planetary surface on which black and white daisies can grow thus influencing the local surface albedo and therefore also the temperature distribution. Since then, variations of daisyworld have been applied to study problems ranging from ecological systems to global climate. Much of the interest in daisyworld models is due to the fact that they enable one to study self-regulating systems. These models are nonlinear, and as such they exhibit sensitive dependence on initial conditions, and depending on the specifics of the model they can also exhibit feedback loops, oscillations, and chaotic behavior. Many daisyworld models are thermodynamic in nature in that they rely on heat flux and temperature gradients. However, what is not well-known is whether, or even why, a daisyworld model might settle into a maximum entropy production (MEP) state. With the aim to better understand these systems, this paper will discuss what is known about the role of MEP in daisyworld models.
Maximum stellar iron core mass
F W Giacobbe
2003-03-01
An analytical method of estimating the mass of a stellar iron core, just prior to core collapse, is described in this paper. The method employed depends, in part, upon an estimate of the true relativistic mass increase experienced by electrons within a highly compressed iron core, just prior to core collapse, and is signiﬁcantly different from a more typical Chandrasekhar mass limit approach. This technique produced a maximum stellar iron core mass value of 2.69 × 1030 kg (1.35 solar masses). This mass value is very near to the typical mass values found for neutron stars in a recent survey of actual neutron star masses. Although slightly lower and higher neutron star masses may also be found, lower mass neutron stars are believed to be formed as a result of enhanced iron core compression due to the weight of non-ferrous matter overlying the iron cores within large stars. And, higher mass neutron stars are likely to be formed as a result of fallback or accretion of additional matter after an initial collapse event involving an iron core having a mass no greater than 2.69 × 1030 kg.
Maximum Matchings via Glauber Dynamics
Jindal, Anant; Pal, Manjish
2011-01-01
In this paper we study the classic problem of computing a maximum cardinality matching in general graphs $G = (V, E)$. The best known algorithm for this problem till date runs in $O(m \\sqrt{n})$ time due to Micali and Vazirani \\cite{MV80}. Even for general bipartite graphs this is the best known running time (the algorithm of Karp and Hopcroft \\cite{HK73} also achieves this bound). For regular bipartite graphs one can achieve an $O(m)$ time algorithm which, following a series of papers, has been recently improved to $O(n \\log n)$ by Goel, Kapralov and Khanna (STOC 2010) \\cite{GKK10}. In this paper we present a randomized algorithm based on the Markov Chain Monte Carlo paradigm which runs in $O(m \\log^2 n)$ time, thereby obtaining a significant improvement over \\cite{MV80}. We use a Markov chain similar to the \\emph{hard-core model} for Glauber Dynamics with \\emph{fugacity} parameter $\\lambda$, which is used to sample independent sets in a graph from the Gibbs Distribution \\cite{V99}, to design a faster algori...
2011-01-10
...: Establishing Maximum Allowable Operating Pressure or Maximum Operating Pressure Using Record Evidence, and... facilities of their responsibilities, under Federal integrity management (IM) regulations, to perform... system, especially when calculating Maximum Allowable Operating Pressure (MAOP) or Maximum Operating...
Ethylene Oxide: Acute Four-Hour and One-Hour Inhalation Toxicity Testing in Rats
William M. Snellings
2011-01-01
Full Text Available Ethylene oxide was tested on groups of rats for either 4-hour or 1-hour inhalation exposure, followed by 14 days of observation. Groups of five Sprague-Dawley rats/sex were exposed, and clinical signs and mortality were recorded. Clinical signs noted included irregular breathing, absence of certain reflexes, and tremors. Rats that died had moderate to severe pulmonary congestion. The calculated LC50 values, reported as ppm by volume (with 95% confidence limits, were as follows. 4-hour LC50 values were 1972 (1887 to 2061 ppm for males; 1537 (1391 to 1698 ppm for females; 1741 (1655 to 1831 ppm for the combined sexes. The 1-hour LC50 values were 5748 (5276 to 6262 ppm for males; 4439 (4034 to 4884 ppm for females; 5029 (4634 to 5459 ppm for the combined sexes.
New opening hours of the gates
GS Department
2009-01-01
Please note the new opening hours of the gates as well as the intersites tunnel from the 19 May 2009: GATE A 7h - 19h GATE B 24h/24 GATE C 7h - 9h\t17h - 19h GATE D 8h - 12h\t13h - 16h GATE E 7h - 9h\t17h - 19h Prévessin 24h/24 The intersites tunnel will be opened from 7h30 to 18h non stop. GS-SEM Group Infrastructure and General Services Department
Gate A: changes to opening hours
2015-01-01
Due to maintenance work, the opening hours of Gate A (near Reception) will be modified between Monday, 13 and Friday, 17 April 2015. During this period, the gate will be open to vehicles between 7 a.m. and 9.30 a.m., then between 4.30 p.m. and 7 p.m. It will be completely closed to traffic between 9.30 a.m. and 4.30 p.m. Pedestrians and cyclists may continue to use the gate. We apologise for any inconvenience and thank you for your understanding.
Temporary new opening hours for Gate C
GS Department
2010-01-01
Please note the new temporary opening hours for the gate C as from 22 September 2010 until 29 October 2010 (working days): Morning: between 7:00 a.m. and 9:00 a.m. Lunch: between 12:00 and 2:00 p.m. Evening: between 5:00 pm and 7:00 p.m. Traffic flow will be permitted in both directions during this period. Please minimize your speed accordingly and respect all road signs. GS-SEM Group General Infrastructure Services Department
Electoral (Amendment) (Hours of Polling) Bill 2013
Butler, Graham; Hunt, Brian; Doyle, Andrew
2013-01-01
Private Members' Bill (legislation) introduced in Dáil Éireann (House of Deputies), Houses of the Oireachtas (Parliament of Ireland). An Act to set voting hours for Dáil Elections, Dáil Bye-Elections, Presidential Elections, European Parliament Elections, Local Government Elections and Referenda...... as being 7.00 a.m. to 10.00 p.m. and for that purpose to amend the Electoral Act 1992, Presidential Elections Act 1993, Referendum Act 1994, European Parliament Elections Act 1997, the Local Government Act 2001 and to provide for related matters....
Plumley, George
2015-01-01
Create and expand feature-rich sites with no programming experience Ready to build, maintain, and expand your web site with WordPress but have no prior programming experience? WordPress 24-Hour Trainer, 3rd Edition is your book-and-video learning solution that walks you step-by-step through all the important features you will need to know. Lessons range from focused, practical everyday tasks to more advanced, creative features. Learn from an industry professional how to enter content, create pages, manage menus, utilize plug-ins, connect to social media, create membership and e-commerce site
The Sherpa Maximum Likelihood Estimator
Nguyen, D.; Doe, S.; Evans, I.; Hain, R.; Primini, F.
2011-07-01
A primary goal for the second release of the Chandra Source Catalog (CSC) is to include X-ray sources with as few as 5 photon counts detected in stacked observations of the same field, while maintaining acceptable detection efficiency and false source rates. Aggressive source detection methods will result in detection of many false positive source candidates. Candidate detections will then be sent to a new tool, the Maximum Likelihood Estimator (MLE), to evaluate the likelihood that a detection is a real source. MLE uses the Sherpa modeling and fitting engine to fit a model of a background and source to multiple overlapping candidate source regions. A background model is calculated by simultaneously fitting the observed photon flux in multiple background regions. This model is used to determine the quality of the fit statistic for a background-only hypothesis in the potential source region. The statistic for a background-plus-source hypothesis is calculated by adding a Gaussian source model convolved with the appropriate Chandra point spread function (PSF) and simultaneously fitting the observed photon flux in each observation in the stack. Since a candidate source may be located anywhere in the field of view of each stacked observation, a different PSF must be used for each observation because of the strong spatial dependence of the Chandra PSF. The likelihood of a valid source being detected is a function of the two statistics (for background alone, and for background-plus-source). The MLE tool is an extensible Python module with potential for use by the general Chandra user.
Vestige: Maximum likelihood phylogenetic footprinting
Maxwell Peter
2005-05-01
Full Text Available Abstract Background Phylogenetic footprinting is the identification of functional regions of DNA by their evolutionary conservation. This is achieved by comparing orthologous regions from multiple species and identifying the DNA regions that have diverged less than neutral DNA. Vestige is a phylogenetic footprinting package built on the PyEvolve toolkit that uses probabilistic molecular evolutionary modelling to represent aspects of sequence evolution, including the conventional divergence measure employed by other footprinting approaches. In addition to measuring the divergence, Vestige allows the expansion of the definition of a phylogenetic footprint to include variation in the distribution of any molecular evolutionary processes. This is achieved by displaying the distribution of model parameters that represent partitions of molecular evolutionary substitutions. Examination of the spatial incidence of these effects across regions of the genome can identify DNA segments that differ in the nature of the evolutionary process. Results Vestige was applied to a reference dataset of the SCL locus from four species and provided clear identification of the known conserved regions in this dataset. To demonstrate the flexibility to use diverse models of molecular evolution and dissect the nature of the evolutionary process Vestige was used to footprint the Ka/Ks ratio in primate BRCA1 with a codon model of evolution. Two regions of putative adaptive evolution were identified illustrating the ability of Vestige to represent the spatial distribution of distinct molecular evolutionary processes. Conclusion Vestige provides a flexible, open platform for phylogenetic footprinting. Underpinned by the PyEvolve toolkit, Vestige provides a framework for visualising the signatures of evolutionary processes across the genome of numerous organisms simultaneously. By exploiting the maximum-likelihood statistical framework, the complex interplay between mutational
Brokk Toggerson
2011-01-01
This year, the 48-hour film project (48hfp) returns to Geneva after a one-year hiatus. Organized by Neal Hartman and the CERN film-making club, Open Your Eyes Films, the 48hfp challenges teams of film-makers to write, shoot, soundtrack and edit a 4 to 7 minute film in 48 hours from 4 to 6 November. At the start of the festival, contestants picked their film genre from a hat. The films will be screened on 8 and 9 November, with the awards presentation on the 9th. The winner will receive a trip to the US to compete in the international version of the competition. “There are so many short films being made now," says Hartman, “I think, however, that the 48hfp allows a critical creative mass to form. The result is that these 20 teams make 20 better films than if each participant were making their own." Each team draws a genre from a hat and is given a character, a prop and a line of dialogue that must appear in their film. The genres run the gamut from &am...
Hybrid methodology for hourly global radiation forecasting in Mediterranean area
Voyant, Cyril; Paoli, Christophe; Nivet, Marie Laure
2012-01-01
The renewable energies prediction and particularly global radiation forecasting is a challenge studied by a growing number of research teams. This paper proposes an original technique to model the insolation time series based on combining Artificial Neural Network (ANN) and Auto-Regressive and Moving Average (ARMA) model. While ANN by its non-linear nature is effective to predict cloudy days, ARMA techniques are more dedicated to sunny days without cloud occurrences. Thus, three hybrids models are suggested: the first proposes simply to use ARMA for 6 months in spring and summer and to use an optimized ANN for the other part of the year; the second model is equivalent to the first but with a seasonal learning; the last model depends on the error occurred the previous hour. These models were used to forecast the hourly global radiation for five places in Mediterranean area. The forecasting performance was compared among several models: the 3 above mentioned models, the best ANN and ARMA for each location. In t...
Site Specific Probable Maximum Precipitation Estimates and Professional Judgement
Hayes, B. D.; Kao, S. C.; Kanney, J. F.; Quinlan, K. R.; DeNeale, S. T.
2015-12-01
State and federal regulatory authorities currently rely upon the US National Weather Service Hydrometeorological Reports (HMRs) to determine probable maximum precipitation (PMP) estimates (i.e., rainfall depths and durations) for estimating flooding hazards for relatively broad regions in the US. PMP estimates for the contributing watersheds upstream of vulnerable facilities are used to estimate riverine flooding hazards while site-specific estimates for small water sheds are appropriate for individual facilities such as nuclear power plants. The HMRs are often criticized due to their limitations on basin size, questionable applicability in regions affected by orographic effects, their lack of consist methods, and generally by their age. HMR-51 for generalized PMP estimates for the United States east of the 105th meridian, was published in 1978 and is sometimes perceived as overly conservative. The US Nuclear Regulatory Commission (NRC), is currently reviewing several flood hazard evaluation reports that rely on site specific PMP estimates that have been commercially developed. As such, NRC has recently investigated key areas of expert judgement via a generic audit and one in-depth site specific review as they relate to identifying and quantifying actual and potential storm moisture sources, determining storm transposition limits, and adjusting available moisture during storm transposition. Though much of the approach reviewed was considered a logical extension of HMRs, two key points of expert judgement stood out for further in-depth review. The first relates primarily to small storms and the use of a heuristic for storm representative dew point adjustment developed for the Electric Power Research Institute by North American Weather Consultants in 1993 in order to harmonize historic storms for which only 12 hour dew point data was available with more recent storms in a single database. The second issue relates to the use of climatological averages for spatially
WIDTHS AND AVERAGE WIDTHS OF SOBOLEV CLASSES
刘永平; 许贵桥
2003-01-01
This paper concerns the problem of the Kolmogorov n-width, the linear n-width, the Gel'fand n-width and the Bernstein n-width of Sobolev classes of the periodicmultivariate functions in the space Lp(Td) and the average Bernstein σ-width, averageKolmogorov σ-widths, the average linear σ-widths of Sobolev classes of the multivariatequantities.
Stochastic averaging of quasi-Hamiltonian systems
朱位秋
1996-01-01
A stochastic averaging method is proposed for quasi-Hamiltonian systems (Hamiltonian systems with light dampings subject to weakly stochastic excitations). Various versions of the method, depending on whether the associated Hamiltonian systems are integrable or nonintegrable, resonant or nonresonant, are discussed. It is pointed out that the standard stochastic averaging method and the stochastic averaging method of energy envelope are special cases of the stochastic averaging method of quasi-Hamiltonian systems and that the results obtained by this method for several examples prove its effectiveness.
NOAA Average Annual Salinity (3-Zone)
California Department of Resources — The 3-Zone Average Annual Salinity Digital Geography is a digital spatial framework developed using geographic information system (GIS) technology. These salinity...
PLANNING OF TRAINING AIRCRAFT FLIGHT HOURS
Visnja Vojvodić Rosenzweig
2010-12-01
Full Text Available Maintenance of the training aircraft causes downtime of operations and thereby reduces the operational availability, which is crucial for flight planning in a training organisation. Manual daily planning within the fleet delivers suboptimal results and often causes discontinued flight of several aircraft that have to be maintained at the same time. Optimal maintenance schedule of training aircraft can be obtained by a sliding scale method. This paper presents a mathematical model of the sliding scale formulated by a mixed integer linear problem. Allocation of flight hours is optimised by using AMPL programming language, assuring that a sufficient number of aircraft are always available for training. The model can be used by a flight dispatch department in a training organisation as a basis for optimised planning and reduction of maintenance downtime.
Dynamic Multiscale Averaging (DMA) of Turbulent Flow
Richard W. Johnson
2012-09-01
A new approach called dynamic multiscale averaging (DMA) for computing the effects of turbulent flow is described. The new method encompasses multiple applications of temporal and spatial averaging, that is, multiscale operations. Initially, a direct numerical simulation (DNS) is performed for a relatively short time; it is envisioned that this short time should be long enough to capture several fluctuating time periods of the smallest scales. The flow field variables are subject to running time averaging during the DNS. After the relatively short time, the time-averaged variables are volume averaged onto a coarser grid. Both time and volume averaging of the describing equations generate correlations in the averaged equations. These correlations are computed from the flow field and added as source terms to the computation on the next coarser mesh. They represent coupling between the two adjacent scales. Since they are computed directly from first principles, there is no modeling involved. However, there is approximation involved in the coupling correlations as the flow field has been computed for only a relatively short time. After the time and spatial averaging operations are applied at a given stage, new computations are performed on the next coarser mesh using a larger time step. The process continues until the coarsest scale needed is reached. New correlations are created for each averaging procedure. The number of averaging operations needed is expected to be problem dependent. The new DMA approach is applied to a relatively low Reynolds number flow in a square duct segment. Time-averaged stream-wise velocity and vorticity contours from the DMA approach appear to be very similar to a full DNS for a similar flow reported in the literature. Expected symmetry for the final results is produced for the DMA method. The results obtained indicate that DMA holds significant potential in being able to accurately compute turbulent flow without modeling for practical
Long working hours in Korea: results of the 2010 Working Conditions Survey.
Park, Jungsun; Kwon, Oh Jun; Kim, Yangho
2012-01-01
Long working hours adversely affect workers' safety and health. In 2004, Korea passed legislation limiting the working week to 40 h, to improve quality-of-life and to increase business competitiveness. In the present study, we explored the characteristics of work in Korea and compared our data of the second Korean Working Conditions Survey (KWCS) with those of the first KWCS. We found that the average number of hours worked weekly has been reduced but the proportions of workers who work for more than 48 h per week has increased over the 4 yr between the two Korean surveys in all categories studied (male, female, employee, self-employed, and employer). We also found that self-employed and employers work much longer hours than do employees, who are protected by the Labor Standards Act. This was particularly true in the accommodation and food service sectors. In conclusion, Korean workers work longer than do workers of EU countries. The use of average figures masks differences in the numbers of working hours among those engaged in various types of employment, or in certain work sectors. Therefore, the Korean government should not simply monitor reductions in average weekly working hours, but should identify employees working for over 60 h weekly, and reduce their working time.
Average Transmission Probability of a Random Stack
Lu, Yin; Miniatura, Christian; Englert, Berthold-Georg
2010-01-01
The transmission through a stack of identical slabs that are separated by gaps with random widths is usually treated by calculating the average of the logarithm of the transmission probability. We show how to calculate the average of the transmission probability itself with the aid of a recurrence relation and derive analytical upper and lower…
Average sampling theorems for shift invariant subspaces
无
2000-01-01
The sampling theorem is one of the most powerful results in signal analysis. In this paper, we study the average sampling on shift invariant subspaces, e.g. wavelet subspaces. We show that if a subspace satisfies certain conditions, then every function in the subspace is uniquely determined and can be reconstructed by its local averages near certain sampling points. Examples are given.
Testing linearity against nonlinear moving average models
de Gooijer, J.G.; Brännäs, K.; Teräsvirta, T.
1998-01-01
Lagrange multiplier (LM) test statistics are derived for testing a linear moving average model against an additive smooth transition moving average model. The latter model is introduced in the paper. The small sample performance of the proposed tests are evaluated in a Monte Carlo study and compared
Averaging Einstein's equations : The linearized case
Stoeger, William R.; Helmi, Amina; Torres, Diego F.
2007-01-01
We introduce a simple and straightforward averaging procedure, which is a generalization of one which is commonly used in electrodynamics, and show that it possesses all the characteristics we require for linearized averaging in general relativity and cosmology for weak-field and perturbed FLRW situ
Averaging Einstein's equations : The linearized case
Stoeger, William R.; Helmi, Amina; Torres, Diego F.
We introduce a simple and straightforward averaging procedure, which is a generalization of one which is commonly used in electrodynamics, and show that it possesses all the characteristics we require for linearized averaging in general relativity and cosmology for weak-field and perturbed FLRW
Average excitation potentials of air and aluminium
Bogaardt, M.; Koudijs, B.
1951-01-01
By means of a graphical method the average excitation potential I may be derived from experimental data. Average values for Iair and IAl have been obtained. It is shown that in representing range/energy relations by means of Bethe's well known formula, I has to be taken as a continuously changing fu
Average Transmission Probability of a Random Stack
Lu, Yin; Miniatura, Christian; Englert, Berthold-Georg
2010-01-01
The transmission through a stack of identical slabs that are separated by gaps with random widths is usually treated by calculating the average of the logarithm of the transmission probability. We show how to calculate the average of the transmission probability itself with the aid of a recurrence relation and derive analytical upper and lower…
2010-07-19
... CFR Part 3015, Subpart V, and the final rule related notice published at 48 FR 29114, June 24, 1983... Average Payments/Maximum Reimbursement Rates AGENCY: Food and Nutrition Service, USDA. ACTION: Notice. SUMMARY: This Notice announces the annual adjustments to the ``national average payments,'' the amount...
New results on averaging theory and applications
Cândido, Murilo R.; Llibre, Jaume
2016-08-01
The usual averaging theory reduces the computation of some periodic solutions of a system of ordinary differential equations, to find the simple zeros of an associated averaged function. When one of these zeros is not simple, i.e., the Jacobian of the averaged function in it is zero, the classical averaging theory does not provide information about the periodic solution associated to a non-simple zero. Here we provide sufficient conditions in order that the averaging theory can be applied also to non-simple zeros for studying their associated periodic solutions. Additionally, we do two applications of this new result for studying the zero-Hopf bifurcation in the Lorenz system and in the Fitzhugh-Nagumo system.
Analogue Divider by Averaging a Triangular Wave
Selvam, Krishnagiri Chinnathambi
2017-08-01
A new analogue divider circuit by averaging a triangular wave using operational amplifiers is explained in this paper. The triangle wave averaging analog divider using operational amplifiers is explained here. The reference triangular waveform is shifted from zero voltage level up towards positive power supply voltage level. Its positive portion is obtained by a positive rectifier and its average value is obtained by a low pass filter. The same triangular waveform is shifted from zero voltage level to down towards negative power supply voltage level. Its negative portion is obtained by a negative rectifier and its average value is obtained by another low pass filter. Both the averaged voltages are combined in a summing amplifier and the summed voltage is given to an op-amp as negative input. This op-amp is configured to work in a negative closed environment. The op-amp output is the divider output.
2010-07-01
... and Secondary National Ambient Air Quality Standards for Ozone I Appendix I to Part 50 Protection of... AIR QUALITY STANDARDS Pt. 50, App. I Appendix I to Part 50—Interpretation of the 8-Hour Primary and... valid day if valid 8-hour averages are available for at least 75% of possible hours in the day (i.e.,...
Picosecond mid-infrared amplifier for high average power.
Botha, LR
2007-04-01
Full Text Available are similar. The saturation fluence for a multi level system can be written as z PhEsat σ υ 2 = With σ the stimulated emission cross section and P the pressure of the laser. 1/z... is essentially the average number of populated rotational levels. For our case z=0.07 and 181054.1 −×=σ cm2. Thus for a 10 atm laser the saturation fluence is: 2 18 1334 /1173 07.01017/12 10109.210626.6 cmmJxEsat = ××× ××× = − − The maximum...
The Average-Case Area of Heilbronn-Type Triangles
Jiang, T.; Li, Ming; Vitányi, Paul
1999-01-01
From among $ {n \\choose 3}$ triangles with vertices chosen from $n$ points in the unit square, let $T$ be the one with the smallest area, and let $A$ be the area of $T$. Heilbronn's triangle problem asks for the maximum value assumed by $A$ over all choices of $n$ points. We consider the average-case: If the $n$ points are chosen independently and at random (with a uniform distribution), then there exist positive constants $c$ and $C$ such that $c/n^3 < \\mu_n < C/n^3$ for all large enough val...
Recent advances in phase shifted time averaging and stroboscopic interferometry
Styk, Adam; Józwik, Michał
2016-08-01
Classical Time Averaging and Stroboscopic Interferometry are widely used for MEMS/MOEMS dynamic behavior investigations. Unfortunately both methods require an extensive measurement and data processing strategies in order to evaluate the information on maximum amplitude at a given load of vibrating object. In this paper the modified strategies of data processing in both techniques are introduced. These modifications allow for fast and reliable calculation of searched value, without additional complication of measurement systems. Through the paper the both approaches are discussed and experimentally verified.
Eaton, Danice K; McKnight-Eily, Lela R; Lowry, Richard; Perry, Geraldine S; Presley-Cantrell, Letitia; Croft, Janet B
2010-04-01
We describe the prevalence of insufficient, borderline, and optimal sleep hours among U.S. high school students on an average school night. Most students (68.9%) reported insufficient sleep, whereas few (7.6%) reported optimal sleep. The prevalence of insufficient sleep was highest among female and black students, and students in grades 11 and 12. Published by Elsevier Inc.
Education on Adult Urinary Incontinence in Nursing School Curricula: Can It Be Done in Two Hours?
Morishita, Lynne; And Others
1994-01-01
Responses from 339 undergraduate nursing programs (74%) showed that 98% included urinary incontinence content in their curricula. Although most agreed the subject was important and felt their teaching was effective, the didactic component averaged two hours, and clinical experience was not systematic; few faculty are prepared to teach this…
Predicting Maximum Sunspot Number in Solar Cycle 24
Nipa J Bhatt; Rajmal Jain; Malini Aggarwal
2009-03-01
A few prediction methods have been developed based on the precursor technique which is found to be successful for forecasting the solar activity. Considering the geomagnetic activity aa indices during the descending phase of the preceding solar cycle as the precursor, we predict the maximum amplitude of annual mean sunspot number in cycle 24 to be 111 ± 21. This suggests that the maximum amplitude of the upcoming cycle 24 will be less than cycles 21–22. Further, we have estimated the annual mean geomagnetic activity aa index for the solar maximum year in cycle 24 to be 20.6 ± 4.7 and the average of the annual mean sunspot number during the descending phase of cycle 24 is estimated to be 48 ± 16.8.
Sunlight Exposure, Work Hours, Caffeine Consumption, and Sleep Duration in the Naval Environment.
Shattuck, Nita L; Matsangas, Panagiotis
2017-06-01
Sailors in the U.S. Navy are habitual shiftworkers, often experiencing circadian misalignment due to their irregular work/rest schedules. This study assessed the effect of sunlight exposure, work hours, and caffeinated beverage consumption on the daily sleep duration of crewmembers of a U.S. Navy ship during a 2-wk underway period. Working in an artificially lit area with no access to sunlight during work hours, U.S. Navy crew members (N = 91) used daily logs to report their daily activity, caffeinated beverage consumption, and exposure to sunlight while off-duty; sleep was assessed by wrist-worn actigraphy. Hours of sunlight exposure, work duration, and the amount of coffee/tea/soft drinks were statistically significant predictors of sleep duration. On average, crewmembers who reported more than one half-hour of sunlight each day slept on average ∼40 min (10%) less than their peers working the same shifts who received less than one half-hour of sunlight (on average 6.05 ± 0.90 h vs. 6.71 ± 0.91 h, respectively). Exposure to sunlight, work hours, and consumption of caffeinated beverages are important factors when planning watchstanding schedules at sea. Even though further research is needed, our results suggest that even brief exposure to sunlight may contribute to circadian misalignment that negatively affects sleep in the operational environment. Educating crewmembers about sleep hygiene, especially the important roles played by sunlight and caffeine, could potentially improve the sleep and fatigue levels of this population of maritime shiftworkers.Shattuck NL, Matsangas P. Sunlight exposure, work hours, caffeine consumption, and sleep duration in the naval environment. Aerosp Med Hum Perform. 2017; 88(6):579-585.
Dexter, Franklin; Epstein, Richard H; Dutton, Richard P; Kordylewski, Hubert; Ledolter, Johannes; Rosenberg, Henry; Hindman, Bradley J
2016-12-01
Anesthesiologists providing care during off hours (ie, weekends or holidays, or cases started during the evening or late afternoon) are more likely to care for patients at greater risk of sustaining major adverse events than when they work during regular hours (eg, Monday through Friday, from 7:00 AM to 2:59 PM). We consider the logical inconsistency of using subspecialty teams during regular hours but not during weekends or evenings. We analyzed data from the Anesthesia Quality Institute's National Anesthesia Clinical Outcomes Registry (NACOR). Among the hospitals in the United States, we estimated the average number of common types of anesthesia procedures (ie, diversity measured as inverse of Herfindahl index), and the average difference in the number of common procedures between 2 off-hours periods (regular hours versus weekends, and regular hours versus evenings). We also used NACOR data to estimate the average similarity in the distributions of procedures between regular hours and weekends and between regular hours and evenings in US facilities. Results are reported as mean ± standard error of the mean among 399 facilities nationwide with weekend cases. The distributions of common procedures were moderately similar (ie, not large, 2x the number of common procedures calculated by facility). The numbers of procedures commonly performed at most facilities are fewer in number than those that are commonly performed nationally. Thus, decisions on anesthesia specialization should be based on quantitative analysis of local data rather than national recommendations using pooled data. By facility, the number of different procedures that take place during regular hours and off hours (diversity) is essentially the same, but there is only moderate similarity in the procedures performed. Thus, at many facilities, anesthesiologists who work principally within a single specialty during regular work hours will likely not have substantial contemporary experience with many
Averaged Lema\\^itre-Tolman-Bondi dynamics
Isidro, Eddy G Chirinos; Piattella, Oliver F; Zimdahl, Winfried
2016-01-01
We consider cosmological backreaction effects in Buchert's averaging formalism on the basis of an explicit solution of the Lema\\^itre-Tolman-Bondi (LTB) dynamics which is linear in the LTB curvature parameter and has an inhomogeneous bang time. The volume Hubble rate is found in terms of the volume scale factor which represents a derivation of the simplest phenomenological solution of Buchert's equations in which the fractional densities corresponding to average curvature and kinematic backreaction are explicitly determined by the parameters of the underlying LTB solution at the boundary of the averaging volume. This configuration represents an exactly solvable toy model but it does not adequately describe our "real" Universe.
Average-passage flow model development
Adamczyk, John J.; Celestina, Mark L.; Beach, Tim A.; Kirtley, Kevin; Barnett, Mark
1989-01-01
A 3-D model was developed for simulating multistage turbomachinery flows using supercomputers. This average passage flow model described the time averaged flow field within a typical passage of a bladed wheel within a multistage configuration. To date, a number of inviscid simulations were executed to assess the resolution capabilities of the model. Recently, the viscous terms associated with the average passage model were incorporated into the inviscid computer code along with an algebraic turbulence model. A simulation of a stage-and-one-half, low speed turbine was executed. The results of this simulation, including a comparison with experimental data, is discussed.
FREQUENTIST MODEL AVERAGING ESTIMATION: A REVIEW
Haiying WANG; Xinyu ZHANG; Guohua ZOU
2009-01-01
In applications, the traditional estimation procedure generally begins with model selection.Once a specific model is selected, subsequent estimation is conducted under the selected model without consideration of the uncertainty from the selection process. This often leads to the underreporting of variability and too optimistic confidence sets. Model averaging estimation is an alternative to this procedure, which incorporates model uncertainty into the estimation process. In recent years, there has been a rising interest in model averaging from the frequentist perspective, and some important progresses have been made. In this paper, the theory and methods on frequentist model averaging estimation are surveyed. Some future research topics are also discussed.
Averaging of Backscatter Intensities in Compounds
Donovan, John J.; Pingitore, Nicholas E.; Westphal, Andrew J.
2002-01-01
Low uncertainty measurements on pure element stable isotope pairs demonstrate that mass has no influence on the backscattering of electrons at typical electron microprobe energies. The traditional prediction of average backscatter intensities in compounds using elemental mass fractions is improperly grounded in mass and thus has no physical basis. We propose an alternative model to mass fraction averaging, based of the number of electrons or protons, termed “electron fraction,” which predicts backscatter yield better than mass fraction averaging. PMID:27446752
Experimental Demonstration of Squeezed State Quantum Averaging
Lassen, Mikael; Sabuncu, Metin; Filip, Radim; Andersen, Ulrik L
2010-01-01
We propose and experimentally demonstrate a universal quantum averaging process implementing the harmonic mean of quadrature variances. The harmonic mean protocol can be used to efficiently stabilize a set of fragile squeezed light sources with statistically fluctuating noise levels. The averaged variances are prepared probabilistically by means of linear optical interference and measurement induced conditioning. We verify that the implemented harmonic mean outperforms the standard arithmetic mean strategy. The effect of quantum averaging is experimentally tested both for uncorrelated and partially correlated noise sources with sub-Poissonian shot noise or super-Poissonian shot noise characteristics.
The Average Lower Connectivity of Graphs
Ersin Aslan
2014-01-01
Full Text Available For a vertex v of a graph G, the lower connectivity, denoted by sv(G, is the smallest number of vertices that contains v and those vertices whose deletion from G produces a disconnected or a trivial graph. The average lower connectivity denoted by κav(G is the value (∑v∈VGsvG/VG. It is shown that this parameter can be used to measure the vulnerability of networks. This paper contains results on bounds for the average lower connectivity and obtains the average lower connectivity of some graphs.
Cosmic inhomogeneities and averaged cosmological dynamics.
Paranjape, Aseem; Singh, T P
2008-10-31
If general relativity (GR) describes the expansion of the Universe, the observed cosmic acceleration implies the existence of a "dark energy." However, while the Universe is on average homogeneous on large scales, it is inhomogeneous on smaller scales. While GR governs the dynamics of the inhomogeneous Universe, the averaged homogeneous Universe obeys modified Einstein equations. Can such modifications alone explain the acceleration? For a simple generic model with realistic initial conditions, we show the answer to be "no." Averaging effects negligibly influence the cosmological dynamics.
Changing mortality and average cohort life expectancy
Schoen, Robert; Canudas-Romo, Vladimir
2005-01-01
of survivorship. An alternative aggregate measure of period mortality which has been seen as less sensitive to period changes, the cross-sectional average length of life (CAL) has been proposed as an alternative, but has received only limited empirical or analytical examination. Here, we introduce a new measure......, the average cohort life expectancy (ACLE), to provide a precise measure of the average length of life of cohorts alive at a given time. To compare the performance of ACLE with CAL and with period and cohort life expectancy, we first use population models with changing mortality. Then the four aggregate...
Average subentropy, coherence and entanglement of random mixed quantum states
Zhang, Lin; Singh, Uttam; Pati, Arun K.
2017-02-01
Compact expressions for the average subentropy and coherence are obtained for random mixed states that are generated via various probability measures. Surprisingly, our results show that the average subentropy of random mixed states approaches the maximum value of the subentropy which is attained for the maximally mixed state as we increase the dimension. In the special case of the random mixed states sampled from the induced measure via partial tracing of random bipartite pure states, we establish the typicality of the relative entropy of coherence for random mixed states invoking the concentration of measure phenomenon. Our results also indicate that mixed quantum states are less useful compared to pure quantum states in higher dimension when we extract quantum coherence as a resource. This is because of the fact that average coherence of random mixed states is bounded uniformly, however, the average coherence of random pure states increases with the increasing dimension. As an important application, we establish the typicality of relative entropy of entanglement and distillable entanglement for a specific class of random bipartite mixed states. In particular, most of the random states in this specific class have relative entropy of entanglement and distillable entanglement equal to some fixed number (to within an arbitrary small error), thereby hugely reducing the complexity of computation of these entanglement measures for this specific class of mixed states.
R Wave Extraction Based on the Maximum First Derivative plus the Maximum Value of the Double Search
Wen-po Yao; Wen-li Yao; Min Wu; Tie-bing Liu
2016-01-01
R-wave detection is the main approach for heart rate variability analysis and clinical application based on R-R interval. The maximum ifrst derivative plus the maximum value of the double search algorithm is applied on electrocardiogram (ECG) of MIH-BIT Arrhythmia Database to extract R wave. Through the study of algorithm's characteristics and R-wave detection method, data segmentation method is modified to improve the detection accuracy. After segmentation modification, average accuracy rate of 6 sets of short ECG data increase from 82.51% to 93.70%, and the average accuracy rate of 11 groups long-range data is 96.61%. Test results prove that the algorithm and segmentation method can accurately locate R wave and have good effectiveness and versatility, but may exist some undetected problems due to algorithm implementation.
Receiver function estimated by maximum entropy deconvolution
吴庆举; 田小波; 张乃铃; 李卫平; 曾融生
2003-01-01
Maximum entropy deconvolution is presented to estimate receiver function, with the maximum entropy as the rule to determine auto-correlation and cross-correlation functions. The Toeplitz equation and Levinson algorithm are used to calculate the iterative formula of error-predicting filter, and receiver function is then estimated. During extrapolation, reflective coefficient is always less than 1, which keeps maximum entropy deconvolution stable. The maximum entropy of the data outside window increases the resolution of receiver function. Both synthetic and real seismograms show that maximum entropy deconvolution is an effective method to measure receiver function in time-domain.
Sea Surface Temperature Average_SST_Master
National Oceanic and Atmospheric Administration, Department of Commerce — Sea surface temperature collected via satellite imagery from http://www.esrl.noaa.gov/psd/data/gridded/data.noaa.ersst.html and averaged for each region using ArcGIS...
Appeals Council Requests - Average Processing Time
Social Security Administration — This dataset provides annual data from 1989 through 2015 for the average processing time (elapsed time in days) for dispositions by the Appeals Council (AC) (both...
Average Vegetation Growth 1990 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1990 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 1997 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1997 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 1992 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1992 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 2001 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 2001 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 1995 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1995 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 2000 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 2000 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 1998 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1998 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 1994 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1994 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
MN Temperature Average (1961-1990) - Line
Minnesota Department of Natural Resources — This data set depicts 30-year averages (1961-1990) of monthly and annual temperatures for Minnesota. Isolines and regions were created using kriging and...
Average Vegetation Growth 1996 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1996 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 2005 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 2005 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
Average Vegetation Growth 1993 - Direct Download
U.S. Geological Survey, Department of the Interior — This map layer is a grid map of 1993 average vegetation growth for Alaska and the conterminous United States. The nominal spatial resolution is 1 kilometer and the...
MN Temperature Average (1961-1990) - Polygon
Minnesota Department of Natural Resources — This data set depicts 30-year averages (1961-1990) of monthly and annual temperatures for Minnesota. Isolines and regions were created using kriging and...
Spacetime Average Density (SAD) Cosmological Measures
Page, Don N
2014-01-01
The measure problem of cosmology is how to obtain normalized probabilities of observations from the quantum state of the universe. This is particularly a problem when eternal inflation leads to a universe of unbounded size so that there are apparently infinitely many realizations or occurrences of observations of each of many different kinds or types, making the ratios ambiguous. There is also the danger of domination by Boltzmann Brains. Here two new Spacetime Average Density (SAD) measures are proposed, Maximal Average Density (MAD) and Biased Average Density (BAD), for getting a finite number of observation occurrences by using properties of the Spacetime Average Density (SAD) of observation occurrences to restrict to finite regions of spacetimes that have a preferred beginning or bounce hypersurface. These measures avoid Boltzmann brain domination and appear to give results consistent with other observations that are problematic for other widely used measures, such as the observation of a positive cosmolo...
A practical guide to averaging functions
Beliakov, Gleb; Calvo Sánchez, Tomasa
2016-01-01
This book offers an easy-to-use and practice-oriented reference guide to mathematical averages. It presents different ways of aggregating input values given on a numerical scale, and of choosing and/or constructing aggregating functions for specific applications. Building on a previous monograph by Beliakov et al. published by Springer in 2007, it outlines new aggregation methods developed in the interim, with a special focus on the topic of averaging aggregation functions. It examines recent advances in the field, such as aggregation on lattices, penalty-based aggregation and weakly monotone averaging, and extends many of the already existing methods, such as: ordered weighted averaging (OWA), fuzzy integrals and mixture functions. A substantial mathematical background is not called for, as all the relevant mathematical notions are explained here and reported on together with a wealth of graphical illustrations of distinct families of aggregation functions. The authors mainly focus on practical applications ...
Rotational averaging of multiphoton absorption cross sections
Friese, Daniel H., E-mail: daniel.h.friese@uit.no; Beerepoot, Maarten T. P.; Ruud, Kenneth [Centre for Theoretical and Computational Chemistry, University of Tromsø — The Arctic University of Norway, N-9037 Tromsø (Norway)
2014-11-28
Rotational averaging of tensors is a crucial step in the calculation of molecular properties in isotropic media. We present a scheme for the rotational averaging of multiphoton absorption cross sections. We extend existing literature on rotational averaging to even-rank tensors of arbitrary order and derive equations that require only the number of photons as input. In particular, we derive the first explicit expressions for the rotational average of five-, six-, and seven-photon absorption cross sections. This work is one of the required steps in making the calculation of these higher-order absorption properties possible. The results can be applied to any even-rank tensor provided linearly polarized light is used.
Rotational averaging of multiphoton absorption cross sections
Friese, Daniel H.; Beerepoot, Maarten T. P.; Ruud, Kenneth
2014-11-01
Rotational averaging of tensors is a crucial step in the calculation of molecular properties in isotropic media. We present a scheme for the rotational averaging of multiphoton absorption cross sections. We extend existing literature on rotational averaging to even-rank tensors of arbitrary order and derive equations that require only the number of photons as input. In particular, we derive the first explicit expressions for the rotational average of five-, six-, and seven-photon absorption cross sections. This work is one of the required steps in making the calculation of these higher-order absorption properties possible. The results can be applied to any even-rank tensor provided linearly polarized light is used.
Monthly snow/ice averages (ISCCP)
National Aeronautics and Space Administration — September Arctic sea ice is now declining at a rate of 11.5 percent per decade, relative to the 1979 to 2000 average. Data from NASA show that the land ice sheets in...
Average Annual Precipitation (PRISM model) 1961 - 1990
U.S. Geological Survey, Department of the Interior — This map layer shows polygons of average annual precipitation in the contiguous United States, for the climatological period 1961-1990. Parameter-elevation...
Maximum Power from a Solar Panel
Michael Miller
2010-01-01
Full Text Available Solar energy has become a promising alternative to conventional fossil fuel sources. Solar panels are used to collect solar radiation and convert it into electricity. One of the techniques used to maximize the effectiveness of this energy alternative is to maximize the power output of the solar collector. In this project the maximum power is calculated by determining the voltage and the current of maximum power. These quantities are determined by finding the maximum value for the equation for power using differentiation. After the maximum values are found for each time of day, each individual quantity, voltage of maximum power, current of maximum power, and maximum power is plotted as a function of the time of day.
Symmetric Euler orientation representations for orientational averaging.
Mayerhöfer, Thomas G
2005-09-01
A new kind of orientation representation called symmetric Euler orientation representation (SEOR) is presented. It is based on a combination of the conventional Euler orientation representations (Euler angles) and Hamilton's quaternions. The properties of the SEORs concerning orientational averaging are explored and compared to those of averaging schemes that are based on conventional Euler orientation representations. To that aim, the reflectance of a hypothetical polycrystalline material with orthorhombic crystal symmetry was calculated. The calculation was carried out according to the average refractive index theory (ARIT [T.G. Mayerhöfer, Appl. Spectrosc. 56 (2002) 1194]). It is shown that the use of averaging schemes based on conventional Euler orientation representations leads to a dependence of the result from the specific Euler orientation representation that was utilized and from the initial position of the crystal. The latter problem can be overcome partly by the introduction of a weighing factor, but only for two-axes-type Euler orientation representations. In case of a numerical evaluation of the average, a residual difference remains also if a two-axes type Euler orientation representation is used despite of the utilization of a weighing factor. In contrast, this problem does not occur if a symmetric Euler orientation representation is used as a matter of principle, while the result of the averaging for both types of orientation representations converges with increasing number of orientations considered in the numerical evaluation. Additionally, the use of a weighing factor and/or non-equally spaced steps in the numerical evaluation of the average is not necessary. The symmetrical Euler orientation representations are therefore ideally suited for the use in orientational averaging procedures.
Cosmic Inhomogeneities and the Average Cosmological Dynamics
Paranjape, Aseem; Singh, T. P.
2008-01-01
If general relativity (GR) describes the expansion of the Universe, the observed cosmic acceleration implies the existence of a `dark energy'. However, while the Universe is on average homogeneous on large scales, it is inhomogeneous on smaller scales. While GR governs the dynamics of the \\emph{in}homogeneous Universe, the averaged \\emph{homogeneous} Universe obeys modified Einstein equations. Can such modifications alone explain the acceleration? For a simple generic model with realistic ini...
Average Bandwidth Allocation Model of WFQ
Tomáš Balogh
2012-01-01
Full Text Available We present a new iterative method for the calculation of average bandwidth assignment to traffic flows using a WFQ scheduler in IP based NGN networks. The bandwidth assignment calculation is based on the link speed, assigned weights, arrival rate, and average packet length or input rate of the traffic flows. We prove the model outcome with examples and simulation results using NS2 simulator.
U.S. Environmental Protection Agency — This EnviroAtlas dataset shows the annual average potential wind energy resource in kilowatt hours per square kilometer per day for each 12-digit Hydrologic Unit...
A global surface drifter data set at hourly resolution
Elipot, Shane; Lumpkin, Rick; Perez, Renellys C.; Lilly, Jonathan M.; Early, Jeffrey J.; Sykulski, Adam M.
2016-05-01
The surface drifting buoys, or drifters, of the Global Drifter Program (GDP) are predominantly tracked by the Argos positioning system, providing drifter locations with O(100 m) errors at nonuniform temporal intervals, with an average interval of 1.2 h since January 2005. This data set is thus a rich and global source of information on high-frequency and small-scale oceanic processes, yet is still relatively understudied because of the challenges associated with its large size and sampling characteristics. A methodology is described to produce a new high-resolution global data set since 2005, consisting of drifter locations and velocities estimated at hourly intervals, along with their respective errors. Locations and velocities are obtained by modeling locally in time trajectories as a first-order polynomial with coefficients obtained by maximizing a likelihood function. This function is derived by modeling the Argos location errors with t location-scale probability distribution functions. The methodology is motivated by analyzing 82 drifters tracked contemporaneously by Argos and by the Global Positioning System, where the latter is assumed to provide true locations. A global spectral analysis of the velocity variance from the new data set reveals a sharply defined ridge of energy closely following the inertial frequency as a function of latitude, distinct energy peaks near diurnal and semidiurnal frequencies, as well as higher-frequency peaks located near tidal harmonics as well as near replicates of the inertial frequency. Compared to the spectra that can be obtained using the standard 6-hourly GDP product, the new data set contains up to 100% more spectral energy at some latitudes.
Barbero, R.; Fowler, H. J.; Lenderink, G.; Blenkinsop, S.
2017-01-01
Although it has been documented that daily precipitation extremes are increasing worldwide, faster increases may be expected for subdaily extremes. Here after a careful quality control procedure, we compared trends in hourly and daily precipitation extremes using a large network of stations across the United States (U.S.) within the 1950-2011 period. A greater number of significant increasing trends in annual and seasonal maximum precipitation were detected from daily extremes, with the primary exception of wintertime. Our results also show that the mean percentage change in annual maximum daily precipitation across the U.S. per global warming degree is 6.9% °C-1 (in agreement with the Clausius-Clapeyron rate) while lower sensitivities were observed for hourly extremes, suggesting that changes in the magnitude of subdaily extremes in response to global warming emerge more slowly than those for daily extremes in the climate record.
Averaged controllability of parameter dependent conservative semigroups
Lohéac, Jérôme; Zuazua, Enrique
2017-02-01
We consider the problem of averaged controllability for parameter depending (either in a discrete or continuous fashion) control systems, the aim being to find a control, independent of the unknown parameters, so that the average of the states is controlled. We do it in the context of conservative models, both in an abstract setting and also analysing the specific examples of the wave and Schrödinger equations. Our first result is of perturbative nature. Assuming the averaging probability measure to be a small parameter-dependent perturbation (in a sense that we make precise) of an atomic measure given by a Dirac mass corresponding to a specific realisation of the system, we show that the averaged controllability property is achieved whenever the system corresponding to the support of the Dirac is controllable. Similar tools can be employed to obtain averaged versions of the so-called Ingham inequalities. Particular attention is devoted to the 1d wave equation in which the time-periodicity of solutions can be exploited to obtain more precise results, provided the parameters involved satisfy Diophantine conditions ensuring the lack of resonances.
Average Temperatures in the Southwestern United States, 2000-2015 Versus Long-Term Average
U.S. Environmental Protection Agency — This indicator shows how the average air temperature from 2000 to 2015 has differed from the long-term average (1895–2015). To provide more detailed information,...
Change in the body temperature of healthy term infant over the first 72 hours of life
李萌霞; 孙革; NEUBAUERHenning
2004-01-01
Objective:To determine the range of body temperature in a group of healthy Chinese term neonates over the first 72 hours of life and to assess the influence of body weight, gestational age and route of delivery.Method: All 200 consecutive cases of neonates delivered at our hospital from March to August 2001 were included in this retrospective study.Temperatures were measured immediately after delivery, after 30 minutes, 1 hour, 2 hours, 8 hours and 15 hours and on the 2nd and 3rd day. Axillary temperatures ranging from 36.5℃ to 37℃ were regarded as normal. No cases of maternal fever or systemic infection of the newborns were discovered. All infants were discharged in good general condition. Results:The mean rectal temperature at birth was 37.19℃. The lowest average temperature was reached at 1 hour after delivery (36.54℃) with a significant difference between natural delivery (36.48℃) and section (36.59℃) (P<0.05).Temperature subsequently rose to 36.70℃ at 8 hours and 36.78℃ at 15 hours (P<0.05).Hypothermia was seen in 51.8% and hypothermia in 42.5% of the patients.On the 3rd day after delivery, 96% of all temperatures were in the normal range. A significant relation was found between hypothermia and both low birth weight (P<0.001) and low gestational age (P<0.05).Conclusion: The reference range presently used did not include all physiological temperatures in the first 72 hours of life. Considering other factors,such as birth weight, route of delivery,gestational age and body temperature on the 2nd and 3rd day of life, may help to correctly assess the significance of temperatures beyond the reference range.
Change in the body temperature of healthy term infant over the first 72 hours of life
LI Meng-xia (李萌霞); SUN Ge (孙革); NEUBAUER Henning
2004-01-01
Objective: To determine the range of body temperature in a group of healthy Chinese term neonates over the first 72 hours of life and to assess the influence of body weight, gestational age and route of delivery. Method: All 200 consecutive cases of neonates delivered at our hospital from March to August 2001 were included in this retrospective study. Temperatures were measured immediately after delivery, after 30 minutes, 1 hour, 2 hours, 8 hours and 15 hours and on the 2nd and 3rd day. Axillary temperatures ranging from 36.5 oC to 37 oC were regarded as normal. No cases of maternal fever or systemic infection of the newborns were discovered. All infants were discharged in good general condition. Results: The mean rectal temperature at birth was 37.19 ℃. The lowest average temperature was reached at 1 hour after delivery (36.54 ℃) with a significant difference between natural delivery (36.48 ℃) and section (36.59 ℃) (P<0.05). Temperature subsequently rose to 36.70 ℃ at 8 hours and 36.78 ℃ at 15 hours (P<0.05). Hypothermia was seen in 51.8% and hypothermia in 42.5% of the patients. On the 3rd day after delivery, 96% of all temperatures were in the normal range. A significant relation was found between hypothermia and both low birth weight (P<0.001) and low gestational age (P<0.05). Conclusion: The reference range presently used did not include all physiological temperatures in the first 72 hours of life. Considering other factors, such as birth weight, route of delivery, gestational age and body temperature on the 2nd and 3rd day of life, may help to correctly assess the significance of temperatures beyond the reference range.
The inverse maximum dynamic flow problem
BAGHERIAN; Mehri
2010-01-01
We consider the inverse maximum dynamic flow (IMDF) problem.IMDF problem can be described as: how to change the capacity vector of a dynamic network as little as possible so that a given feasible dynamic flow becomes a maximum dynamic flow.After discussing some characteristics of this problem,it is converted to a constrained minimum dynamic cut problem.Then an efficient algorithm which uses two maximum dynamic flow algorithms is proposed to solve the problem.
Cosmic structure, averaging and dark energy
Wiltshire, David L
2013-01-01
These lecture notes review the theoretical problems associated with coarse-graining the observed inhomogeneous structure of the universe at late epochs, of describing average cosmic evolution in the presence of growing inhomogeneity, and of relating average quantities to physical observables. In particular, a detailed discussion of the timescape scenario is presented. In this scenario, dark energy is realized as a misidentification of gravitational energy gradients which result from gradients in the kinetic energy of expansion of space, in the presence of density and spatial curvature gradients that grow large with the growth of structure. The phenomenology and observational tests of the timescape model are discussed in detail, with updated constraints from Planck satellite data. In addition, recent results on the variation of the Hubble expansion on < 100/h Mpc scales are discussed. The spherically averaged Hubble law is significantly more uniform in the rest frame of the Local Group of galaxies than in t...
Books average previous decade of economic misery.
R Alexander Bentley
Full Text Available For the 20(th century since the Depression, we find a strong correlation between a 'literary misery index' derived from English language books and a moving average of the previous decade of the annual U.S. economic misery index, which is the sum of inflation and unemployment rates. We find a peak in the goodness of fit at 11 years for the moving average. The fit between the two misery indices holds when using different techniques to measure the literary misery index, and this fit is significantly better than other possible correlations with different emotion indices. To check the robustness of the results, we also analysed books written in German language and obtained very similar correlations with the German economic misery index. The results suggest that millions of books published every year average the authors' shared economic experiences over the past decade.
Benchmarking statistical averaging of spectra with HULLAC
Klapisch, Marcel; Busquet, Michel
2008-11-01
Knowledge of radiative properties of hot plasmas is important for ICF, astrophysics, etc When mid-Z or high-Z elements are present, the spectra are so complex that one commonly uses statistically averaged description of atomic systems [1]. In a recent experiment on Fe[2], performed under controlled conditions, high resolution transmission spectra were obtained. The new version of HULLAC [3] allows the use of the same model with different levels of details/averaging. We will take advantage of this feature to check the effect of averaging with comparison with experiment. [1] A Bar-Shalom, J Oreg, and M Klapisch, J. Quant. Spectros. Rad. Transf. 65, 43 (2000). [2] J. E. Bailey, G. A. Rochau, C. A. Iglesias et al., Phys. Rev. Lett. 99, 265002-4 (2007). [3]. M. Klapisch, M. Busquet, and A. Bar-Shalom, AIP Conference Proceedings 926, 206-15 (2007).
Stochastic Averaging and Stochastic Extremum Seeking
Liu, Shu-Jun
2012-01-01
Stochastic Averaging and Stochastic Extremum Seeking develops methods of mathematical analysis inspired by the interest in reverse engineering and analysis of bacterial convergence by chemotaxis and to apply similar stochastic optimization techniques in other environments. The first half of the text presents significant advances in stochastic averaging theory, necessitated by the fact that existing theorems are restricted to systems with linear growth, globally exponentially stable average models, vanishing stochastic perturbations, and prevent analysis over infinite time horizon. The second half of the text introduces stochastic extremum seeking algorithms for model-free optimization of systems in real time using stochastic perturbations for estimation of their gradients. Both gradient- and Newton-based algorithms are presented, offering the user the choice between the simplicity of implementation (gradient) and the ability to achieve a known, arbitrary convergence rate (Newton). The design of algorithms...
Conway, Sadie H; Pompeii, Lisa A; Roberts, Robert E; Follis, Jack L; Gimeno, David
2016-03-01
The aim of this study was to examine the presence of a dose-response relationship between work hours and incident cardiovascular disease (CVD) in a representative sample of U.S. workers. A retrospective cohort study of 1926 individuals from the Panel Study of Income Dynamics (1986 to 2011) employed for at least 10 years. Restricted cubic spline regression was used to estimate the dose-response relationship of work hours with CVD. A dose-response relationship was observed in which an average workweek of 46 hours or more for at least 10 years was associated with an increased risk of CVD. Compared with working 45 hours per week, working an additional 10 hours per week or more for at least 10 years increased CVD risk by at least 16%. Working more than 45 work hours per week for at least 10 years may be an independent risk factor for CVD.
Books average previous decade of economic misery.
Bentley, R Alexander; Acerbi, Alberto; Ormerod, Paul; Lampos, Vasileios
2014-01-01
For the 20(th) century since the Depression, we find a strong correlation between a 'literary misery index' derived from English language books and a moving average of the previous decade of the annual U.S. economic misery index, which is the sum of inflation and unemployment rates. We find a peak in the goodness of fit at 11 years for the moving average. The fit between the two misery indices holds when using different techniques to measure the literary misery index, and this fit is significantly better than other possible correlations with different emotion indices. To check the robustness of the results, we also analysed books written in German language and obtained very similar correlations with the German economic misery index. The results suggest that millions of books published every year average the authors' shared economic experiences over the past decade.
High Average Power Yb:YAG Laser
Zapata, L E; Beach, R J; Payne, S A
2001-05-23
We are working on a composite thin-disk laser design that can be scaled as a source of high brightness laser power for tactical engagement and other high average power applications. The key component is a diffusion-bonded composite comprising a thin gain-medium and thicker cladding that is strikingly robust and resolves prior difficulties with high average power pumping/cooling and the rejection of amplified spontaneous emission (ASE). In contrast to high power rods or slabs, the one-dimensional nature of the cooling geometry and the edge-pump geometry scale gracefully to very high average power. The crucial design ideas have been verified experimentally. Progress this last year included: extraction with high beam quality using a telescopic resonator, a heterogeneous thin film coating prescription that meets the unusual requirements demanded by this laser architecture, thermal management with our first generation cooler. Progress was also made in design of a second-generation laser.
Sefik M.Bajmak
2013-11-01
Full Text Available Work together more cooling source (refrigeration machines the system of centralized supply cooling energy ( SCSCE is a way to achieve cost-effective operation and safe and rational supply consumption area with cool water for central cooling and air conditioning . Maximum energy needs cold water occurs rarely , because the extremely high temperatures occur rarely . Therefore , the total cooling load is divided into basic and peak . One of the main characteristics that define the justification of the use of coupled processes and sizes hourly coefficient centralized supply of cold water, temperature regime , or hour coefficient district cooling . Determination of Optimal hour coefficient district cooling is one of the most techno economic tasks at the design of the system of centralized supply cold water for air conditioning and industrial building social housing and business districts .
Maximum permissible voltage of YBCO coated conductors
Wen, J.; Lin, B.; Sheng, J.; Xu, J.; Jin, Z. [Department of Electrical Engineering, Shanghai Jiao Tong University, Shanghai (China); Hong, Z., E-mail: zhiyong.hong@sjtu.edu.cn [Department of Electrical Engineering, Shanghai Jiao Tong University, Shanghai (China); Wang, D.; Zhou, H.; Shen, X.; Shen, C. [Qingpu Power Supply Company, State Grid Shanghai Municipal Electric Power Company, Shanghai (China)
2014-06-15
Highlights: • We examine three kinds of tapes’ maximum permissible voltage. • We examine the relationship between quenching duration and maximum permissible voltage. • Continuous I{sub c} degradations under repetitive quenching where tapes reaching maximum permissible voltage. • The relationship between maximum permissible voltage and resistance, temperature. - Abstract: Superconducting fault current limiter (SFCL) could reduce short circuit currents in electrical power system. One of the most important thing in developing SFCL is to find out the maximum permissible voltage of each limiting element. The maximum permissible voltage is defined as the maximum voltage per unit length at which the YBCO coated conductors (CC) do not suffer from critical current (I{sub c}) degradation or burnout. In this research, the time of quenching process is changed and voltage is raised until the I{sub c} degradation or burnout happens. YBCO coated conductors test in the experiment are from American superconductor (AMSC) and Shanghai Jiao Tong University (SJTU). Along with the quenching duration increasing, the maximum permissible voltage of CC decreases. When quenching duration is 100 ms, the maximum permissible of SJTU CC, 12 mm AMSC CC and 4 mm AMSC CC are 0.72 V/cm, 0.52 V/cm and 1.2 V/cm respectively. Based on the results of samples, the whole length of CCs used in the design of a SFCL can be determined.
The modulated average structure of mullite.
Birkenstock, Johannes; Petříček, Václav; Pedersen, Bjoern; Schneider, Hartmut; Fischer, Reinhard X
2015-06-01
Homogeneous and inclusion-free single crystals of 2:1 mullite (Al(4.8)Si(1.2)O(9.6)) grown by the Czochralski technique were examined by X-ray and neutron diffraction methods. The observed diffuse scattering together with the pattern of satellite reflections confirm previously published data and are thus inherent features of the mullite structure. The ideal composition was closely met as confirmed by microprobe analysis (Al(4.82 (3))Si(1.18 (1))O(9.59 (5))) and by average structure refinements. 8 (5) to 20 (13)% of the available Si was found in the T* position of the tetrahedra triclusters. The strong tendencey for disorder in mullite may be understood from considerations of hypothetical superstructures which would have to be n-fivefold with respect to the three-dimensional average unit cell of 2:1 mullite and n-fourfold in case of 3:2 mullite. In any of these the possible arrangements of the vacancies and of the tetrahedral units would inevitably be unfavorable. Three directions of incommensurate modulations were determined: q1 = [0.3137 (2) 0 ½], q2 = [0 0.4021 (5) 0.1834 (2)] and q3 = [0 0.4009 (5) -0.1834 (2)]. The one-dimensional incommensurately modulated crystal structure associated with q1 was refined for the first time using the superspace approach. The modulation is dominated by harmonic occupational modulations of the atoms in the di- and the triclusters of the tetrahedral units in mullite. The modulation amplitudes are small and the harmonic character implies that the modulated structure still represents an average structure in the overall disordered arrangement of the vacancies and of the tetrahedral structural units. In other words, when projecting the local assemblies at the scale of a few tens of average mullite cells into cells determined by either one of the modulation vectors q1, q2 or q3 a weak average modulation results with slightly varying average occupation factors for the tetrahedral units. As a result, the real
A singularity theorem based on spatial averages
J M M Senovilla
2007-07-01
Inspired by Raychaudhuri's work, and using the equation named after him as a basic ingredient, a new singularity theorem is proved. Open non-rotating Universes, expanding everywhere with a non-vanishing spatial average of the matter variables, show severe geodesic incompletness in the past. Another way of stating the result is that, under the same conditions, any singularity-free model must have a vanishing spatial average of the energy density (and other physical variables). This is very satisfactory and provides a clear decisive difference between singular and non-singular cosmologies.
Average: the juxtaposition of procedure and context
Watson, Jane; Chick, Helen; Callingham, Rosemary
2014-09-01
This paper presents recent data on the performance of 247 middle school students on questions concerning average in three contexts. Analysis includes considering levels of understanding linking definition and context, performance across contexts, the relative difficulty of tasks, and difference in performance for male and female students. The outcomes lead to a discussion of the expectations of the curriculum and its implementation, as well as assessment, in relation to students' skills in carrying out procedures and their understanding about the meaning of average in context.
SOURCE TERMS FOR AVERAGE DOE SNF CANISTERS
K. L. Goluoglu
2000-06-09
The objective of this calculation is to generate source terms for each type of Department of Energy (DOE) spent nuclear fuel (SNF) canister that may be disposed of at the potential repository at Yucca Mountain. The scope of this calculation is limited to generating source terms for average DOE SNF canisters, and is not intended to be used for subsequent calculations requiring bounding source terms. This calculation is to be used in future Performance Assessment calculations, or other shielding or thermal calculations requiring average source terms.
An approximate analytical approach to resampling averages
Malzahn, Dorthe; Opper, M.
2004-01-01
Using a novel reformulation, we develop a framework to compute approximate resampling data averages analytically. The method avoids multiple retraining of statistical models on the samples. Our approach uses a combination of the replica "trick" of statistical physics and the TAP approach for appr......Using a novel reformulation, we develop a framework to compute approximate resampling data averages analytically. The method avoids multiple retraining of statistical models on the samples. Our approach uses a combination of the replica "trick" of statistical physics and the TAP approach...
Grassmann Averages for Scalable Robust PCA
Hauberg, Søren; Feragen, Aasa; Black, Michael J.
2014-01-01
As the collection of large datasets becomes increasingly automated, the occurrence of outliers will increase—“big data” implies “big outliers”. While principal component analysis (PCA) is often used to reduce the size of data, and scalable solutions exist, it is well-known that outliers can...... to vectors (subspaces) or elements of vectors; we focus on the latter and use a trimmed average. The resulting Trimmed Grassmann Average (TGA) is particularly appropriate for computer vision because it is robust to pixel outliers. The algorithm has low computational complexity and minimal memory requirements...
Model averaging and muddled multimodel inferences.
Cade, Brian S
2015-09-01
Three flawed practices associated with model averaging coefficients for predictor variables in regression models commonly occur when making multimodel inferences in analyses of ecological data. Model-averaged regression coefficients based on Akaike information criterion (AIC) weights have been recommended for addressing model uncertainty but they are not valid, interpretable estimates of partial effects for individual predictors when there is multicollinearity among the predictor variables. Multicollinearity implies that the scaling of units in the denominators of the regression coefficients may change across models such that neither the parameters nor their estimates have common scales, therefore averaging them makes no sense. The associated sums of AIC model weights recommended to assess relative importance of individual predictors are really a measure of relative importance of models, with little information about contributions by individual predictors compared to other measures of relative importance based on effects size or variance reduction. Sometimes the model-averaged regression coefficients for predictor variables are incorrectly used to make model-averaged predictions of the response variable when the models are not linear in the parameters. I demonstrate the issues with the first two practices using the college grade point average example extensively analyzed by Burnham and Anderson. I show how partial standard deviations of the predictor variables can be used to detect changing scales of their estimates with multicollinearity. Standardizing estimates based on partial standard deviations for their variables can be used to make the scaling of the estimates commensurate across models, a necessary but not sufficient condition for model averaging of the estimates to be sensible. A unimodal distribution of estimates and valid interpretation of individual parameters are additional requisite conditions. The standardized estimates or equivalently the t
Model averaging and muddled multimodel inferences
Cade, Brian S.
2015-01-01
Three flawed practices associated with model averaging coefficients for predictor variables in regression models commonly occur when making multimodel inferences in analyses of ecological data. Model-averaged regression coefficients based on Akaike information criterion (AIC) weights have been recommended for addressing model uncertainty but they are not valid, interpretable estimates of partial effects for individual predictors when there is multicollinearity among the predictor variables. Multicollinearity implies that the scaling of units in the denominators of the regression coefficients may change across models such that neither the parameters nor their estimates have common scales, therefore averaging them makes no sense. The associated sums of AIC model weights recommended to assess relative importance of individual predictors are really a measure of relative importance of models, with little information about contributions by individual predictors compared to other measures of relative importance based on effects size or variance reduction. Sometimes the model-averaged regression coefficients for predictor variables are incorrectly used to make model-averaged predictions of the response variable when the models are not linear in the parameters. I demonstrate the issues with the first two practices using the college grade point average example extensively analyzed by Burnham and Anderson. I show how partial standard deviations of the predictor variables can be used to detect changing scales of their estimates with multicollinearity. Standardizing estimates based on partial standard deviations for their variables can be used to make the scaling of the estimates commensurate across models, a necessary but not sufficient condition for model averaging of the estimates to be sensible. A unimodal distribution of estimates and valid interpretation of individual parameters are additional requisite conditions. The standardized estimates or equivalently the
CMAQv502 Base AQS Hourly site compare output
U.S. Environmental Protection Agency — Monthly AQS hourly site compare output files for the CMAQv502 Base simulation. Monthly files contain hourly paired model/ob data for the AQS network. These data were...
Rush Hour Pollution May Be Worse Than Thought
... medlineplus.gov/news/fullstory_167478.html Rush Hour Pollution May Be Worse Than Thought Higher measurements seen ... in rush-hour traffic? New research finds air pollution levels in cars are much higher than previously ...
A Systematic Review of the Effects of Resident Duty Hour Restrictions in Surgery
Devitt, Katharine S.; Keshet, Itay; Spicer, Jonathan; Imrie, Kevin; Feldman, Liane; Cools-Lartigue, Jonathan; Kayssi, Ahmed; Lipsman, Nir; Elmi, Maryam; Kulkarni, Abhaya V.; Parshuram, Chris; Mainprize, Todd; Warren, Richard J.; Fata, Paola; Gorman, M. Sean; Feinberg, Stan; Rutka, James
2014-01-01
Background: In 2003, the Accreditation Council for Graduate Medical Education (ACGME) mandated 80-hour resident duty limits. In 2011 the ACGME mandated 16-hour duty maximums for PGY1 (post graduate year) residents. The stated goals were to improve patient safety, resident well-being, and education. A systematic review and meta-analysis were performed to evaluate the impact of resident duty hours (RDH) on clinical and educational outcomes in surgery. Methods: A systematic review (1980–2013) was executed on CINAHL, Cochrane Database, Embase, Medline, and Scopus. Quality of articles was assessed using the GRADE guidelines. Sixteen-hour shifts and night float systems were analyzed separately. Articles that examined mortality data were combined in a random-effects meta-analysis to evaluate the impact of RDH on patient mortality. Results: A total of 135 articles met the inclusion criteria. Among these, 42% (N = 57) were considered moderate-high quality. There was no overall improvement in patient outcomes as a result of RDH; however, some studies suggest increased complication rates in high-acuity patients. There was no improvement in education related to RDH restrictions, and performance on certification examinations has declined in some specialties. Survey studies revealed a perception of worsened education and patient safety. There were improvements in resident wellness after the 80-hour workweek, but there was little improvement or negative effects on wellness after 16-hour duty maximums were implemented. Conclusions: Recent RDH changes are not consistently associated with improvements in resident well-being, and have negative impacts on patient outcomes and performance on certification examinations. Greater flexibility to accommodate resident training needs is required. Further erosion of training time should be considered with great caution. PMID:24662409
Rolling bearing feature frequency extraction using extreme average envelope decomposition
Shi, Kunju; Liu, Shulin; Jiang, Chao; Zhang, Hongli
2016-09-01
The vibration signal contains a wealth of sensitive information which reflects the running status of the equipment. It is one of the most important steps for precise diagnosis to decompose the signal and extracts the effective information properly. The traditional classical adaptive signal decomposition method, such as EMD, exists the problems of mode mixing, low decomposition accuracy etc. Aiming at those problems, EAED(extreme average envelope decomposition) method is presented based on EMD. EAED method has three advantages. Firstly, it is completed through midpoint envelopment method rather than using maximum and minimum envelopment respectively as used in EMD. Therefore, the average variability of the signal can be described accurately. Secondly, in order to reduce the envelope errors during the signal decomposition, replacing two envelopes with one envelope strategy is presented. Thirdly, the similar triangle principle is utilized to calculate the time of extreme average points accurately. Thus, the influence of sampling frequency on the calculation results can be significantly reduced. Experimental results show that EAED could separate out single frequency components from a complex signal gradually. EAED could not only isolate three kinds of typical bearing fault characteristic of vibration frequency components but also has fewer decomposition layers. EAED replaces quadratic enveloping to an envelope which ensuring to isolate the fault characteristic frequency under the condition of less decomposition layers. Therefore, the precision of signal decomposition is improved.
On the average uncertainty for systems with nonlinear coupling
Nelson, Kenric P.; Umarov, Sabir R.; Kon, Mark A.
2017-02-01
The increased uncertainty and complexity of nonlinear systems have motivated investigators to consider generalized approaches to defining an entropy function. New insights are achieved by defining the average uncertainty in the probability domain as a transformation of entropy functions. The Shannon entropy when transformed to the probability domain is the weighted geometric mean of the probabilities. For the exponential and Gaussian distributions, we show that the weighted geometric mean of the distribution is equal to the density of the distribution at the location plus the scale (i.e. at the width of the distribution). The average uncertainty is generalized via the weighted generalized mean, in which the moment is a function of the nonlinear source. Both the Rényi and Tsallis entropies transform to this definition of the generalized average uncertainty in the probability domain. For the generalized Pareto and Student's t-distributions, which are the maximum entropy distributions for these generalized entropies, the appropriate weighted generalized mean also equals the density of the distribution at the location plus scale. A coupled entropy function is proposed, which is equal to the normalized Tsallis entropy divided by one plus the coupling.
Parameterized Traveling Salesman Problem: Beating the Average
Gutin, G.; Patel, V.
2016-01-01
In the traveling salesman problem (TSP), we are given a complete graph Kn together with an integer weighting w on the edges of Kn, and we are asked to find a Hamilton cycle of Kn of minimum weight. Let h(w) denote the average weight of a Hamilton cycle of Kn for the weighting w. Vizing in 1973 asked
On averaging methods for partial differential equations
Verhulst, F.
2001-01-01
The analysis of weakly nonlinear partial differential equations both qualitatively and quantitatively is emerging as an exciting eld of investigation In this report we consider specic results related to averaging but we do not aim at completeness The sections and contain important material which
Discontinuities and hysteresis in quantized average consensus
Ceragioli, Francesca; Persis, Claudio De; Frasca, Paolo
2011-01-01
We consider continuous-time average consensus dynamics in which the agents’ states are communicated through uniform quantizers. Solutions to the resulting system are defined in the Krasowskii sense and are proven to converge to conditions of ‘‘practical consensus’’. To cope with undesired chattering
Bayesian Averaging is Well-Temperated
Hansen, Lars Kai
2000-01-01
Bayesian predictions are stochastic just like predictions of any other inference scheme that generalize from a finite sample. While a simple variational argument shows that Bayes averaging is generalization optimal given that the prior matches the teacher parameter distribution the situation...
A Functional Measurement Study on Averaging Numerosity
Tira, Michael D.; Tagliabue, Mariaelena; Vidotto, Giulio
2014-01-01
In two experiments, participants judged the average numerosity between two sequentially presented dot patterns to perform an approximate arithmetic task. In Experiment 1, the response was given on a 0-20 numerical scale (categorical scaling), and in Experiment 2, the response was given by the production of a dot pattern of the desired numerosity…
Generalized Jackknife Estimators of Weighted Average Derivatives
Cattaneo, Matias D.; Crump, Richard K.; Jansson, Michael
With the aim of improving the quality of asymptotic distributional approximations for nonlinear functionals of nonparametric estimators, this paper revisits the large-sample properties of an important member of that class, namely a kernel-based weighted average derivative estimator. Asymptotic li...
Bootstrapping Density-Weighted Average Derivatives
Cattaneo, Matias D.; Crump, Richard K.; Jansson, Michael
Employing the "small bandwidth" asymptotic framework of Cattaneo, Crump, and Jansson (2009), this paper studies the properties of a variety of bootstrap-based inference procedures associated with the kernel-based density-weighted averaged derivative estimator proposed by Powell, Stock, and Stoker...
Quantum Averaging of Squeezed States of Light
Squeezing has been recognized as the main resource for quantum information processing and an important resource for beating classical detection strategies. It is therefore of high importance to reliably generate stable squeezing over longer periods of time. The averaging procedure for a single qu...
Bayesian Model Averaging for Propensity Score Analysis
Kaplan, David; Chen, Jianshen
2013-01-01
The purpose of this study is to explore Bayesian model averaging in the propensity score context. Previous research on Bayesian propensity score analysis does not take into account model uncertainty. In this regard, an internally consistent Bayesian framework for model building and estimation must also account for model uncertainty. The…
A dynamic analysis of moving average rules
Chiarella, C.; He, X.Z.; Hommes, C.H.
2006-01-01
The use of various moving average (MA) rules remains popular with financial market practitioners. These rules have recently become the focus of a number empirical studies, but there have been very few studies of financial market models where some agents employ technical trading rules of the type
Average utility maximization: A preference foundation
A.V. Kothiyal (Amit); V. Spinu (Vitalie); P.P. Wakker (Peter)
2014-01-01
textabstractThis paper provides necessary and sufficient preference conditions for average utility maximization over sequences of variable length. We obtain full generality by using a new algebraic technique that exploits the richness structure naturally provided by the variable length of the sequen
High average-power induction linacs
Prono, D.S.; Barrett, D.; Bowles, E.; Caporaso, G.J.; Chen, Yu-Jiuan; Clark, J.C.; Coffield, F.; Newton, M.A.; Nexsen, W.; Ravenscroft, D.
1989-03-15
Induction linear accelerators (LIAs) are inherently capable of accelerating several thousand amperes of /approximately/ 50-ns duration pulses to > 100 MeV. In this paper we report progress and status in the areas of duty factor and stray power management. These technologies are vital if LIAs are to attain high average power operation. 13 figs.
High Average Power Optical FEL Amplifiers
Ben-Zvi, I; Litvinenko, V
2005-01-01
Historically, the first demonstration of the FEL was in an amplifier configuration at Stanford University. There were other notable instances of amplifying a seed laser, such as the LLNL amplifier and the BNL ATF High-Gain Harmonic Generation FEL. However, for the most part FELs are operated as oscillators or self amplified spontaneous emission devices. Yet, in wavelength regimes where a conventional laser seed can be used, the FEL can be used as an amplifier. One promising application is for very high average power generation, for instance a 100 kW average power FEL. The high electron beam power, high brightness and high efficiency that can be achieved with photoinjectors and superconducting energy recovery linacs combine well with the high-gain FEL amplifier to produce unprecedented average power FELs with some advantages. In addition to the general features of the high average power FEL amplifier, we will look at a 100 kW class FEL amplifier is being designed to operate on the 0.5 ampere Energy Recovery Li...
Full averaging of fuzzy impulsive differential inclusions
Natalia V. Skripnik
2010-09-01
Full Text Available In this paper the substantiation of the method of full averaging for fuzzy impulsive differential inclusions is studied. We extend the similar results for impulsive differential inclusions with Hukuhara derivative (Skripnik, 2007, for fuzzy impulsive differential equations (Plotnikov and Skripnik, 2009, and for fuzzy differential inclusions (Skripnik, 2009.
Materials for high average power lasers
Marion, J.E.; Pertica, A.J.
1989-01-01
Unique materials properties requirements for solid state high average power (HAP) lasers dictate a materials development research program. A review of the desirable laser, optical and thermo-mechanical properties for HAP lasers precedes an assessment of the development status for crystalline and glass hosts optimized for HAP lasers. 24 refs., 7 figs., 1 tab.
A dynamic analysis of moving average rules
C. Chiarella; X.Z. He; C.H. Hommes
2006-01-01
The use of various moving average (MA) rules remains popular with financial market practitioners. These rules have recently become the focus of a number empirical studies, but there have been very few studies of financial market models where some agents employ technical trading rules of the type use
19 CFR 101.6 - Hours of business.
2010-04-01
... 19 Customs Duties 1 2010-04-01 2010-04-01 false Hours of business. 101.6 Section 101.6 Customs... GENERAL PROVISIONS § 101.6 Hours of business. Except as specified in paragraphs (a) through (g) of this section, each CBP office shall be open for the transactions of general CBP business between the hours of 8...
"Hour of Code": Can It Change Students' Attitudes toward Programming?
Du, Jie; Wimmer, Hayden; Rada, Roy
2016-01-01
The Hour of Code is a one-hour introduction to computer science organized by Code.org, a non-profit dedicated to expanding participation in computer science. This study investigated the impact of the Hour of Code on students' attitudes towards computer programming and their knowledge of programming. A sample of undergraduate students from two…
Hourly Electricity Prices in Day-Ahead Markets
R. Huisman (Ronald); C. Huurman; R.J. Mahieu (Ronald)
2007-01-01
textabstractThis paper focuses on the characteristics of hourly electricity prices in day-ahead markets. In these markets, quotes for day-ahead delivery of electricity are submitted simultaneously for all hours in the next day. The same information set is used for quoting all hours of the day. The d
48 CFR 16.602 - Labor-hour contracts.
2010-10-01
... METHODS AND CONTRACT TYPES TYPES OF CONTRACTS Time-and-Materials, Labor-Hour, and Letter Contracts 16.602 Labor-hour contracts. Description. A labor-hour contract is a variation of the time-and-materials contract, differing only in that materials are not supplied by the contractor. See 12.207(b),...
Improvements in PIXE analysis of hourly particulate matter samples
Calzolai, G., E-mail: calzolai@fi.infn.it [Department of Physics and Astronomy, University of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); National Institute of Nuclear Physics (INFN), Division of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); Lucarelli, F. [Department of Physics and Astronomy, University of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); National Institute of Nuclear Physics (INFN), Division of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); Chiari, M.; Nava, S. [National Institute of Nuclear Physics (INFN), Division of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); Giannoni, M. [National Institute of Nuclear Physics (INFN), Division of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); Department of Chemistry, University of Florence, Via della Lastruccia 3, 50019 Sesto Fiorentino (Italy); Carraresi, L. [Department of Physics and Astronomy, University of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); National Institute of Nuclear Physics (INFN), Division of Florence, Via G. Sansone 1, 50019 Sesto Fiorentino (Italy); Prati, P. [Department of Physics, University of Genoa and INFN Division of Genoa, Via Dodecaneso 33, 16146 Genoa (Italy); Vecchi, R. [Department of Physics, Università degli Studi di Milano and INFN Division of Milan, Via Celoria 16, 20133 Milan (Italy)
2015-11-15
Most air quality studies on particulate matter (PM) are based on 24-h averaged data; however, many PM emissions as well as their atmospheric dilution processes change within a few hours. Samplings of PM with 1-h resolution can be performed by the streaker sampler (PIXE International Corporation), which is designed to separate the fine (aerodynamic diameter less than 2.5 μm) and the coarse (aerodynamic diameter between 2.5 and 10 μm) fractions of PM. These samples are efficiently analyzed by Particle Induced X-ray Emission (PIXE) at the LABEC laboratory of INFN in Florence (Italy), equipped with a 3 MV Tandetron accelerator, thanks to an optimized external-beam set-up, a convenient choice of the beam energy and suitable collecting substrates. A detailed description of the adopted set-up and results from a methodological study on the detection limits for the selection of the optimal beam energy are shown; the outcomes of the research on alternative collecting substrates, which produce a lower background during the measurements, and with lower contaminations, are also discussed.
Improvements in PIXE analysis of hourly particulate matter samples
Calzolai, G.; Lucarelli, F.; Chiari, M.; Nava, S.; Giannoni, M.; Carraresi, L.; Prati, P.; Vecchi, R.
2015-11-01
Most air quality studies on particulate matter (PM) are based on 24-h averaged data; however, many PM emissions as well as their atmospheric dilution processes change within a few hours. Samplings of PM with 1-h resolution can be performed by the streaker sampler (PIXE International Corporation), which is designed to separate the fine (aerodynamic diameter less than 2.5 μm) and the coarse (aerodynamic diameter between 2.5 and 10 μm) fractions of PM. These samples are efficiently analyzed by Particle Induced X-ray Emission (PIXE) at the LABEC laboratory of INFN in Florence (Italy), equipped with a 3 MV Tandetron accelerator, thanks to an optimized external-beam set-up, a convenient choice of the beam energy and suitable collecting substrates. A detailed description of the adopted set-up and results from a methodological study on the detection limits for the selection of the optimal beam energy are shown; the outcomes of the research on alternative collecting substrates, which produce a lower background during the measurements, and with lower contaminations, are also discussed.
Generalised maximum entropy and heterogeneous technologies
Oude Lansink, A.G.J.M.
1999-01-01
Generalised maximum entropy methods are used to estimate a dual model of production on panel data of Dutch cash crop farms over the period 1970-1992. The generalised maximum entropy approach allows a coherent system of input demand and output supply equations to be estimated for each farm in the sam
20 CFR 229.48 - Family maximum.
2010-04-01
... month on one person's earnings record is limited. This limited amount is called the family maximum. The family maximum used to adjust the social security overall minimum rate is based on the employee's Overall..., when any of the persons entitled to benefits on the insured individual's compensation would, except...
LIANG Hong; ZHANG Renhe; LIU Jingmiao; SUN Zhian; CHENG Xinghong
2012-01-01
In this study,the clear sky hourly global and net solar irradiances at the surface determined using SUNFLUX,a simple parameterization scheme,for three stations (Gaize,Naqu,and Lhasa) on the Tibetan Plateau were evaluated against observation data.Our modeled results agree well with observations.The correlation coefficients between modeled and observed values were ＞0.99 for all three stations.The relative error of modeled results,in average was ＜ 7％,and the root-mean-square variance was ＜ 27 W m-2.The solar irradiances in the radiation model were slightly overestimated compared with observation data;there were at least two likely causes.First,the radiative effects of aerosols were not included in the radiation model.Second,solar irradiances determined by thermopile pyranometers include a thermal offset error that causes solar radiation to be slightly underestimated.The solar radiation absorbed by the ozone and water vapor was estimated. The results show that monthly mean solar radiation absorbed by the ozone is ＜ 2％ of the global solar radiation (＜ 14 W m-2).Solar radiation absorbed by water vapor is stronger in summer than in winter.The maximum amount of monthly mean solar radiation absorbed by water vapor can be up to 13％ of the global solar radiation (95W m-2).This indicates that water vapor measurements with high precision are very important for precise determination of solar radiation.
Thrombectomy within 8 hours after symptom onset in ischemic stroke.
Jovin, Tudor G; Chamorro, Angel; Cobo, Erik; de Miquel, María A; Molina, Carlos A; Rovira, Alex; San Román, Luis; Serena, Joaquín; Abilleira, Sonia; Ribó, Marc; Millán, Mònica; Urra, Xabier; Cardona, Pere; López-Cancio, Elena; Tomasello, Alejandro; Castaño, Carlos; Blasco, Jordi; Aja, Lucía; Dorado, Laura; Quesada, Helena; Rubiera, Marta; Hernandez-Pérez, María; Goyal, Mayank; Demchuk, Andrew M; von Kummer, Rüdiger; Gallofré, Miquel; Dávalos, Antoni
2015-06-11
We aimed to assess the safety and efficacy of thrombectomy for the treatment of stroke in a trial embedded within a population-based stroke reperfusion registry. During a 2-year period at four centers in Catalonia, Spain, we randomly assigned 206 patients who could be treated within 8 hours after the onset of symptoms of acute ischemic stroke to receive either medical therapy (including intravenous alteplase when eligible) and endovascular therapy with the Solitaire stent retriever (thrombectomy group) or medical therapy alone (control group). All patients had confirmed proximal anterior circulation occlusion and the absence of a large infarct on neuroimaging. In all study patients, the use of alteplase either did not achieve revascularization or was contraindicated. The primary outcome was the severity of global disability at 90 days, as measured on the modified Rankin scale (ranging from 0 [no symptoms] to 6 [death]). Although the maximum planned sample size was 690, enrollment was halted early because of loss of equipoise after positive results for thrombectomy were reported from other similar trials. Thrombectomy reduced the severity of disability over the range of the modified Rankin scale (adjusted odds ratio for improvement of 1 point, 1.7; 95% confidence interval [CI], 1.05 to 2.8) and led to higher rates of functional independence (a score of 0 to 2) at 90 days (43.7% vs. 28.2%; adjusted odds ratio, 2.1; 95% CI, 1.1 to 4.0). At 90 days, the rates of symptomatic intracranial hemorrhage were 1.9% in both the thrombectomy group and the control group (P=1.00), and rates of death were 18.4% and 15.5%, respectively (P=0.60). Registry data indicated that only eight patients who met the eligibility criteria were treated outside the trial at participating hospitals. Among patients with anterior circulation stroke who could be treated within 8 hours after symptom onset, stent retriever thrombectomy reduced the severity of post-stroke disability and increased the
Kos, Bor; Valič, Blaž; Kotnik, Tadej; Gajšek, Peter
2012-10-07
Induction heating equipment is a source of strong and nonhomogeneous magnetic fields, which can exceed occupational reference levels. We investigated a case of an induction tempering tunnel furnace. Measurements of the emitted magnetic flux density (B) were performed during its operation and used to validate a numerical model of the furnace. This model was used to compute the values of B and the induced in situ electric field (E) for 15 different body positions relative to the source. For each body position, the computed B values were used to determine their maximum and average values, using six spatial averaging schemes (9-285 averaging points) and two averaging algorithms (arithmetic mean and quadratic mean). Maximum and average B values were compared to the ICNIRP reference level, and E values to the ICNIRP basic restriction. Our results show that in nonhomogeneous fields, the maximum B is an overly conservative predictor of overexposure, as it yields many false positives. The average B yielded fewer false positives, but as the number of averaging points increased, false negatives emerged. The most reliable averaging schemes were obtained for averaging over the torso with quadratic averaging, with no false negatives even for the maximum number of averaging points investigated.
Duality of Maximum Entropy and Minimum Divergence
Shinto Eguchi
2014-06-01
Full Text Available We discuss a special class of generalized divergence measures by the use of generator functions. Any divergence measure in the class is separated into the difference between cross and diagonal entropy. The diagonal entropy measure in the class associates with a model of maximum entropy distributions; the divergence measure leads to statistical estimation via minimization, for arbitrarily giving a statistical model. The dualistic relationship between the maximum entropy model and the minimum divergence estimation is explored in the framework of information geometry. The model of maximum entropy distributions is characterized to be totally geodesic with respect to the linear connection associated with the divergence. A natural extension for the classical theory for the maximum likelihood method under the maximum entropy model in terms of the Boltzmann-Gibbs-Shannon entropy is given. We discuss the duality in detail for Tsallis entropy as a typical example.
Maximum power analysis of photovoltaic module in Ramadi city
Shahatha Salim, Majid; Mohammed Najim, Jassim [College of Science, University of Anbar (Iraq); Mohammed Salih, Salih [Renewable Energy Research Center, University of Anbar (Iraq)
2013-07-01
Performance of photovoltaic (PV) module is greatly dependent on the solar irradiance, operating temperature, and shading. Solar irradiance can have a significant impact on power output of PV module and energy yield. In this paper, a maximum PV power which can be obtain in Ramadi city (100km west of Baghdad) is practically analyzed. The analysis is based on real irradiance values obtained as the first time by using Soly2 sun tracker device. Proper and adequate information on solar radiation and its components at a given location is very essential in the design of solar energy systems. The solar irradiance data in Ramadi city were analyzed based on the first three months of 2013. The solar irradiance data are measured on earth's surface in the campus area of Anbar University. Actual average data readings were taken from the data logger of sun tracker system, which sets to save the average readings for each two minutes and based on reading in each one second. The data are analyzed from January to the end of March-2013. Maximum daily readings and monthly average readings of solar irradiance have been analyzed to optimize the output of photovoltaic solar modules. The results show that the system sizing of PV can be reduced by 12.5% if a tracking system is used instead of fixed orientation of PV modules.
Maximum power analysis of photovoltaic module in Ramadi city
Majid Shahatha Salim, Jassim Mohammed Najim, Salih Mohammed Salih
2013-01-01
Full Text Available Performance of photovoltaic (PV module is greatly dependent on the solar irradiance, operating temperature, and shading. Solar irradiance can have a significant impact on power output of PV module and energy yield. In this paper, a maximum PV power which can be obtain in Ramadi city (100km west of Baghdad is practically analyzed. The analysis is based on real irradiance values obtained as the first time by using Soly2 sun tracker device. Proper and adequate information on solar radiation and its components at a given location is very essential in the design of solar energy systems. The solar irradiance data in Ramadi city were analyzed based on the first three months of 2013. The solar irradiance data are measured on earth's surface in the campus area of Anbar University. Actual average data readings were taken from the data logger of sun tracker system, which sets to save the average readings for each two minutes and based on reading in each one second. The data are analyzed from January to the end of March-2013. Maximum daily readings and monthly average readings of solar irradiance have been analyzed to optimize the output of photovoltaic solar modules. The results show that the system sizing of PV can be reduced by 12.5% if a tracking system is used instead of fixed orientation of PV modules.
Campbell, Bruce C V; Tu, Hans T H; Christensen, Søren; Desmond, Patricia M; Levi, Christopher R; Bladin, Christopher F; Hjort, Niels; Ashkanian, Mahmoud; Sølling, Christine; Donnan, Geoffrey A; Davis, Stephen M; Ostergaard, Leif; Parsons, Mark W
2012-01-01
Imaging is used as a surrogate for clinical outcome in early-phase stroke trials. Assessment of infarct growth earlier than the standard 90 days used for clinical end points may be equally accurate and more practical. To compare assessment of the effect of reperfusion therapies using 24-hour vs day 90 magnetic resonance imaging. Infarct volume was assessed on diffusion-weighted imaging (DWI) at baseline and 24 hours after stroke onset and on fluid-attenuated inversion recovery images at day 90. The DWI and fluid-attenuated inversion recovery lesions were manually outlined by 2 independent raters, and the volumes were averaged. Interrater consistency was assessed using the median difference in lesion volume between raters. Referral center. Patients Imaging data were available for 83 patients; 77 of these patients received thrombolysis. Infarct volume at 24 hours and 90 days. The 24-hour DWI infarct volume had a strong linear correlation with day 90 fluid-attenuated inversion recovery infarct volume (r = 0.98, 95% confidence interval, 0.97-0.99). Recanalization had a significant effect on infarct evolution between baseline and 24 hours but not between 24 hours and day 90. Infarct growth from baseline was significantly reduced by recanalization, whether assessed at 24 hours or day 90. Infarct volume at either time point predicted functional outcome independent of age and baseline stroke severity. Interrater agreement was better for DWI than fluid-attenuated inversion recovery (1.4 mL [8%] vs 1.8 mL [17%]; P = .002). Assessment of final infarct volume using DWI at 24 hours captures the effect of reperfusion therapies on infarct growth and predicts functional outcome similarly to imaging at day 90. This has the potential to reduce loss to follow-up in trials and may add early prognostic information in clinical practice.
2,445 Hours of Code: What I Learned from Facilitating Hour of Code Events in High School Libraries
Colby, Jennifer
2015-01-01
This article describes a school librarian's experience with initiating an Hour of Code event for her school's student body. Hadi Partovi of Code.org conceived the Hour of Code "to get ten million students to try one hour of computer science" (Partovi, 2013a), which is implemented during Computer Science Education Week with a goal of…
Averaged Extended Tree Augmented Naive Classifier
Aaron Meehan
2015-07-01
Full Text Available This work presents a new general purpose classifier named Averaged Extended Tree Augmented Naive Bayes (AETAN, which is based on combining the advantageous characteristics of Extended Tree Augmented Naive Bayes (ETAN and Averaged One-Dependence Estimator (AODE classifiers. We describe the main properties of the approach and algorithms for learning it, along with an analysis of its computational time complexity. Empirical results with numerous data sets indicate that the new approach is superior to ETAN and AODE in terms of both zero-one classification accuracy and log loss. It also compares favourably against weighted AODE and hidden Naive Bayes. The learning phase of the new approach is slower than that of its competitors, while the time complexity for the testing phase is similar. Such characteristics suggest that the new classifier is ideal in scenarios where online learning is not required.
ANALYSIS OF THE FACTORS AFFECTING THE AVERAGE
Carmen BOGHEAN
2013-12-01
Full Text Available Productivity in agriculture most relevantly and concisely expresses the economic efficiency of using the factors of production. Labour productivity is affected by a considerable number of variables (including the relationship system and interdependence between factors, which differ in each economic sector and influence it, giving rise to a series of technical, economic and organizational idiosyncrasies. The purpose of this paper is to analyse the underlying factors of the average work productivity in agriculture, forestry and fishing. The analysis will take into account the data concerning the economically active population and the gross added value in agriculture, forestry and fishing in Romania during 2008-2011. The distribution of the average work productivity per factors affecting it is conducted by means of the u-substitution method.
Time-average dynamic speckle interferometry
Vladimirov, A. P.
2014-05-01
For the study of microscopic processes occurring at structural level in solids and thin biological objects, a method of dynamic speckle interferometry successfully applied. However, the method has disadvantages. The purpose of the report is to acquaint colleagues with the method of averaging in time in dynamic speckle - interferometry of microscopic processes, allowing eliminating shortcomings. The main idea of the method is the choice the averaging time, which exceeds the characteristic time correlation (relaxation) the most rapid process. The method theory for a thin phase and the reflecting object is given. The results of the experiment on the high-cycle fatigue of steel and experiment to estimate the biological activity of a monolayer of cells, cultivated on a transparent substrate is given. It is shown that the method allows real-time visualize the accumulation of fatigue damages and reliably estimate the activity of cells with viruses and without viruses.
Effects of bruxism on the maximum bite force
Todić Jelena T.
2017-01-01
Full Text Available Background/Aim. Bruxism is a parafunctional activity of the masticatory system, which is characterized by clenching or grinding of teeth. The purpose of this study was to determine whether the presence of bruxism has impact on maximum bite force, with particular reference to the potential impact of gender on bite force values. Methods. This study included two groups of subjects: without and with bruxism. The presence of bruxism in the subjects was registered using a specific clinical questionnaire on bruxism and physical examination. The subjects from both groups were submitted to the procedure of measuring the maximum bite pressure and occlusal contact area using a single-sheet pressure-sensitive films (Fuji Prescale MS and HS Film. Maximal bite force was obtained by multiplying maximal bite pressure and occlusal contact area values. Results. The average values of maximal bite force were significantly higher in the subjects with bruxism compared to those without bruxism (p 0.01. Maximal bite force was significantly higher in the males compared to the females in all segments of the research. Conclusion. The presence of bruxism influences the increase in the maximum bite force as shown in this study. Gender is a significant determinant of bite force. Registration of maximum bite force can be used in diagnosing and analysing pathophysiological events during bruxism.
Average Annual Rainfall over the Globe
Agrawal, D. C.
2013-01-01
The atmospheric recycling of water is a very important phenomenon on the globe because it not only refreshes the water but it also redistributes it over land and oceans/rivers/lakes throughout the globe. This is made possible by the solar energy intercepted by the Earth. The half of the globe facing the Sun, on the average, intercepts 1.74 ×…
Endogenous average cost based access pricing
Fjell, Kenneth; Foros, Øystein; Pal, Debashis
2006-01-01
We consider an industry where a downstream competitor requires access to an upstream facility controlled by a vertically integrated and regulated incumbent. The literature on access pricing assumes the access price to be exogenously fixed ex-ante. We analyze an endogenous average cost based access pricing rule, where both firms realize the interdependence among their quantities and the regulated access price. Endogenous access pricing neutralizes the artificial cost advantag...
The Ghirlanda-Guerra identities without averaging
Chatterjee, Sourav
2009-01-01
The Ghirlanda-Guerra identities are one of the most mysterious features of spin glasses. We prove the GG identities in a large class of models that includes the Edwards-Anderson model, the random field Ising model, and the Sherrington-Kirkpatrick model in the presence of a random external field. Previously, the GG identities were rigorously proved only `on average' over a range of temperatures or under small perturbations.
Average Annual Rainfall over the Globe
Agrawal, D. C.
2013-01-01
The atmospheric recycling of water is a very important phenomenon on the globe because it not only refreshes the water but it also redistributes it over land and oceans/rivers/lakes throughout the globe. This is made possible by the solar energy intercepted by the Earth. The half of the globe facing the Sun, on the average, intercepts 1.74 ×…
Average Light Intensity Inside a Photobioreactor
Herby Jean
2011-01-01
Full Text Available For energy production, microalgae are one of the few alternatives with high potential. Similar to plants, algae require energy acquired from light sources to grow. This project uses calculus to determine the light intensity inside of a photobioreactor filled with algae. Under preset conditions along with estimated values, we applied Lambert-Beer's law to formulate an equation to calculate how much light intensity escapes a photobioreactor and determine the average light intensity that was present inside the reactor.
Geomagnetic effects on the average surface temperature
Ballatore, P.
Several results have previously shown as the solar activity can be related to the cloudiness and the surface solar radiation intensity (Svensmark and Friis-Christensen, J. Atmos. Sol. Terr. Phys., 59, 1225, 1997; Veretenenkoand Pudovkin, J. Atmos. Sol. Terr. Phys., 61, 521, 1999). Here, the possible relationships between the averaged surface temperature and the solar wind parameters or geomagnetic activity indices are investigated. The temperature data used are the monthly SST maps (generated at RAL and available from the related ESRIN/ESA database) that represent the averaged surface temperature with a spatial resolution of 0.5°x0.5° and cover the entire globe. The interplanetary data and the geomagnetic data are from the USA National Space Science Data Center. The time interval considered is 1995-2000. Specifically, possible associations and/or correlations of the average temperature with the interplanetary magnetic field Bz component and with the Kp index are considered and differentiated taking into account separate geographic and geomagnetic planetary regions.
Unscrambling The "Average User" Of Habbo Hotel
Mikael Johnson
2007-01-01
Full Text Available The “user” is an ambiguous concept in human-computer interaction and information systems. Analyses of users as social actors, participants, or configured users delineate approaches to studying design-use relationships. Here, a developer’s reference to a figure of speech, termed the “average user,” is contrasted with design guidelines. The aim is to create an understanding about categorization practices in design through a case study about the virtual community, Habbo Hotel. A qualitative analysis highlighted not only the meaning of the “average user,” but also the work that both the developer and the category contribute to this meaning. The average user a represents the unknown, b influences the boundaries of the target user groups, c legitimizes the designer to disregard marginal user feedback, and d keeps the design space open, thus allowing for creativity. The analysis shows how design and use are intertwined and highlights the developers’ role in governing different users’ interests.
On Backus average for generally anisotropic layers
Bos, Len; Slawinski, Michael A; Stanoev, Theodore
2016-01-01
In this paper, following the Backus (1962) approach, we examine expressions for elasticity parameters of a homogeneous generally anisotropic medium that is long-wave-equivalent to a stack of thin generally anisotropic layers. These expressions reduce to the results of Backus (1962) for the case of isotropic and transversely isotropic layers. In over half-a-century since the publications of Backus (1962) there have been numerous publications applying and extending that formulation. However, neither George Backus nor the authors of the present paper are aware of further examinations of mathematical underpinnings of the original formulation; hence, this paper. We prove that---within the long-wave approximation---if the thin layers obey stability conditions then so does the equivalent medium. We examine---within the Backus-average context---the approximation of the average of a product as the product of averages, and express it as a proposition in terms of an upper bound. In the presented examination we use the e...
A simple algorithm for averaging spike trains.
Julienne, Hannah; Houghton, Conor
2013-02-25
Although spike trains are the principal channel of communication between neurons, a single stimulus will elicit different spike trains from trial to trial. This variability, in both spike timings and spike number can obscure the temporal structure of spike trains and often means that computations need to be run on numerous spike trains in order to extract features common across all the responses to a particular stimulus. This can increase the computational burden and obscure analytical results. As a consequence, it is useful to consider how to calculate a central spike train that summarizes a set of trials. Indeed, averaging responses over trials is routine for other signal types. Here, a simple method for finding a central spike train is described. The spike trains are first mapped to functions, these functions are averaged, and a greedy algorithm is then used to map the average function back to a spike train. The central spike trains are tested for a large data set. Their performance on a classification-based test is considerably better than the performance of the medoid spike trains.
Changing mortality and average cohort life expectancy
Robert Schoen
2005-10-01
Full Text Available Period life expectancy varies with changes in mortality, and should not be confused with the life expectancy of those alive during that period. Given past and likely future mortality changes, a recent debate has arisen on the usefulness of the period life expectancy as the leading measure of survivorship. An alternative aggregate measure of period mortality which has been seen as less sensitive to period changes, the cross-sectional average length of life (CAL has been proposed as an alternative, but has received only limited empirical or analytical examination. Here, we introduce a new measure, the average cohort life expectancy (ACLE, to provide a precise measure of the average length of life of cohorts alive at a given time. To compare the performance of ACLE with CAL and with period and cohort life expectancy, we first use population models with changing mortality. Then the four aggregate measures of mortality are calculated for England and Wales, Norway, and Switzerland for the years 1880 to 2000. CAL is found to be sensitive to past and present changes in death rates. ACLE requires the most data, but gives the best representation of the survivorship of cohorts present at a given time.
Disk-averaged synthetic spectra of Mars
Tinetti, Giovanna; Meadows, Victoria S.; Crisp, David; Fong, William; Velusamy, Thangasamy; Snively, Heather
2005-01-01
The principal goal of the NASA Terrestrial Planet Finder (TPF) and European Space Agency's Darwin mission concepts is to directly detect and characterize extrasolar terrestrial (Earthsized) planets. This first generation of instruments is expected to provide disk-averaged spectra with modest spectral resolution and signal-to-noise. Here we use a spatially and spectrally resolved model of a Mars-like planet to study the detectability of a planet's surface and atmospheric properties from disk-averaged spectra. We explore the detectability as a function of spectral resolution and wavelength range, for both the proposed visible coronograph (TPFC) and mid-infrared interferometer (TPF-I/Darwin) architectures. At the core of our model is a spectrum-resolving (line-by-line) atmospheric/surface radiative transfer model. This model uses observational data as input to generate a database of spatially resolved synthetic spectra for a range of illumination conditions and viewing geometries. The model was validated against spectra recorded by the Mars Global Surveyor-Thermal Emission Spectrometer and the Mariner 9-Infrared Interferometer Spectrometer. Results presented here include disk-averaged synthetic spectra, light curves, and the spectral variability at visible and mid-infrared wavelengths for Mars as a function of viewing angle, illumination, and season. We also considered the differences in the spectral appearance of an increasingly ice-covered Mars, as a function of spectral resolution, signal-to-noise and integration time for both TPF-C and TPFI/ Darwin.
Spatial averaging infiltration model for layered soil
HU HePing; YANG ZhiYong; TIAN FuQiang
2009-01-01
To quantify the influences of soil heterogeneity on infiltration, a spatial averaging infiltration model for layered soil (SAI model) is developed by coupling the spatial averaging approach proposed by Chen et al. and the Generalized Green-Ampt model proposed by Jia et al. In the SAI model, the spatial heterogeneity along the horizontal direction is described by a probability distribution function, while that along the vertical direction is represented by the layered soils. The SAI model is tested on a typical soil using Monte Carlo simulations as the base model. The results show that the SAI model can directly incorporate the influence of spatial heterogeneity on infiltration on the macro scale. It is also found that the homogeneous assumption of soil hydraulic conductivity along the horizontal direction will overestimate the infiltration rate, while that along the vertical direction will underestimate the infiltration rate significantly during rainstorm periods. The SAI model is adopted in the spatial averaging hydrological model developed by the authors, and the results prove that it can be applied in the macro-scale hydrological and land surface process modeling in a promising way.
Spatial averaging infiltration model for layered soil
无
2009-01-01
To quantify the influences of soil heterogeneity on infiltration, a spatial averaging infiltration model for layered soil (SAI model) is developed by coupling the spatial averaging approach proposed by Chen et al. and the Generalized Green-Ampt model proposed by Jia et al. In the SAI model, the spatial hetero- geneity along the horizontal direction is described by a probability distribution function, while that along the vertical direction is represented by the layered soils. The SAI model is tested on a typical soil using Monte Carlo simulations as the base model. The results show that the SAI model can directly incorporate the influence of spatial heterogeneity on infiltration on the macro scale. It is also found that the homogeneous assumption of soil hydraulic conductivity along the horizontal direction will overes- timate the infiltration rate, while that along the vertical direction will underestimate the infiltration rate significantly during rainstorm periods. The SAI model is adopted in the spatial averaging hydrological model developed by the authors, and the results prove that it can be applied in the macro-scale hy- drological and land surface process modeling in a promising way.
Disk-averaged synthetic spectra of Mars
Tinetti, G; Fong, W; Meadows, V S; Snively, H; Velusamy, T; Crisp, David; Fong, William; Meadows, Victoria S.; Snively, Heather; Tinetti, Giovanna; Velusamy, Thangasamy
2004-01-01
The principal goal of the NASA Terrestrial Planet Finder (TPF) and ESA Darwin mission concepts is to directly detect and characterize extrasolar terrestrial (Earth-sized) planets. This first generation of instruments is expected to provide disk-averaged spectra with modest spectral resolution and signal-to-noise. Here we use a spatially and spectrally resolved model of the planet Mars to study the detectability of a planet's surface and atmospheric properties from disk-averaged spectra as a function of spectral resolution and wavelength range, for both the proposed visible coronograph (TPF-C) and mid-infrared interferometer (TPF-I/Darwin) architectures. At the core of our model is a spectrum-resolving (line-by-line) atmospheric/surface radiative transfer model which uses observational data as input to generate a database of spatially-resolved synthetic spectra for a range of illumination conditions (phase angles) and viewing geometries. Results presented here include disk averaged synthetic spectra, light-cur...
Disk-averaged synthetic spectra of Mars
Tinetti, Giovanna; Meadows, Victoria S.; Crisp, David; Fong, William; Velusamy, Thangasamy; Snively, Heather
2005-01-01
The principal goal of the NASA Terrestrial Planet Finder (TPF) and European Space Agency's Darwin mission concepts is to directly detect and characterize extrasolar terrestrial (Earthsized) planets. This first generation of instruments is expected to provide disk-averaged spectra with modest spectral resolution and signal-to-noise. Here we use a spatially and spectrally resolved model of a Mars-like planet to study the detectability of a planet's surface and atmospheric properties from disk-averaged spectra. We explore the detectability as a function of spectral resolution and wavelength range, for both the proposed visible coronograph (TPFC) and mid-infrared interferometer (TPF-I/Darwin) architectures. At the core of our model is a spectrum-resolving (line-by-line) atmospheric/surface radiative transfer model. This model uses observational data as input to generate a database of spatially resolved synthetic spectra for a range of illumination conditions and viewing geometries. The model was validated against spectra recorded by the Mars Global Surveyor-Thermal Emission Spectrometer and the Mariner 9-Infrared Interferometer Spectrometer. Results presented here include disk-averaged synthetic spectra, light curves, and the spectral variability at visible and mid-infrared wavelengths for Mars as a function of viewing angle, illumination, and season. We also considered the differences in the spectral appearance of an increasingly ice-covered Mars, as a function of spectral resolution, signal-to-noise and integration time for both TPF-C and TPFI/ Darwin.
Adaptive Parallel Tempering for Stochastic Maximum Likelihood Learning of RBMs
Desjardins, Guillaume; Bengio, Yoshua
2010-01-01
Restricted Boltzmann Machines (RBM) have attracted a lot of attention of late, as one the principle building blocks of deep networks. Training RBMs remains problematic however, because of the intractibility of their partition function. The maximum likelihood gradient requires a very robust sampler which can accurately sample from the model despite the loss of ergodicity often incurred during learning. While using Parallel Tempering in the negative phase of Stochastic Maximum Likelihood (SML-PT) helps address the issue, it imposes a trade-off between computational complexity and high ergodicity, and requires careful hand-tuning of the temperatures. In this paper, we show that this trade-off is unnecessary. The choice of optimal temperatures can be automated by minimizing average return time (a concept first proposed by [Katzgraber et al., 2006]) while chains can be spawned dynamically, as needed, thus minimizing the computational overhead. We show on a synthetic dataset, that this results in better likelihood ...
High-resolution grids of hourly meteorological variables for Germany
Krähenmann, S.; Walter, A.; Brienen, S.; Imbery, F.; Matzarakis, A.
2016-12-01
We present a 1-km2 gridded German dataset of hourly surface climate variables covering the period 1995 to 2012. The dataset comprises 12 variables including temperature, dew point, cloud cover, wind speed and direction, global and direct shortwave radiation, down- and up-welling longwave radiation, sea level pressure, relative humidity and vapour pressure. This dataset was constructed statistically from station data, satellite observations and model data. It is outstanding in terms of spatial and temporal resolution and in the number of climate variables. For each variable, we employed the most suitable gridding method and combined the best of several information sources, including station records, satellite-derived data and data from a regional climate model. A module to estimate urban heat island intensity was integrated for air and dew point temperature. Owing to the low density of available synop stations, the gridded dataset does not capture all variations that may occur at a resolution of 1 km2. This applies to areas of complex terrain (all the variables), and in particular to wind speed and the radiation parameters. To achieve maximum precision, we used all observational information when it was available. This, however, leads to inhomogeneities in station network density and affects the long-term consistency of the dataset. A first climate analysis for Germany was conducted. The Rhine River Valley, for example, exhibited more than 100 summer days in 2003, whereas in 1996, the number was low everywhere in Germany. The dataset is useful for applications in various climate-related studies, hazard management and for solar or wind energy applications and it is available via doi: 10.5676/DWD_CDC/TRY_Basis_v001.
Resident education in developmental behavioral pediatrics: the impact of the 80-hour work week.
D'Sa, Viren A; High, Pamela C
2010-04-01
This study assessed the structure of pediatric resident developmental behavioral pediatrics (DBP) rotations in the context of the Accreditation Council for Graduate Medical Education duty hour mandates. : We distributed an online survey addressing rotation structure, call schedule, and impact of the duty hour policy to resident DBP rotation directors in 114 of 204 pediatric residency programs in the United States and Canada and received responses from 81 programs (71% response rate). Seventy-five percent of respondents reported an average of 16% reduction in their DBP rotation after implementation of the Duty Hours rule in 2003. More programs having overnight calls during the rotation reported decreases versus those without overnight calls (91% vs 52%, p rotations. Attention may need to be given to the consequences for resident education in DBP.
Analytic continuation average spectrum method for transport in quantum liquids
Kletenik-Edelman, Orly [School of Chemistry, Sackler Faculty of Exact Sciences, Tel Aviv University, Tel Aviv 69978 (Israel); Rabani, Eran, E-mail: rabani@tau.ac.il [School of Chemistry, Sackler Faculty of Exact Sciences, Tel Aviv University, Tel Aviv 69978 (Israel); Reichman, David R. [Department of Chemistry, Columbia University, 3000 Broadway, New York, NY 10027 (United States)
2010-05-12
Recently, we have applied the analytic continuation averaged spectrum method (ASM) to calculate collective density fluctuations in quantum liquid . Unlike the maximum entropy (MaxEnt) method, the ASM approach is capable of revealing resolved modes in the dynamic structure factor in agreement with experiments. In this work we further develop the ASM to study single-particle dynamics in quantum liquids with dynamical susceptibilities that are characterized by a smooth spectrum. Surprisingly, we find that for the power spectrum of the velocity autocorrelation function there are pronounced differences in comparison with the MaxEnt approach, even for this simple case of smooth unimodal dynamic response. We show that for liquid para-hydrogen the ASM is closer to the centroid molecular dynamics (CMD) result while for normal liquid helium it agrees better with the quantum mode coupling theory (QMCT) and with the MaxEnt approach.
Extracting Credible Dependencies for Averaged One-Dependence Estimator Analysis
LiMin Wang
2014-01-01
Full Text Available Of the numerous proposals to improve the accuracy of naive Bayes (NB by weakening the conditional independence assumption, averaged one-dependence estimator (AODE demonstrates remarkable zero-one loss performance. However, indiscriminate superparent attributes will bring both considerable computational cost and negative effect on classification accuracy. In this paper, to extract the most credible dependencies we present a new type of seminaive Bayesian operation, which selects superparent attributes by building maximum weighted spanning tree and removes highly correlated children attributes by functional dependency and canonical cover analysis. Our extensive experimental comparison on UCI data sets shows that this operation efficiently identifies possible superparent attributes at training time and eliminates redundant children attributes at classification time.
Analysis of 72-hour sterility of common pediatric continuous intravenous infusions.
Piro, Christina C; Davis, Jennifer; Frelix, Arlesia; Grisso, Alison G; Sinclair-Pingel, Julie; Willingham, Harold; Wright, Lorianne; Potts, Amy L
2009-01-01
Patient morbidity and mortality associated with contaminated and improperly prepared sterile products has captured national attention. In response, both the United States Pharmacopeia (USP) and Centers for Disease Control (CDC) have published recommendations in an effort to minimize the risk of infection. While the CDC recommends that administration sets are not changed more frequently than every 72 hours, the USP recommends a maximum beyond use date of 48 hours. Neither organization provides specific guidance on expiration dating once the intravenous drug is dispensed. Likewise, neither addresses the length of time that a bag containing medication for continuous infusion may hang once administration to the patient has begun. We evaluated the sterility of medications that are commonly administered by continuous infusion to pediatric patients. Because frequent manipulation of infusion and administration sets may predispose the patient to adverse events, we evaluated sterility for extended beyond use dating up to 72 hours. Thirty-five common intravenous (IV) continuous infusions using 94 standard concentrations and diluents were identified. IV solutions were mixed using sterile technique in the laminar flow hood in accordance with USP guidelines. Medications were excluded for short stability, short durations of use or high cost. A sample from each solution was tested for contamination or bacterial growth at 72 hours. Any visible discoloration suggesting physical instability was also evaluated. None of the syringes or chambers resulted in contamination, bacterial growth or discoloration after 72 hours. This study provides sufficient data that these compounded sterile products may be stored using a beyond use date up to 72 hours for a number of commonly used continuous IV infusions in pediatric patients. In our institution, this allows for a more convenient and consistent change of both administration sets and continuous infusions at 72 hours to potentially minimize
The maximum intelligible range of the human voice
Boren, Braxton
This dissertation examines the acoustics of the spoken voice at high levels and the maximum number of people that could hear such a voice unamplified in the open air. In particular, it examines an early auditory experiment by Benjamin Franklin which sought to determine the maximum intelligible crowd for the Anglican preacher George Whitefield in the eighteenth century. Using Franklin's description of the experiment and a noise source on Front Street, the geometry and diffraction effects of such a noise source are examined to more precisely pinpoint Franklin's position when Whitefield's voice ceased to be intelligible. Based on historical maps, drawings, and prints, the geometry and material of Market Street is constructed as a computer model which is then used to construct an acoustic cone tracing model. Based on minimal values of the Speech Transmission Index (STI) at Franklin's position, Whitefield's on-axis Sound Pressure Level (SPL) at 1 m is determined, leading to estimates centering around 90 dBA. Recordings are carried out on trained actors and singers to determine their maximum time-averaged SPL at 1 m. This suggests that the greatest average SPL achievable by the human voice is 90-91 dBA, similar to the median estimates for Whitefield's voice. The sites of Whitefield's largest crowds are acoustically modeled based on historical evidence and maps. Based on Whitefield's SPL, the minimal STI value, and the crowd's background noise, this allows a prediction of the minimally intelligible area for each site. These yield maximum crowd estimates of 50,000 under ideal conditions, while crowds of 20,000 to 30,000 seem more reasonable when the crowd was reasonably quiet and Whitefield's voice was near 90 dBA.
Rouzi, Abdulrahim A; Alsahly, Nora; Alamoudi, Rana; Almansouri, Nisma; Alsinani, Nawal; Alkafy, Souzan; Rozzah, Rayyan; Abduljabbar, Hassan
2017-04-01
Misoprostol is an effective agent for the induction of labor. Existing guidelines recommend oral misoprostol solution 25 μg every 2 hours. However, more research is required to optimize the use of oral misoprostol solution for the induction of labor. The purpose of this study was to compare efficacy and safety of hourly titrated-dose oral misoprostol solution with static-dose oral misoprostol solution every 2 hours for labor induction. In this randomized controlled study, oral misoprostol solution was administered as (1) 20 μg hourly (≤4 doses) that was increased in the absence of regular uterine contractions to 40 μg hourly (≤4 doses) and then to 60 μg hourly (≤16 doses) or (2) 25 μg every 2 hours until active labor began (≤12 doses). A sample size of 146 women was planned with the use of a projected 95% rate for the primary endpoint (vaginal delivery within 24 hours) for hourly titrated-dose misoprostol and 80% rate for static-dose misoprostol every 2 hours. Safety outcomes included maternal morbidity and adverse neonatal outcomes. From December 2013 to July 2015, 146 women were assigned randomly to treatment. Demographic and clinical factors were similar between groups, except for age. Vaginal delivery was achieved within 24 hours in 47 women (64.4%) who received hourly titrated-doses of misoprostol solution and 48 women (65.8%) who received 2-hourly static-dose misoprostol solution (P=1.00). Rates of vaginal delivery within 24 hours did not differ significantly between treatment groups for women who were nulliparous (P=1.00) or who had postterm pregnancies (P=.66), a Bishop score of ≤3 (P=.84), or oxytocin augmentation (P=.83). Cesarean deliveries were performed within 24 hours in 9 women who received hourly titrated-dose misoprostol solution and 2 women who received 2-hourly static-dose misoprostol solution (P=.056). Pyrexia and meconium-stained liquor occurred more frequently with the hourly titrated-dose regimen. The static-dose oral
Hourly Wind Speed Interval Prediction in Arid Regions
Chaouch, M.; Ouarda, T.
2013-12-01
The long and extended warm and dry summers, the low rate of rain and humidity are the main factors that explain the increase of electricity consumption in hot arid regions. In such regions, the ventilating and air-conditioning installations, that are typically the most energy-intensive among energy consumption activities, are essential for securing healthy, safe and suitable indoor thermal conditions for building occupants and stored materials. The use of renewable energy resources such as solar and wind represents one of the most relevant solutions to overcome the increase of the electricity demand challenge. In the recent years, wind energy is gaining more importance among the researchers worldwide. Wind energy is intermittent in nature and hence the power system scheduling and dynamic control of wind turbine requires an estimate of wind energy. Accurate forecast of wind speed is a challenging task for the wind energy research field. In fact, due to the large variability of wind speed caused by the unpredictable and dynamic nature of the earth's atmosphere, there are many fluctuations in wind power production. This inherent variability of wind speed is the main cause of the uncertainty observed in wind power generation. Furthermore, producing wind power forecasts might be obtained indirectly by modeling the wind speed series and then transforming the forecasts through a power curve. Wind speed forecasting techniques have received substantial attention recently and several models have been developed. Basically two main approaches have been proposed in the literature: (1) physical models such as Numerical Weather Forecast and (2) statistical models such as Autoregressive integrated moving average (ARIMA) models, Neural Networks. While the initial focus in the literature has been on point forecasts, the need to quantify forecast uncertainty and communicate the risk of extreme ramp events has led to an interest in producing probabilistic forecasts. In short term
Mapping the MPM maximum flow algorithm on GPUs
Solomon, Steven; Thulasiraman, Parimala
2010-11-01
The GPU offers a high degree of parallelism and computational power that developers can exploit for general purpose parallel applications. As a result, a significant level of interest has been directed towards GPUs in recent years. Regular applications, however, have traditionally been the focus of work on the GPU. Only very recently has there been a growing number of works exploring the potential of irregular applications on the GPU. We present a work that investigates the feasibility of Malhotra, Pramodh Kumar and Maheshwari's "MPM" maximum flow algorithm on the GPU that achieves an average speedup of 8 when compared to a sequential CPU implementation.
Modified Weighting for Calculating the Average Concentration of Non-Point Source Pollutant
牟瑞芳
2004-01-01
The concentration of runoff depends upon that of soil loss and the latter is assumed to be linear to the value of EI that equals the product of total storm energy E times the maximum 30-min intensity I30 for a given rainstorm. Usually, the maximum accumulative amount of rain for a rainstorm might bring on the maximum amount of runoff, but it does not equal the maximum erosion and not always lead the maximum concentration. Thus, the average concentration weighted by amount of runoff is somewhat unreasonable. An improvement for the calculation method of non-point source pollution load put forward by professor Li Huaien is proposed. In replacement of the weight of runoff, EI value of a single rainstorm is introduced as a new weight. An example of Fujing River watershed shows that its application is effective.
A step towards using urinary nitrogen as a tool for validating 24-hour dietary recall interviews.
Kahn, H A
1987-09-01
In an active population, can urinary nitrogen output provide a satisfactory basis for estimating protein intake without preliminary screening to rule out metabolic disease? A successful demonstration would justify comparing average protein intake for active groups derived from 24-hour recall interviews with that obtained from 24-hour urine specimens. This could then be a practical test of the hypothesis that 24-hour recall data provide valid estimates for groups. Volunteers (45) from a university population limited their diet for four days to a commercially prepared liquid diet and provided a 24-hour urine specimen beginning on the morning of the fourth day. For men, average daily intake of protein was 112.8 grams and the estimated intake based on urinary nitrogen was 114.5. Corresponding data for women were 84.6 and 83.9. Screening for metabolic disease seems unnecessary but persons restricting their intake to less than their caloric needs will have their protein intake overestimated on the basis of urinary nitrogen measurement.
A dual method for maximum entropy restoration
Smith, C. B.
1979-01-01
A simple iterative dual algorithm for maximum entropy image restoration is presented. The dual algorithm involves fewer parameters than conventional minimization in the image space. Minicomputer test results for Fourier synthesis with inadequate phantom data are given.
Maximum Throughput in Multiple-Antenna Systems
Zamani, Mahdi
2012-01-01
The point-to-point multiple-antenna channel is investigated in uncorrelated block fading environment with Rayleigh distribution. The maximum throughput and maximum expected-rate of this channel are derived under the assumption that the transmitter is oblivious to the channel state information (CSI), however, the receiver has perfect CSI. First, we prove that in multiple-input single-output (MISO) channels, the optimum transmission strategy maximizing the throughput is to use all available antennas and perform equal power allocation with uncorrelated signals. Furthermore, to increase the expected-rate, multi-layer coding is applied. Analogously, we establish that sending uncorrelated signals and performing equal power allocation across all available antennas at each layer is optimum. A closed form expression for the maximum continuous-layer expected-rate of MISO channels is also obtained. Moreover, we investigate multiple-input multiple-output (MIMO) channels, and formulate the maximum throughput in the asympt...
Photoemission spectromicroscopy with MAXIMUM at Wisconsin
Ng, W.; Ray-Chaudhuri, A.K.; Cole, R.K.; Wallace, J.; Crossley, S.; Crossley, D.; Chen, G.; Green, M.; Guo, J.; Hansen, R.W.C.; Cerrina, F.; Margaritondo, G. (Dept. of Electrical Engineering, Dept. of Physics and Synchrotron Radiation Center, Univ. of Wisconsin, Madison (USA)); Underwood, J.H.; Korthright, J.; Perera, R.C.C. (Center for X-ray Optics, Accelerator and Fusion Research Div., Lawrence Berkeley Lab., CA (USA))
1990-06-01
We describe the development of the scanning photoemission spectromicroscope MAXIMUM at the Wisoncsin Synchrotron Radiation Center, which uses radiation from a 30-period undulator. The article includes a discussion of the first tests after the initial commissioning. (orig.).
Maximum-likelihood method in quantum estimation
Paris, M G A; Sacchi, M F
2001-01-01
The maximum-likelihood method for quantum estimation is reviewed and applied to the reconstruction of density matrix of spin and radiation as well as to the determination of several parameters of interest in quantum optics.
Okuda, Koichi; Hashimoto, Mitsumasa [Kanazawa Medical University, Department of Physics, Kahoku, Ishikawa (Japan); Nakajima, Kenichi; Matsuo, Shinro; Taki, Junichi; Kinuya, Seigo [Kanazawa University Hospital, Department of Nuclear Medicine, Kanazawa, Ishikawa (Japan); Sugino, Shuichi [Okayama Kyokuto Hospital, Department of Radiology, Okayama, Okayama (Japan); Kirihara, Yumiko [FUJIFILM RI Pharma Co., Ltd., Chuo-Ku, Tokyo (Japan)
2016-02-15
The washout rate (WR) has been used in {sup 123}I-metaiodobenzylguanidine (MIBG) imaging to evaluate cardiac sympathetic innervation. However, WR varies depending on the time between the early and late MIBG scans. Late scans are performed at either 3 or 4 hours after injection of MIBG. The aim of this study was to directly compare the WR at 3 hours (WR{sub 3h}) with the WR at 4 hours (WR{sub 4h}). We hypothesized that the cardiac count would reduce linearly between the 3-hour and 4-hour scans. A linear regression model for cardiac counts at two time-points was generated. We enrolled a total of 96 patients who underwent planar {sup 123}I-MIBG scintigraphy early (15 min) and during the late phase at both 3 and 4 hours. Patients were randomly divided into two groups: a model-creation group (group 1) and a clinical validation group (group 2). Cardiac counts at 15 minutes (count{sub early}), 3 hours (count{sub 3h}) and 4 hours (count{sub 4h}) were measured. Cardiac count{sub 4h} was mathematically estimated using the linear regression model from count{sub early} and count{sub 3h}. In group 1, the actual cardiac count{sub 4h}/count{sub early} was highly significantly correlated with count{sub 3h}/count{sub early} (r = 0.979). In group 2, the average estimated count{sub 4h} was 92.8 ± 31.9, and there was no significant difference between this value and the actual count{sub 4h} (91.9 ± 31.9). Bland-Altman analysis revealed a small bias of -0.9 with 95 % limits of agreement of -6.2 and +4.3. WR{sub 4h} calculated using the estimated cardiac count{sub 4h} was comparable to the actual WR{sub 4h} (24.3 ± 9.6 % vs. 25.1 ± 9.7 %, p = ns). Bland-Altman analysis and the intraclass correlation coefficient showed that there was excellent agreement between the estimated and actual WR{sub 4h}. The linear regression model that we used accurately estimated cardiac count{sub 4h} using count{sub early} and count{sub 3h}. Moreover, WR{sub 4h} that was mathematically calculated using
De Luca, G.; Magnus, J.R.
2011-01-01
This article is concerned with the estimation of linear regression models with uncertainty about the choice of the explanatory variables. We introduce the Stata commands bma and wals which implement, respectively, the exact Bayesian Model Averaging (BMA) estimator and the Weighted Average Least Squa
Entanglement in random pure states: spectral density and average von Neumann entropy
Kumar, Santosh; Pandey, Akhilesh, E-mail: skumar.physics@gmail.com, E-mail: ap0700@mail.jnu.ac.in [School of Physical Sciences, Jawaharlal Nehru University, New Delhi 110 067 (India)
2011-11-04
Quantum entanglement plays a crucial role in quantum information, quantum teleportation and quantum computation. The information about the entanglement content between subsystems of the composite system is encoded in the Schmidt eigenvalues. We derive here closed expressions for the spectral density of Schmidt eigenvalues for all three invariant classes of random matrix ensembles. We also obtain exact results for average von Neumann entropy. We find that maximum average entanglement is achieved if the system belongs to the symplectic invariant class. (paper)
The maximum entropy technique. System's statistical description
Belashev, B Z
2002-01-01
The maximum entropy technique (MENT) is applied for searching the distribution functions of physical values. MENT takes into consideration the demand of maximum entropy, the characteristics of the system and the connection conditions, naturally. It is allowed to apply MENT for statistical description of closed and open systems. The examples in which MENT had been used for the description of the equilibrium and nonequilibrium states and the states far from the thermodynamical equilibrium are considered
19 CFR 114.23 - Maximum period.
2010-04-01
... 19 Customs Duties 1 2010-04-01 2010-04-01 false Maximum period. 114.23 Section 114.23 Customs... CARNETS Processing of Carnets § 114.23 Maximum period. (a) A.T.A. carnet. No A.T.A. carnet with a period of validity exceeding 1 year from date of issue shall be accepted. This period of validity cannot be...
Maximum-Likelihood Detection Of Noncoherent CPM
Divsalar, Dariush; Simon, Marvin K.
1993-01-01
Simplified detectors proposed for use in maximum-likelihood-sequence detection of symbols in alphabet of size M transmitted by uncoded, full-response continuous phase modulation over radio channel with additive white Gaussian noise. Structures of receivers derived from particular interpretation of maximum-likelihood metrics. Receivers include front ends, structures of which depends only on M, analogous to those in receivers of coherent CPM. Parts of receivers following front ends have structures, complexity of which would depend on N.
A sixth order averaged vector field method
Li, Haochen; Wang, Yushun; Qin, Mengzhao
2014-01-01
In this paper, based on the theory of rooted trees and B-series, we propose the concrete formulas of the substitution law for the trees of order =5. With the help of the new substitution law, we derive a B-series integrator extending the averaged vector field (AVF) method to high order. The new integrator turns out to be of order six and exactly preserves energy for Hamiltonian systems. Numerical experiments are presented to demonstrate the accuracy and the energy-preserving property of the s...
Phase-averaged transport for quasiperiodic Hamiltonians
Bellissard, J; Schulz-Baldes, H
2002-01-01
For a class of discrete quasi-periodic Schroedinger operators defined by covariant re- presentations of the rotation algebra, a lower bound on phase-averaged transport in terms of the multifractal dimensions of the density of states is proven. This result is established under a Diophantine condition on the incommensuration parameter. The relevant class of operators is distinguished by invariance with respect to symmetry automorphisms of the rotation algebra. It includes the critical Harper (almost-Mathieu) operator. As a by-product, a new solution of the frame problem associated with Weyl-Heisenberg-Gabor lattices of coherent states is given.
Sparsity averaging for radio-interferometric imaging
Carrillo, Rafael E; Wiaux, Yves
2014-01-01
We propose a novel regularization method for compressive imaging in the context of the compressed sensing (CS) theory with coherent and redundant dictionaries. Natural images are often complicated and several types of structures can be present at once. It is well known that piecewise smooth images exhibit gradient sparsity, and that images with extended structures are better encapsulated in wavelet frames. Therefore, we here conjecture that promoting average sparsity or compressibility over multiple frames rather than single frames is an extremely powerful regularization prior.
Fluctuations of wavefunctions about their classical average
Benet, L [Centro Internacional de Ciencias, Ciudad Universitaria, Chamilpa, Cuernavaca (Mexico); Flores, J [Centro Internacional de Ciencias, Ciudad Universitaria, Chamilpa, Cuernavaca (Mexico); Hernandez-Saldana, H [Centro Internacional de Ciencias, Ciudad Universitaria, Chamilpa, Cuernavaca (Mexico); Izrailev, F M [Centro Internacional de Ciencias, Ciudad Universitaria, Chamilpa, Cuernavaca (Mexico); Leyvraz, F [Centro Internacional de Ciencias, Ciudad Universitaria, Chamilpa, Cuernavaca (Mexico); Seligman, T H [Centro Internacional de Ciencias, Ciudad Universitaria, Chamilpa, Cuernavaca (Mexico)
2003-02-07
Quantum-classical correspondence for the average shape of eigenfunctions and the local spectral density of states are well-known facts. In this paper, the fluctuations of the quantum wavefunctions around the classical value are discussed. A simple random matrix model leads to a Gaussian distribution of the amplitudes whose width is determined by the classical shape of the eigenfunction. To compare this prediction with numerical calculations in chaotic models of coupled quartic oscillators, we develop a rescaling method for the components. The expectations are broadly confirmed, but deviations due to scars are observed. This effect is much reduced when both Hamiltonians have chaotic dynamics.
The average free volume model for liquids
Yu, Yang
2014-01-01
In this work, the molar volume thermal expansion coefficient of 59 room temperature ionic liquids is compared with their van der Waals volume Vw. Regular correlation can be discerned between the two quantities. An average free volume model, that considers the particles as hard core with attractive force, is proposed to explain the correlation in this study. A combination between free volume and Lennard-Jones potential is applied to explain the physical phenomena of liquids. Some typical simple liquids (inorganic, organic, metallic and salt) are introduced to verify this hypothesis. Good agreement from the theory prediction and experimental data can be obtained.
Fluctuations of wavefunctions about their classical average
Bénet, L; Hernandez-Saldana, H; Izrailev, F M; Leyvraz, F; Seligman, T H
2003-01-01
Quantum-classical correspondence for the average shape of eigenfunctions and the local spectral density of states are well-known facts. In this paper, the fluctuations of the quantum wavefunctions around the classical value are discussed. A simple random matrix model leads to a Gaussian distribution of the amplitudes whose width is determined by the classical shape of the eigenfunction. To compare this prediction with numerical calculations in chaotic models of coupled quartic oscillators, we develop a rescaling method for the components. The expectations are broadly confirmed, but deviations due to scars are observed. This effect is much reduced when both Hamiltonians have chaotic dynamics.
Grassmann Averages for Scalable Robust PCA
2014-01-01
As the collection of large datasets becomes increasingly automated, the occurrence of outliers will increase—“big data” implies “big outliers”. While principal component analysis (PCA) is often used to reduce the size of data, and scalable solutions exist, it is well-known that outliers can arbitrarily corrupt the results. Unfortunately, state-of-the-art approaches for robust PCA do not scale beyond small-to-medium sized datasets. To address this, we introduce the Grassmann Average (GA), whic...
2013-09-24
... the maximum on-duty periods and minimum off-duty periods for railroad employees performing certain... Employee Has Initiated an On-duty Period on Six Consecutive Days (Section III.D of the Second Interim... Limitation for Freight Train Employees and the Requirement of at Least 48 or 72 Hours Off Duty at the Home...
Marcum, Jennifer L; Browning, Steven R; Reed, Deborah B; Charnigo, Richard J
2011-01-01
The average age of United States farmers has been increasing for 20 years. The objective is to examine the factors associated with hours worked among farmers age 50 and older. A cohort of Kentucky and South Carolina farmers (n = 1394) over age 50 were surveyed annually during 2002-2005. Of those that reported any farm work, males worked 24 mean hours/week and females worked 14 mean hours/week. Greater satisfaction and more experience farming, increased acreage, and presence of animals significantly increased estimated hours farmed, whereas chronic health problems, although prevalent, had a minor role in determining work hours.
SEXUAL DIMORPHISM OF MAXIMUM FEMORAL LENGTH
Pandya A M
2011-04-01
Full Text Available Sexual identification from the skeletal parts has medico legal and anthropological importance. Present study aims to obtain values of maximum femoral length and to evaluate its possible usefulness in determining correct sexual identification. Study sample consisted of 184 dry, normal, adult, human femora (136 male & 48 female from skeletal collections of Anatomy department, M. P. Shah Medical College, Jamnagar, Gujarat. Maximum length of femur was considered as maximum vertical distance between upper end of head of femur and the lowest point on femoral condyle, measured with the osteometric board. Mean Values obtained were, 451.81 and 417.48 for right male and female, and 453.35 and 420.44 for left male and female respectively. Higher value in male was statistically highly significant (P< 0.001 on both sides. Demarking point (D.P. analysis of the data showed that right femora with maximum length more than 476.70 were definitely male and less than 379.99 were definitely female; while for left bones, femora with maximum length more than 484.49 were definitely male and less than 385.73 were definitely female. Maximum length identified 13.43% of right male femora, 4.35% of right female femora, 7.25% of left male femora and 8% of left female femora. [National J of Med Res 2011; 1(2.000: 67-70
Tamer Khatib
2015-01-01
Full Text Available This paper presents a model for predicting hourly solar radiation data using daily solar radiation averages. The proposed model is a generalized regression artificial neural network. This model has three inputs, namely, mean daily solar radiation, hour angle, and sunset hour angle. The output layer has one node which is mean hourly solar radiation. The training and development of the proposed model are done using MATLAB and 43800 records of hourly global solar radiation. The results show that the proposed model has better prediction accuracy compared to some empirical and statistical models. Two error statistics are used in this research to evaluate the proposed model, namely, mean absolute percentage error and root mean square error. These values for the proposed model are 11.8% and −3.1%, respectively. Finally, the proposed model shows better ability in overcoming the sophistic nature of the solar radiation data.
Why the long hours? Job demands and social exchange dynamics.
Genin, Emilie; Haines, Victor Y; Pelletier, David; Rousseau, Vincent; Marchand, Alain
2016-11-22
This study investigates the determinants of long working hours from the perspectives of the demand-control model [Karasek, 1979] and social exchange theory [Blau, 1964; Goulder, 1960]. These two theoretical perspectives are tested to understand why individuals work longer (or shorter) hours. The hypotheses are tested with a representative sample of 1,604 employed Canadians. In line with Karasek's model, the results support that high job demands are positively associated with longer work hours. The social exchange perspective would predict a positive association between skill discretion and work hours. This hypothesis was supported for individuals with a higher education degree. Finally, the results support a positive association between active jobs and longer work hours. Our research suggests that job demands and social exchange dynamics need to be considered together in the explanation of longer (or shorter) work hours.
Detrending moving average algorithm for multifractals
Gu, Gao-Feng; Zhou, Wei-Xing
2010-07-01
The detrending moving average (DMA) algorithm is a widely used technique to quantify the long-term correlations of nonstationary time series and the long-range correlations of fractal surfaces, which contains a parameter θ determining the position of the detrending window. We develop multifractal detrending moving average (MFDMA) algorithms for the analysis of one-dimensional multifractal measures and higher-dimensional multifractals, which is a generalization of the DMA method. The performance of the one-dimensional and two-dimensional MFDMA methods is investigated using synthetic multifractal measures with analytical solutions for backward (θ=0) , centered (θ=0.5) , and forward (θ=1) detrending windows. We find that the estimated multifractal scaling exponent τ(q) and the singularity spectrum f(α) are in good agreement with the theoretical values. In addition, the backward MFDMA method has the best performance, which provides the most accurate estimates of the scaling exponents with lowest error bars, while the centered MFDMA method has the worse performance. It is found that the backward MFDMA algorithm also outperforms the multifractal detrended fluctuation analysis. The one-dimensional backward MFDMA method is applied to analyzing the time series of Shanghai Stock Exchange Composite Index and its multifractal nature is confirmed.
Averaged null energy condition from causality
Hartman, Thomas; Kundu, Sandipan; Tajdini, Amirhossein
2017-07-01
Unitary, Lorentz-invariant quantum field theories in flat spacetime obey mi-crocausality: commutators vanish at spacelike separation. For interacting theories in more than two dimensions, we show that this implies that the averaged null energy, ∫ duT uu , must be non-negative. This non-local operator appears in the operator product expansion of local operators in the lightcone limit, and therefore contributes to n-point functions. We derive a sum rule that isolates this contribution and is manifestly positive. The argument also applies to certain higher spin operators other than the stress tensor, generating an infinite family of new constraints of the form ∫ duX uuu··· u ≥ 0. These lead to new inequalities for the coupling constants of spinning operators in conformal field theory, which include as special cases (but are generally stronger than) the existing constraints from the lightcone bootstrap, deep inelastic scattering, conformal collider methods, and relative entropy. We also comment on the relation to the recent derivation of the averaged null energy condition from relative entropy, and suggest a more general connection between causality and information-theoretic inequalities in QFT.
Local average height distribution of fluctuating interfaces
Smith, Naftali R.; Meerson, Baruch; Sasorov, Pavel V.
2017-01-01
Height fluctuations of growing surfaces can be characterized by the probability distribution of height in a spatial point at a finite time. Recently there has been spectacular progress in the studies of this quantity for the Kardar-Parisi-Zhang (KPZ) equation in 1 +1 dimensions. Here we notice that, at or above a critical dimension, the finite-time one-point height distribution is ill defined in a broad class of linear surface growth models unless the model is regularized at small scales. The regularization via a system-dependent small-scale cutoff leads to a partial loss of universality. As a possible alternative, we introduce a local average height. For the linear models, the probability density of this quantity is well defined in any dimension. The weak-noise theory for these models yields the "optimal path" of the interface conditioned on a nonequilibrium fluctuation of the local average height. As an illustration, we consider the conserved Edwards-Wilkinson (EW) equation, where, without regularization, the finite-time one-point height distribution is ill defined in all physical dimensions. We also determine the optimal path of the interface in a closely related problem of the finite-time height-difference distribution for the nonconserved EW equation in 1 +1 dimension. Finally, we discuss a UV catastrophe in the finite-time one-point distribution of height in the (nonregularized) KPZ equation in 2 +1 dimensions.
Asymptotic Time Averages and Frequency Distributions
Muhammad El-Taha
2016-01-01
Full Text Available Consider an arbitrary nonnegative deterministic process (in a stochastic setting {X(t, t≥0} is a fixed realization, i.e., sample-path of the underlying stochastic process with state space S=(-∞,∞. Using a sample-path approach, we give necessary and sufficient conditions for the long-run time average of a measurable function of process to be equal to the expectation taken with respect to the same measurable function of its long-run frequency distribution. The results are further extended to allow unrestricted parameter (time space. Examples are provided to show that our condition is not superfluous and that it is weaker than uniform integrability. The case of discrete-time processes is also considered. The relationship to previously known sufficient conditions, usually given in stochastic settings, will also be discussed. Our approach is applied to regenerative processes and an extension of a well-known result is given. For researchers interested in sample-path analysis, our results will give them the choice to work with the time average of a process or its frequency distribution function and go back and forth between the two under a mild condition.
“Friday off”: Reducing Working Hours in Europe
Jack Rossiter; Hugh O.'Flynn; Michael Kalush; Giorgos Kallis; Nicholas Ashford
2013-01-01
This article explores the pros and cons for reducing working hours in Europe. To arrive to an informed judgment we review critically the theoretical and empirical literature, mostly from economics, concerning the relation between working hours on the one hand, and productivity, employment, quality of life, and the environment, on the other. We adopt a binary economics distinction between capital and labor productiveness, and are concerned with how working hours may be reduced without harming ...
Ergogenic demands of a 24 hour cycling event.
White, J A; Ward, C; Nelson, H
1984-09-01
The maximal aerobic performance (VO2 max) and energy costs of cycling at various power outputs and equivalent road speeds of a highly trained endurance cyclist (age 23.4 yrs, height 1.95 m, weight 73.1 kg), were measured in the laboratory on an eddy-current cycle ergometer, and the physiological responses related to determinations made during a 24 h cycling time trial event, using continuous ECG recording from which estimates of ergogenic demands were obtained. The cyclist covered a distance of 694 km during the event at an average speed of 28.9 km.h-1 which corresponded to an equivalent oxygen cost of 38.5 ml.kg-1 min-1 and represented approximately 55% of his VO2 max. During the event, the cyclist expended an estimated 82,680 kJ of energy, of which approximately 44,278 kJ (54%) were supplied by repeated feedings of liquids, solids and semi-solids and some 38,402 kJ (46%) came from the stored energy reserves which resulted in a 1.19 kg loss of body weight during the event. The energy demands of the activity were more than three times greater than the highest recorded values of severe industrial work, and similar to the hourly rates of expenditure of shorter duration competitive events, but above the highest values reported over other extreme endurance events over the same period of time. The results thus represent near maximal levels of sustainable ergogenic effort by man over a complete 24 h cycle.
BINDER DRAINAGE TEST FOR POROUS MIXTURES MADE BY VARYING THE MAXIMUM AGGREGATE SIZES
Hardiman Hardiman
2004-01-01
Full Text Available Binder drainage occurs with mixes of small aggregate surface area particularly porous asphalt. The binder drainage test, developed by the Transport Research Laboratory, UK, is commonly used to set an upper limit on the acceptable binder content for a porous mix. This paper presents the results of a laboratory investigation to determine the effects of different binder types on the binder drainage characteristics of porous mix made of various maximum aggregate sizes 20, 14 and 10 mm. Two types of binder were used, conventional 60/70 pen bitumen, and styrene butadiene styrene (SBS modified bitumen. The amount of binder lost through drainage after three hours at the maximum mixing temperature were measured in duplicate for mixes of different maximum sizes and binder contents. The maximum mixing temperature adopted depends on the types of binder used. The retained binder is plotted against the initial mixed binder content, together with the line of equality where the retained binder equals the mixed binder content. The results indicate the significant contribution of using SBS modified bitumen to increase the target bitumen binder content. Their significance is discussed in terms of target binder content, the critical binder content, the maximum mixed binder content and the maximum retained binder content values obtained from the binder drainage test. It was concluded that increasing maximum aggregate sizes decrease the maximum retained binder content, critical binder content, target binder content, maximum mixed binder content, and mixed content for both binders, but however for all mixtures, SBS is the highest.
Working hours and health behaviour among nurses at public hospitals
Juliana da Costa Fernandes
2013-09-01
Full Text Available OBJECTIVE: to analyse the differences between genders in the description in the professional, domestic and total work hours and assess its association with health-related behaviour among nurses. METHODS: this is a transversal study carried out in 18 different public hospitals in the municipality of Rio de Janeiro. The data collection procedure was based on questionnaires. All nurses working with assistance were considered eligible (n=2,279. RESULTS: men and women showed significant differences in relation to working hours. The female group showed longer domestic and total work hours when compared to the group of men. In contrast, the number of hours spent on professional work was higher among men. For the women, both the professional hours and total work hours were often associated with excessive consumption of fried food and also coffee, lack of physical exercise and also the greater occurrence of overweight and obesity. CONCLUSION: both the professional hours and the domestic work hours need to be taken into account in studies about health, self-care and also the care provided within the context of nursing workers, particularly among women. The results add weight to the need for actions for health promotion in this occupational group and the importance of assessing the impact of long working hours on the health of workers.
Thorp, Jonathon; Dattalo, Melissa; Ghanem, Khalil G; Christmas, Colleen
2016-12-01
Training programs have implemented the 2011 ACGME duty hour regulations (DHR) using "workload compression" (WLC) strategies, attempting to fit similar clinical responsibilities into fewer working hours, or workload reduction (WLR) approaches, reducing the number of patient encounters per trainee. Many have expressed concern that these strategies could negatively impact patient care and learner outcomes. This study evaluates the medical knowledge and clinical impact of a WLR intervention in a single institution. Nonrandomized intervention study with comparison to a historical control study among 58 PGY-1 internal medicine trainees in the 2 years after duty hour implementation [exposure cohort (EC), 7/1/2011-6/30/2013], compared to 2 years before implementation [comparison cohort (CC), 7/1/2009-6/30/2011]. Process outcomes were average inpatient encounters, average new inpatient admissions, and average scheduled outpatient encounters per PGY-1 year. Performance outcomes included trainee inpatient and outpatient days on service, In-Training Examination (ITE) scores as an objective surrogate of medical knowledge, Case-Mix Index (CMI), and quality of care measures (30-day readmission rate, 30-day mortality rate, and average length of stay). Baseline characteristics and average numbers of inpatient encounters per PGY-1 class were similar between the EC and CC. However, the EC experienced fewer new inpatient admissions (157.47 ± 40.47 vs. 181.72 ± 25.45; p theoretical concern that reducing PGY-1 inpatient admissions volumes may negatively impact education and clinical care measures, this study found no evidence of such a trade-off.
Asymmetric network connectivity using weighted harmonic averages
Morrison, Greg; Mahadevan, L.
2011-02-01
We propose a non-metric measure of the "closeness" felt between two nodes in an undirected, weighted graph using a simple weighted harmonic average of connectivity, that is a real-valued Generalized Erdös Number (GEN). While our measure is developed with a collaborative network in mind, the approach can be of use in a variety of artificial and real-world networks. We are able to distinguish between network topologies that standard distance metrics view as identical, and use our measure to study some simple analytically tractable networks. We show how this might be used to look at asymmetry in authorship networks such as those that inspired the integer Erdös numbers in mathematical coauthorships. We also show the utility of our approach to devise a ratings scheme that we apply to the data from the NetFlix prize, and find a significant improvement using our method over a baseline.
Averaged Null Energy Condition from Causality
Hartman, Thomas; Tajdini, Amirhossein
2016-01-01
Unitary, Lorentz-invariant quantum field theories in flat spacetime obey microcausality: commutators vanish at spacelike separation. For interacting theories in more than two dimensions, we show that this implies that the averaged null energy, $\\int du T_{uu}$, must be positive. This non-local operator appears in the operator product expansion of local operators in the lightcone limit, and therefore contributes to $n$-point functions. We derive a sum rule that isolates this contribution and is manifestly positive. The argument also applies to certain higher spin operators other than the stress tensor, generating an infinite family of new constraints of the form $\\int du X_{uuu\\cdots u} \\geq 0$. These lead to new inequalities for the coupling constants of spinning operators in conformal field theory, which include as special cases (but are generally stronger than) the existing constraints from the lightcone bootstrap, deep inelastic scattering, conformal collider methods, and relative entropy. We also comment ...
Average Gait Differential Image Based Human Recognition
Jinyan Chen
2014-01-01
Full Text Available The difference between adjacent frames of human walking contains useful information for human gait identification. Based on the previous idea a silhouettes difference based human gait recognition method named as average gait differential image (AGDI is proposed in this paper. The AGDI is generated by the accumulation of the silhouettes difference between adjacent frames. The advantage of this method lies in that as a feature image it can preserve both the kinetic and static information of walking. Comparing to gait energy image (GEI, AGDI is more fit to representation the variation of silhouettes during walking. Two-dimensional principal component analysis (2DPCA is used to extract features from the AGDI. Experiments on CASIA dataset show that AGDI has better identification and verification performance than GEI. Comparing to PCA, 2DPCA is a more efficient and less memory storage consumption feature extraction method in gait based recognition.
Geographic Gossip: Efficient Averaging for Sensor Networks
Dimakis, Alexandros G; Wainwright, Martin J
2007-01-01
Gossip algorithms for distributed computation are attractive due to their simplicity, distributed nature, and robustness in noisy and uncertain environments. However, using standard gossip algorithms can lead to a significant waste in energy by repeatedly recirculating redundant information. For realistic sensor network model topologies like grids and random geometric graphs, the inefficiency of gossip schemes is related to the slow mixing times of random walks on the communication graph. We propose and analyze an alternative gossiping scheme that exploits geographic information. By utilizing geographic routing combined with a simple resampling method, we demonstrate substantial gains over previously proposed gossip protocols. For regular graphs such as the ring or grid, our algorithm improves standard gossip by factors of $n$ and $\\sqrt{n}$ respectively. For the more challenging case of random geometric graphs, our algorithm computes the true average to accuracy $\\epsilon$ using $O(\\frac{n^{1.5}}{\\sqrt{\\log ...
Bivariate phase-rectified signal averaging
Schumann, Aicko Y; Bauer, Axel; Schmidt, Georg
2008-01-01
Phase-Rectified Signal Averaging (PRSA) was shown to be a powerful tool for the study of quasi-periodic oscillations and nonlinear effects in non-stationary signals. Here we present a bivariate PRSA technique for the study of the inter-relationship between two simultaneous data recordings. Its performance is compared with traditional cross-correlation analysis, which, however, does not work well for non-stationary data and cannot distinguish the coupling directions in complex nonlinear situations. We show that bivariate PRSA allows the analysis of events in one signal at times where the other signal is in a certain phase or state; it is stable in the presence of noise and impassible to non-stationarities.
Maximum permissible voltage of YBCO coated conductors
Wen, J.; Lin, B.; Sheng, J.; Xu, J.; Jin, Z.; Hong, Z.; Wang, D.; Zhou, H.; Shen, X.; Shen, C.
2014-06-01
Superconducting fault current limiter (SFCL) could reduce short circuit currents in electrical power system. One of the most important thing in developing SFCL is to find out the maximum permissible voltage of each limiting element. The maximum permissible voltage is defined as the maximum voltage per unit length at which the YBCO coated conductors (CC) do not suffer from critical current (Ic) degradation or burnout. In this research, the time of quenching process is changed and voltage is raised until the Ic degradation or burnout happens. YBCO coated conductors test in the experiment are from American superconductor (AMSC) and Shanghai Jiao Tong University (SJTU). Along with the quenching duration increasing, the maximum permissible voltage of CC decreases. When quenching duration is 100 ms, the maximum permissible of SJTU CC, 12 mm AMSC CC and 4 mm AMSC CC are 0.72 V/cm, 0.52 V/cm and 1.2 V/cm respectively. Based on the results of samples, the whole length of CCs used in the design of a SFCL can be determined.
Computing Rooted and Unrooted Maximum Consistent Supertrees
van Iersel, Leo
2009-01-01
A chief problem in phylogenetics and database theory is the computation of a maximum consistent tree from a set of rooted or unrooted trees. A standard input are triplets, rooted binary trees on three leaves, or quartets, unrooted binary trees on four leaves. We give exact algorithms constructing rooted and unrooted maximum consistent supertrees in time O(2^n n^5 m^2 log(m)) for a set of m triplets (quartets), each one distinctly leaf-labeled by some subset of n labels. The algorithms extend to weighted triplets (quartets). We further present fast exact algorithms for constructing rooted and unrooted maximum consistent trees in polynomial space. Finally, for a set T of m rooted or unrooted trees with maximum degree D and distinctly leaf-labeled by some subset of a set L of n labels, we compute, in O(2^{mD} n^m m^5 n^6 log(m)) time, a tree distinctly leaf-labeled by a maximum-size subset X of L that all trees in T, when restricted to X, are consistent with.
Maximum magnitude earthquakes induced by fluid injection
McGarr, Arthur F.
2014-01-01
Analysis of numerous case histories of earthquake sequences induced by fluid injection at depth reveals that the maximum magnitude appears to be limited according to the total volume of fluid injected. Similarly, the maximum seismic moment seems to have an upper bound proportional to the total volume of injected fluid. Activities involving fluid injection include (1) hydraulic fracturing of shale formations or coal seams to extract gas and oil, (2) disposal of wastewater from these gas and oil activities by injection into deep aquifers, and (3) the development of enhanced geothermal systems by injecting water into hot, low-permeability rock. Of these three operations, wastewater disposal is observed to be associated with the largest earthquakes, with maximum magnitudes sometimes exceeding 5. To estimate the maximum earthquake that could be induced by a given fluid injection project, the rock mass is assumed to be fully saturated, brittle, to respond to injection with a sequence of earthquakes localized to the region weakened by the pore pressure increase of the injection operation and to have a Gutenberg-Richter magnitude distribution with a b value of 1. If these assumptions correctly describe the circumstances of the largest earthquake, then the maximum seismic moment is limited to the volume of injected liquid times the modulus of rigidity. Observations from the available case histories of earthquakes induced by fluid injection are consistent with this bound on seismic moment. In view of the uncertainties in this analysis, however, this should not be regarded as an absolute physical limit.
Maximum magnitude earthquakes induced by fluid injection
McGarr, A.
2014-02-01
Analysis of numerous case histories of earthquake sequences induced by fluid injection at depth reveals that the maximum magnitude appears to be limited according to the total volume of fluid injected. Similarly, the maximum seismic moment seems to have an upper bound proportional to the total volume of injected fluid. Activities involving fluid injection include (1) hydraulic fracturing of shale formations or coal seams to extract gas and oil, (2) disposal of wastewater from these gas and oil activities by injection into deep aquifers, and (3) the development of enhanced geothermal systems by injecting water into hot, low-permeability rock. Of these three operations, wastewater disposal is observed to be associated with the largest earthquakes, with maximum magnitudes sometimes exceeding 5. To estimate the maximum earthquake that could be induced by a given fluid injection project, the rock mass is assumed to be fully saturated, brittle, to respond to injection with a sequence of earthquakes localized to the region weakened by the pore pressure increase of the injection operation and to have a Gutenberg-Richter magnitude distribution with a b value of 1. If these assumptions correctly describe the circumstances of the largest earthquake, then the maximum seismic moment is limited to the volume of injected liquid times the modulus of rigidity. Observations from the available case histories of earthquakes induced by fluid injection are consistent with this bound on seismic moment. In view of the uncertainties in this analysis, however, this should not be regarded as an absolute physical limit.
Maximum embryo absorbed dose from intravenous urography: interhospital variations
Damilakis, J.; Perisinakis, K. [University of Crete (Greece). Dept. of Medical Physics; Koukourakis, M. [University of Crete (Greece). Dept. of Radiology; Gourtsoyiannis, N. [University Hospital of Iraklion, Crete (Greece). Dept. of Radiotherapy
1997-12-01
The purpose of this study was to determine the maximum embryo dose during intravenous urography (IVU) examinations, when inadvertent irradiation of a pregnant woman occurs, and to investigate the variation of doses received from different institutions. Doses at average embryo depth from IVU examinations have been measured in four institutions using a Rando phantom and thermoluminescent crystals. In order to estimate the maximum range of embryo doses, radiologists were asked to carry out the examinations with the same technique as in female patients with acute ureteral obstruction. The range of doses estimated at embryo depth for the institutions participating in this study was 5.77 to 35.2 mGy. The considerable interhospital variation found in dose can be explained by different equipment and techniques used. A simple method of estimating embryo dose from pelvic radiographs reported previously was found to be also applicable to IVU examinations. Absorbed dose at 6 cm, the average embryo depth, was found significantly less than 50 mGy. (Author).
Does a 48-hour rule predict outcomes in patients with acute sigmoid diverticulitis?
Evans, Jessica; Kozol, Robert; Frederick, Wayne; Voytavich, Anthony; Pennoyer, William; Lukianoff, Alexandra; Lardner, Jennifer
2008-03-01
Sigmoid diverticulitis is an infection that resolves with conservative management in 70-85% of patients. Some patients require prolonged hospitalization or surgery during their admission. It has been taught that one should expect clinical improvement within 48 h. In this study, we examined whether basic clinical parameters (the maximum temperature and leukocyte count) of patients would predict improvement and discharge as expected, or prolonged hospitalization. Data was acquired from 198 patients admitted with acute sigmoid diverticulitis as confirmed by computed tomography (CT) scanning and physical exam. One hundred sixty-five patients recovered without surgery with an average hospital stay of 4 days: 120 were discharged within 4 days, whereas 45 patients required longer stays. Nineteen patients underwent surgery early during their admission (within 48 h). Fourteen patients did not improve over time and required surgery later during their hospital stay. The daily maximum temperature and leukocyte count of patients with prolonged stays was compared to the patients who were discharged within 4 days using analysis of variance analysis. The average maximum temperature and leukocyte count on admission were not statistically different between the groups; therefore, maximum temperature and leukocyte count on admission alone are not predictive. After the first 24 h, however, one could see a statistically significant difference in maximum temperature (p=0.004). The leukocyte count responded significantly by hospital day 2 (p=0.003). Both trends were significant through hospital day 4. Patients with a noticeable drop in leukocyte count and maximum temperature over the first 48 h of medical management were predictably discharged early on oral antibiotics. Patients failing to improve at 48 h required prolonged stays or surgery. By observing early trends in leukocyte count and maximum temperature of patients with diverticulitis, one can predict whether they will recover
Generation and applications of high average power Mid-IR supercontinuum in chalcogenide fibres
Petersen, Christian Rosenberg
2016-01-01
Mid-infrared supercontinuum with up to 54.8 mW average power, and maximum bandwidth of 1.77-8.66 Î¼m is demonstrated as a result of pumping tapered chalcogenide photonic crystal fibers with a MHz parametric source at 4 Î¼m....
2013-08-05
... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF AGRICULTURE Food and Nutrition Service National School Lunch, Special Milk, and School Breakfast Programs, National Average Payments/Maximum Reimbursement Rates Correction In notice document 2013-17990, appearing on...
Maximum Multiflow in Wireless Network Coding
Zhou, Jin-Yi; Jiang, Yong; Zheng, Hai-Tao
2012-01-01
In a multihop wireless network, wireless interference is crucial to the maximum multiflow (MMF) problem, which studies the maximum throughput between multiple pairs of sources and sinks. In this paper, we observe that network coding could help to decrease the impacts of wireless interference, and propose a framework to study the MMF problem for multihop wireless networks with network coding. Firstly, a network model is set up to describe the new conflict relations modified by network coding. Then, we formulate a linear programming problem to compute the maximum throughput and show its superiority over one in networks without coding. Finally, the MMF problem in wireless network coding is shown to be NP-hard and a polynomial approximation algorithm is proposed.
An improved maximum power point tracking method for photovoltaic systems
Tafticht, T.; Agbossou, K.; Doumbia, M.L.; Cheriti, A. [Institut de recherche sur l' hydrogene, Departement de genie electrique et genie informatique, Universite du Quebec a Trois-Rivieres, C.P. 500, Trois-Rivieres (QC) (Canada)
2008-07-15
In most of the maximum power point tracking (MPPT) methods described currently in the literature, the optimal operation point of the photovoltaic (PV) systems is estimated by linear approximations. However these approximations can lead to less than optimal operating conditions and hence reduce considerably the performances of the PV system. This paper proposes a new approach to determine the maximum power point (MPP) based on measurements of the open-circuit voltage of the PV modules, and a nonlinear expression for the optimal operating voltage is developed based on this open-circuit voltage. The approach is thus a combination of the nonlinear and perturbation and observation (P and O) methods. The experimental results show that the approach improves clearly the tracking efficiency of the maximum power available at the output of the PV modules. The new method reduces the oscillations around the MPP, and increases the average efficiency of the MPPT obtained. The new MPPT method will deliver more power to any generic load or energy storage media. (author)
75 FR 71545 - Changes to NARA Facilities' Hours of Operation
2010-11-24
... Personnel Records Center archival research room is located at 9700 Page Ave., St. Louis, MO 63132-5100. The... located at 700 Pennsylvania Avenue, NW., Washington, DC 20408. Hours for the Research Center and the Central Research Room are posted at http://www.archives.gov . The exhibit areas' hours of operation are...
75 FR 54543 - Changes to NARA Facilities' Hours of Operation
2010-09-08
... archival research room is located at 9700 Page Ave., St. Louis, MO 63132-5100. The telephone number is 314... Building is located at 700 Pennsylvania Avenue, NW., Washington, DC 20408. Hours for the Research Center and the Central Research Room are posted at http://www.archives.gov . The exhibit areas' hours of...
Contact Hours in Dutch and Vietnamese Higher Education: A Comparison
Quyen, Do Thi Ngoc
2009-01-01
The number of contact hours, one of the important institutional context factors, was examined and compared between Dutch and Vietnamese higher education at institute and student levels in Psychology and Business and/or Economics specializations. The quantity of contact hours per credit point given by institutions was investigated in a number of…
Night Owl: Maryland's After-Hours Reference Service.
Duke, Deborah C.
1994-01-01
Discusses "Night Owl," a Maryland public library's after hours telephone reference service. Issues include project start-up, user profiles, types of questions, volume, after hours reference accessibility, security, costs, service limits, publicity, staffing, and employee turnover. Similar services in other states are cited. (Contains six…
Human prolactin - 24-hour pattern with increased release during sleep.
Sassin, J. F.; Weitzman, E. D.; Kapen, S.; Frantz, A. G.
1972-01-01
Human prolactin was measured in plasma by radioimmunoassay at 20-minute intervals for a 24-hour period in each of six normal adults, whose sleep-wake cycles were monitored polygraphically. A marked diurnal variation in plasma concentrations was demonstrated, with highest values during sleep. Periods of episodic release occurred throughout the 24 hours.
No Correlation Between Work-Hours and Operative Volumes
Kjærgaard, Jane; Sillesen, Martin; Beier-Holgersen, Randi
2016-01-01
OBJECTIVE: Since 2003, United States residents have been limited to an 80-hour workweek. This has prompted concerns of reduced educational quality, especially inadequate operating exposure. In contrast, the Danish surgical specialty-training program mandates a cap on working hours of 37 per week....
With long hours of work, might depression then lurk?
Varma, Anshu; Marott, Jacob Louis; Stoltenberg, Christian Ditlev Gabriel;
2012-01-01
The aim of this study was to examine depression as a potential negative health effect of long work hours, anticipating an exposure-response relationship.......The aim of this study was to examine depression as a potential negative health effect of long work hours, anticipating an exposure-response relationship....
39 CFR 959.3 - Office, business hours.
2010-07-01
... 39 Postal Service 1 2010-07-01 2010-07-01 false Office, business hours. 959.3 Section 959.3 Postal Service UNITED STATES POSTAL SERVICE PROCEDURES RULES OF PRACTICE IN PROCEEDINGS RELATIVE TO THE PRIVATE EXPRESS STATUTES § 959.3 Office, business hours. The offices of the officials mentioned in these rules...
39 CFR 952.4 - Office business hours.
2010-07-01
... 39 Postal Service 1 2010-07-01 2010-07-01 false Office business hours. 952.4 Section 952.4 Postal Service UNITED STATES POSTAL SERVICE PROCEDURES RULES OF PRACTICE IN PROCEEDINGS RELATIVE TO FALSE REPRESENTATION AND LOTTERY ORDERS § 952.4 Office business hours. The offices of the officials mentioned in...
39 CFR 954.4 - Office business hours.
2010-07-01
... 39 Postal Service 1 2010-07-01 2010-07-01 false Office business hours. 954.4 Section 954.4 Postal Service UNITED STATES POSTAL SERVICE PROCEDURES RULES OF PRACTICE IN PROCEEDINGS RELATIVE TO THE DENIAL, SUSPENSION, OR REVOCATION OF PERIODICALS MAIL PRIVILEGES § 954.4 Office business hours. The offices of...
75 FR 19555 - NARA Facility Locations and Hours
2010-04-15
..., and 1280 [FDMS Docket NARA-10-0002] RIN 3095-AB66 NARA Facility Locations and Hours AGENCY: National... Regional Records Service Office, and the addition of facility information for the locations of two NARA... facility added. * * * * * PART 1253--LOCATION OF RECORDS AND HOURS OF USE 0 3. The authority citation for...
12 CFR 905.3 - Location and business hours.
2010-01-01
... Location and business hours. (a) Location. All office units of the Finance Board are located at 1777 F... 12 Banks and Banking 7 2010-01-01 2010-01-01 false Location and business hours. 905.3 Section 905.3 Banks and Banking FEDERAL HOUSING FINANCE BOARD FEDERAL HOUSING FINANCE BOARD ORGANIZATION...
19 CFR 132.3 - Observation of official hours.
2010-04-01
... 19 Customs Duties 1 2010-04-01 2010-04-01 false Observation of official hours. 132.3 Section 132.3 Customs Duties U.S. CUSTOMS AND BORDER PROTECTION, DEPARTMENT OF HOMELAND SECURITY; DEPARTMENT OF THE TREASURY QUOTAS General Provisions § 132.3 Observation of official hours. An entry summary for consumption...
5 CFR 551.421 - Regular working hours.
2010-01-01
... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false Regular working hours. 551.421 Section 551.421 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PAY ADMINISTRATION UNDER THE FAIR LABOR STANDARDS ACT Hours of Work Application of Principles in Relation to...
The Wiener maximum quadratic assignment problem
Cela, Eranda; Woeginger, Gerhard J
2011-01-01
We investigate a special case of the maximum quadratic assignment problem where one matrix is a product matrix and the other matrix is the distance matrix of a one-dimensional point set. We show that this special case, which we call the Wiener maximum quadratic assignment problem, is NP-hard in the ordinary sense and solvable in pseudo-polynomial time. Our approach also yields a polynomial time solution for the following problem from chemical graph theory: Find a tree that maximizes the Wiener index among all trees with a prescribed degree sequence. This settles an open problem from the literature.
Maximum confidence measurements via probabilistic quantum cloning
Zhang Wen-Hai; Yu Long-Bao; Cao Zhuo-Liang; Ye Liu
2013-01-01
Probabilistic quantum cloning (PQC) cannot copy a set of linearly dependent quantum states.In this paper,we show that if incorrect copies are allowed to be produced,linearly dependent quantum states may also be cloned by the PQC.By exploiting this kind of PQC to clone a special set of three linearly dependent quantum states,we derive the upper bound of the maximum confidence measure of a set.An explicit transformation of the maximum confidence measure is presented.
Maximum floodflows in the conterminous United States
Crippen, John R.; Bue, Conrad D.
1977-01-01
Peak floodflows from thousands of observation sites within the conterminous United States were studied to provide a guide for estimating potential maximum floodflows. Data were selected from 883 sites with drainage areas of less than 10,000 square miles (25,900 square kilometers) and were grouped into regional sets. Outstanding floods for each region were plotted on graphs, and envelope curves were computed that offer reasonable limits for estimates of maximum floods. The curves indicate that floods may occur that are two to three times greater than those known for most streams.
Revealing the Maximum Strength in Nanotwinned Copper
Lu, L.; Chen, X.; Huang, Xiaoxu
2009-01-01
The strength of polycrystalline materials increases with decreasing grain size. Below a critical size, smaller grains might lead to softening, as suggested by atomistic simulations. The strongest size should arise at a transition in deformation mechanism from lattice dislocation activities to grain...... boundary–related processes. We investigated the maximum strength of nanotwinned copper samples with different twin thicknesses. We found that the strength increases with decreasing twin thickness, reaching a maximum at 15 nanometers, followed by a softening at smaller values that is accompanied by enhanced...
The Maximum Resource Bin Packing Problem
Boyar, J.; Epstein, L.; Favrholdt, L.M.
2006-01-01
Usually, for bin packing problems, we try to minimize the number of bins used or in the case of the dual bin packing problem, maximize the number or total size of accepted items. This paper presents results for the opposite problems, where we would like to maximize the number of bins used...... algorithms, First-Fit-Increasing and First-Fit-Decreasing for the maximum resource variant of classical bin packing. For the on-line variant, we define maximum resource variants of classical and dual bin packing. For dual bin packing, no on-line algorithm is competitive. For classical bin packing, we find...
Maximum entropy analysis of EGRET data
Pohl, M.; Strong, A.W.
1997-01-01
EGRET data are usually analysed on the basis of the Maximum-Likelihood method \\cite{ma96} in a search for point sources in excess to a model for the background radiation (e.g. \\cite{hu97}). This method depends strongly on the quality of the background model, and thus may have high systematic unce...... uncertainties in region of strong and uncertain background like the Galactic Center region. Here we show images of such regions obtained by the quantified Maximum-Entropy method. We also discuss a possible further use of MEM in the analysis of problematic regions of the sky....
Revealing the Maximum Strength in Nanotwinned Copper
Lu, L.; Chen, X.; Huang, Xiaoxu
2009-01-01
The strength of polycrystalline materials increases with decreasing grain size. Below a critical size, smaller grains might lead to softening, as suggested by atomistic simulations. The strongest size should arise at a transition in deformation mechanism from lattice dislocation activities to grain...... boundary–related processes. We investigated the maximum strength of nanotwinned copper samples with different twin thicknesses. We found that the strength increases with decreasing twin thickness, reaching a maximum at 15 nanometers, followed by a softening at smaller values that is accompanied by enhanced...
Maximum phytoplankton concentrations in the sea
Jackson, G.A.; Kiørboe, Thomas
2008-01-01
A simplification of plankton dynamics using coagulation theory provides predictions of the maximum algal concentration sustainable in aquatic systems. These predictions have previously been tested successfully against results from iron fertilization experiments. We extend the test to data collected...... in the North Atlantic as part of the Bermuda Atlantic Time Series program as well as data collected off Southern California as part of the Southern California Bight Study program. The observed maximum particulate organic carbon and volumetric particle concentrations are consistent with the predictions...
Jan Vagedes
Full Text Available OBJECTIVE: The number of desaturations determined in recordings of pulse oximeter saturation (SpO2 primarily depends on the time over which values are averaged. As the averaging time in pulse oximeters is not standardized, it varies considerably between centers. To make SpO2 data comparable, it is thus desirable to have a formula that allows conversion between desaturation rates obtained using different averaging times for various desaturation levels and minimal durations. METHODS: Oxygen saturation was measured for 170 hours in 12 preterm infants with a mean number of 65 desaturations <90% per hour of arbitrary duration by using a pulse oximeter in a 2-4 s averaging mode. Using 7 different averaging times between 3 and 16 seconds, the raw red-to-infrared data were reprocessed to determine the number of desaturations (D. The whole procedure was carried out for 7 different minimal desaturation durations (≥ 1, ≥ 5, ≥ 10, ≥ 15, ≥ 20, ≥ 25, ≥ 30 s below SpO2 threshold values of 80%, 85% or 90% to finally reach a conversion formula. The formula was validated by splitting the infants into two groups of six children each and using one group each as a training set and the other one as a test set. RESULTS: Based on the linear relationship found between the logarithm of the desaturation rate and the logarithm of the averaging time, the conversion formula is: D2 = D1 (T2/T1(c, where D2 is the desaturation rate for the desired averaging time T2, and D1 is the desaturation rate for the original averaging time T1, with the exponent c depending on the desaturation threshold and the minimal desaturation duration. The median error when applying this formula was 2.6%. CONCLUSION: This formula enables the conversion of desaturation rates between different averaging times for various desaturation thresholds and minimal desaturation durations.
Employment Standards Administration (DOL), Washington, DC.
At the request of the Employment Standards Administration, the Bureau of Labor Statistics conducted a nationwide survey of private nonfarm industries, designed to provide comprehensive frequency distributions of nonsupervisory employees by average straight-time hourly earnings and weekly hours of work by coverage status under the Fair Labor…
Erceg, Marinko; Becić, Kristijan
2008-01-01
The use of closed-suction drainage systems after total hip and knee replacement is a common practice. It is believed that drainage reduces the haematomas and infection. The usual time to remove drains is 48 hours. Usefulness of the long time drainage of the haematomas and possible harmfulness of infection is still a subject of discussion. In this article, using our own operative material, we want to show the justification of the 48 hour duration of closed-suction drainage. The investigation was done on the patients with hip replacement (45 patients) and knee replacement (11 patients). The amount of blood in the drainage systems in the first 24 hours after the surgery and after 24 hours till the removal of drainage was observed. The results show that the amount of blood lost in the drainage system in the first 24 hours was on average 95.29%, and only 4.71% in the next 24 hours. This article supports the opinion that postoperative drainage has to be removed after 24 hours, i.e. the 48 hours-drainage seems not to be justified.
Industrial Applications of High Average Power FELS
Shinn, Michelle D
2005-01-01
The use of lasers for material processing continues to expand, and the annual sales of such lasers exceeds $1 B (US). Large scale (many m2) processing of materials require the economical production of laser powers of the tens of kilowatts, and therefore are not yet commercial processes, although they have been demonstrated. The development of FELs based on superconducting RF (SRF) linac technology provides a scaleable path to laser outputs above 50 kW in the IR, rendering these applications economically viable, since the cost/photon drops as the output power increases. This approach also enables high average power ~ 1 kW output in the UV spectrum. Such FELs will provide quasi-cw (PRFs in the tens of MHz), of ultrafast (pulsewidth ~ 1 ps) output with very high beam quality. This talk will provide an overview of applications tests by our facility's users such as pulsed laser deposition, laser ablation, and laser surface modification, as well as present plans that will be tested with our upgraded FELs. These upg...
A new approach for Bayesian model averaging
TIAN XiangJun; XIE ZhengHui; WANG AiHui; YANG XiaoChun
2012-01-01
Bayesian model averaging (BMA) is a recently proposed statistical method for calibrating forecast ensembles from numerical weather models.However,successful implementation of BMA requires accurate estimates of the weights and variances of the individual competing models in the ensemble.Two methods,namely the Expectation-Maximization (EM) and the Markov Chain Monte Carlo (MCMC) algorithms,are widely used for BMA model training.Both methods have their own respective strengths and weaknesses.In this paper,we first modify the BMA log-likelihood function with the aim of removing the additional limitation that requires that the BMA weights add to one,and then use a limited memory quasi-Newtonian algorithm for solving the nonlinear optimization problem,thereby formulating a new approach for BMA (referred to as BMA-BFGS).Several groups of multi-model soil moisture simulation experiments from three land surface models show that the performance of BMA-BFGS is similar to the MCMC method in terms of simulation accuracy,and that both are superior to the EM algorithm.On the other hand,the computational cost of the BMA-BFGS algorithm is substantially less than for MCMC and is almost equivalent to that for EM.
Calculating Free Energies Using Average Force
Darve, Eric; Pohorille, Andrew; DeVincenzi, Donald L. (Technical Monitor)
2001-01-01
A new, general formula that connects the derivatives of the free energy along the selected, generalized coordinates of the system with the instantaneous force acting on these coordinates is derived. The instantaneous force is defined as the force acting on the coordinate of interest so that when it is subtracted from the equations of motion the acceleration along this coordinate is zero. The formula applies to simulations in which the selected coordinates are either unconstrained or constrained to fixed values. It is shown that in the latter case the formula reduces to the expression previously derived by den Otter and Briels. If simulations are carried out without constraining the coordinates of interest, the formula leads to a new method for calculating the free energy changes along these coordinates. This method is tested in two examples - rotation around the C-C bond of 1,2-dichloroethane immersed in water and transfer of fluoromethane across the water-hexane interface. The calculated free energies are compared with those obtained by two commonly used methods. One of them relies on determining the probability density function of finding the system at different values of the selected coordinate and the other requires calculating the average force at discrete locations along this coordinate in a series of constrained simulations. The free energies calculated by these three methods are in excellent agreement. The relative advantages of each method are discussed.
Maximum-power-point tracking control of solar heating system
Huang, Bin-Juine
2012-11-01
The present study developed a maximum-power point tracking control (MPPT) technology for solar heating system to minimize the pumping power consumption at an optimal heat collection. The net solar energy gain Q net (=Q s-W p/η e) was experimentally found to be the cost function for MPPT with maximum point. The feedback tracking control system was developed to track the optimal Q net (denoted Q max). A tracking filter which was derived from the thermal analytical model of the solar heating system was used to determine the instantaneous tracking target Q max(t). The system transfer-function model of solar heating system was also derived experimentally using a step response test and used in the design of tracking feedback control system. The PI controller was designed for a tracking target Q max(t) with a quadratic time function. The MPPT control system was implemented using a microprocessor-based controller and the test results show good tracking performance with small tracking errors. It is seen that the average mass flow rate for the specific test periods in five different days is between 18.1 and 22.9kg/min with average pumping power between 77 and 140W, which is greatly reduced as compared to the standard flow rate at 31kg/min and pumping power 450W which is based on the flow rate 0.02kg/sm 2 defined in the ANSI/ASHRAE 93-1986 Standard and the total collector area 25.9m 2. The average net solar heat collected Q net is between 8.62 and 14.1kW depending on weather condition. The MPPT control of solar heating system has been verified to be able to minimize the pumping energy consumption with optimal solar heat collection. © 2012 Elsevier Ltd.
Volume calculation of the spur gear billet for cold precision forging with average circle method
Wangjun Cheng; Chengzhong Chi; Yongzhen Wang; Peng Lin; Wei Liang; Chen Li
2014-01-01
Forging spur gears are widely used in the driving system of mining machinery and equipment due to their higher strength and dimensional accuracy. For the purpose of precisely calculating the volume of cylindrical spur gear billet in cold precision forging, a new theoretical method named average circle method was put forward. With this method, a series of gear billet volumes were calculated. Comparing with the accurate three-dimensional modeling method, the accuracy of average circle method by theoretical calculation was estimated and the maximum relative error of average circle method was less than 1.5%, which was in good agreement with the experimental results. Relative errors of the calculated and the experimental for obtaining the gear billet volumes with reference circle method are larger than those of the average circle method. It shows that average circle method possesses a higher calculation accuracy than reference circle method (traditional method), which should be worth popularizing widely in calculation of spur gear billet volume.
Analysis of Photovoltaic Maximum Power Point Trackers
Veerachary, Mummadi
The photovoltaic generator exhibits a non-linear i-v characteristic and its maximum power point (MPP) varies with solar insolation. An intermediate switch-mode dc-dc converter is required to extract maximum power from the photovoltaic array. In this paper buck, boost and buck-boost topologies are considered and a detailed mathematical analysis, both for continuous and discontinuous inductor current operation, is given for MPP operation. The conditions on the connected load values and duty ratio are derived for achieving the satisfactory maximum power point operation. Further, it is shown that certain load values, falling out of the optimal range, will drive the operating point away from the true maximum power point. Detailed comparison of various topologies for MPPT is given. Selection of the converter topology for a given loading is discussed. Detailed discussion on circuit-oriented model development is given and then MPPT effectiveness of various converter systems is verified through simulations. Proposed theory and analysis is validated through experimental investigations.
On maximum cycle packings in polyhedral graphs
Peter Recht
2014-04-01
Full Text Available This paper addresses upper and lower bounds for the cardinality of a maximum vertex-/edge-disjoint cycle packing in a polyhedral graph G. Bounds on the cardinality of such packings are provided, that depend on the size, the order or the number of faces of G, respectively. Polyhedral graphs are constructed, that attain these bounds.
Hard graphs for the maximum clique problem
Hoede, Cornelis
1988-01-01
The maximum clique problem is one of the NP-complete problems. There are graphs for which a reduction technique exists that transforms the problem for these graphs into one for graphs with specific properties in polynomial time. The resulting graphs do not grow exponentially in order and number. Gra
Maximum Likelihood Estimation of Search Costs
J.L. Moraga-Gonzalez (José Luis); M.R. Wildenbeest (Matthijs)
2006-01-01
textabstractIn a recent paper Hong and Shum (forthcoming) present a structural methodology to estimate search cost distributions. We extend their approach to the case of oligopoly and present a maximum likelihood estimate of the search cost distribution. We apply our method to a data set of online p
Weak Scale From the Maximum Entropy Principle
Hamada, Yuta; Kawana, Kiyoharu
2015-01-01
The theory of multiverse and wormholes suggests that the parameters of the Standard Model are fixed in such a way that the radiation of the $S^{3}$ universe at the final stage $S_{rad}$ becomes maximum, which we call the maximum entropy principle. Although it is difficult to confirm this principle generally, for a few parameters of the Standard Model, we can check whether $S_{rad}$ actually becomes maximum at the observed values. In this paper, we regard $S_{rad}$ at the final stage as a function of the weak scale ( the Higgs expectation value ) $v_{h}$, and show that it becomes maximum around $v_{h}={\\cal{O}}(300\\text{GeV})$ when the dimensionless couplings in the Standard Model, that is, the Higgs self coupling, the gauge couplings, and the Yukawa couplings are fixed. Roughly speaking, we find that the weak scale is given by \\begin{equation} v_{h}\\sim\\frac{T_{BBN}^{2}}{M_{pl}y_{e}^{5}},\
Weak scale from the maximum entropy principle
Hamada, Yuta; Kawai, Hikaru; Kawana, Kiyoharu
2015-03-01
The theory of the multiverse and wormholes suggests that the parameters of the Standard Model (SM) are fixed in such a way that the radiation of the S3 universe at the final stage S_rad becomes maximum, which we call the maximum entropy principle. Although it is difficult to confirm this principle generally, for a few parameters of the SM, we can check whether S_rad actually becomes maximum at the observed values. In this paper, we regard S_rad at the final stage as a function of the weak scale (the Higgs expectation value) vh, and show that it becomes maximum around vh = {{O}} (300 GeV) when the dimensionless couplings in the SM, i.e., the Higgs self-coupling, the gauge couplings, and the Yukawa couplings are fixed. Roughly speaking, we find that the weak scale is given by vh ˜ T_{BBN}2 / (M_{pl}ye5), where ye is the Yukawa coupling of electron, T_BBN is the temperature at which the Big Bang nucleosynthesis starts, and M_pl is the Planck mass.
Global characterization of the Holocene Thermal Maximum
Renssen, H.; Seppä, H.; Crosta, X.; Goosse, H.; Roche, D.M.V.A.P.
2012-01-01
We analyze the global variations in the timing and magnitude of the Holocene Thermal Maximum (HTM) and their dependence on various forcings in transient simulations covering the last 9000 years (9 ka), performed with a global atmosphere-ocean-vegetation model. In these experiments, we consider the i
Instance Optimality of the Adaptive Maximum Strategy
L. Diening; C. Kreuzer; R. Stevenson
2016-01-01
In this paper, we prove that the standard adaptive finite element method with a (modified) maximum marking strategy is instance optimal for the total error, being the square root of the squared energy error plus the squared oscillation. This result will be derived in the model setting of Poisson’s e
Maximum Phonation Time: Variability and Reliability
R. Speyer; H.C.A. Bogaardt; V.L. Passos; N.P.H.D. Roodenburg; A. Zumach; M.A.M. Heijnen; L.W.J. Baijens; S.J.H.M. Fleskens; J.W. Brunings
2010-01-01
The objective of the study was to determine maximum phonation time reliability as a function of the number of trials, days, and raters in dysphonic and control subjects. Two groups of adult subjects participated in this reliability study: a group of outpatients with functional or organic dysphonia v
Maximum likelihood estimation of fractionally cointegrated systems
Lasak, Katarzyna
In this paper we consider a fractionally cointegrated error correction model and investigate asymptotic properties of the maximum likelihood (ML) estimators of the matrix of the cointe- gration relations, the degree of fractional cointegration, the matrix of the speed of adjustment...
Maximum likelihood estimation for integrated diffusion processes
Baltazar-Larios, Fernando; Sørensen, Michael
EM-algorithm to obtain maximum likelihood estimates of the parameters in the diffusion model. As part of the algorithm, we use a recent simple method for approximate simulation of diffusion bridges. In simulation studies for the Ornstein-Uhlenbeck process and the CIR process the proposed method works...
Maximum gain of Yagi-Uda arrays
Bojsen, J.H.; Schjær-Jacobsen, Hans; Nilsson, E.
1971-01-01
Numerical optimisation techniques have been used to find the maximum gain of some specific parasitic arrays. The gain of an array of infinitely thin, equispaced dipoles loaded with arbitrary reactances has been optimised. The results show that standard travelling-wave design methods are not optimum....... Yagi–Uda arrays with equal and unequal spacing have also been optimised with experimental verification....
杨树政; 林理彬
2002-01-01
We have found that the nonthermal radiation of a nonstationary Kerr-Newman black hole is affected by interstellar materials. In particular, the interstellar gas deeply influences the average range of nonthermal radiation particles, while the average range depends on the maximum energy of the radiation and the energy extent of the radiation.
42 CFR 495.308 - Net average allowable costs as the basis for determining the incentive payment.
2010-10-01
... 42 Public Health 5 2010-10-01 2010-10-01 false Net average allowable costs as the basis for... Net average allowable costs as the basis for determining the incentive payment. (a) The first year of..., implementation or upgrade of certified electronic health records technology. (2) The maximum net...
Static stretching can impair explosive performance for at least 24 hours.
Haddad, Monoem; Dridi, Amir; Chtara, Moktar; Chaouachi, Anis; Wong, Del P; Behm, David; Chamari, Karim
2014-01-01
The aim of this study was to compare the effects of static vs. dynamic stretching (DS) on explosive performances and repeated sprint ability (RSA) after a 24-hour delay. Sixteen young male soccer players performed 15 minutes of static stretching (SS), DS, or a no-stretch control condition (CC) 24 hours before performing explosive performances and RSA tests. This was a within-subject repeated measures study with SS, DS, and CC being counterbalanced. Stretching protocols included 2 sets of 7 minutes 30 seconds (2 repetitions of 30 seconds with a 15-second passive recovery) for 5 muscle groups (quadriceps, hamstring, calves, adductors, and hip flexors). Twenty-four hours later (without any kind of stretching in warm-up), the players were tested for the 30-m sprint test (with 10- and 20-m lap times), 5 jump test (5JT), and RSA test. Significant differences were observed between CC, SS, and DS with 5JT (F = 9.99, p effect size [ES] = 0.40), 10-m sprint time (F = 46.52, p stretching protocols in the total time (F = 1.55, p > 0.05), average time (F = 1.53, p > 0.05), and fastest time (F = 2.30, p > 0.05), except for the decline index (F = 3.54, p negative effect on explosive performances up to 24 hours poststretching with no major effects on the RSA. Conversely, the DS of the same muscle groups are highly recommended 24 hours before performing sprint and long-jump performances. In conclusion, the positive effects of DS on explosive performances seem to persist for 24 hours.
Impact of working hours on sleep and mental health.
Afonso, P; Fonseca, M; Pires, J F
2017-07-01
The number of hours people are required to work has a pervasive influence on both physical and mental health. Excessive working hours can also negatively affect sleep quality. The impact at work of mental health problems can have serious consequences for individuals' as well as for organizations' productivity. To evaluate differences in sleep quality and anxiety and depression symptoms between longer working hours group (LWHG) and regular working hours group (RWHG). To examine factors influencing weekly working hours, sleep quality and anxiety and depressive symptoms. Participants were divided into two groups, RWHG and LWHG, based on working hours, with a cut-off of 48 h per week. We used the Hospital Anxiety and Depression Scale (HADS) to assess anxiety and depression symptoms and the Pittsburgh Sleep Quality Index (PSQI) to measure the quality and patterns of sleep. The response rate was 23%. Among the 429 study participants, those in the LWHG group (n = 256, 53%) had significantly more depressive and anxiety symptoms and worse sleep quality than those in RWHG (n = 223, 47%). Working time was significantly positively correlated with higher corporate position and HADS scores. Moreover, HADS scores were positively correlated with PSQI scores and negatively correlated with age. This study suggests that longer working hours are associated with poorer mental health status and increasing levels of anxiety and depression symptoms. There was a positive correlation between these symptoms and sleep disturbances.
Jun He; Xin Yao
2004-01-01
Most of works on the time complexity analysis of evolutionary algorithms have always focused on some artificial binary problems.The time complexity of the algorithms for combinatorial optimisation has not been well understood.This paper considers the time complexity of an evolutionary algorithm for a classical combinatorial optimisation problem,to find the maximum cardinality matching in a graph.It is shown that the evolutionary algorithm can produce a matching with nearly maximum cardinality in average polynomial time.
Yu, Hwa-Lung; Wang, Chih-Hsin
2013-02-05
Understanding the daily changes in ambient air quality concentrations is important to the assessing human exposure and environmental health. However, the fine temporal scales (e.g., hourly) involved in this assessment often lead to high variability in air quality concentrations. This is because of the complex short-term physical and chemical mechanisms among the pollutants. Consequently, high heterogeneity is usually present in not only the averaged pollution levels, but also the intraday variance levels of the daily observations of ambient concentration across space and time. This characteristic decreases the estimation performance of common techniques. This study proposes a novel quantile-based Bayesian maximum entropy (QBME) method to account for the nonstationary and nonhomogeneous characteristics of ambient air pollution dynamics. The QBME method characterizes the spatiotemporal dependence among the ambient air quality levels based on their location-specific quantiles and accounts for spatiotemporal variations using a local weighted smoothing technique. The epistemic framework of the QBME method can allow researchers to further consider the uncertainty of space-time observations. This study presents the spatiotemporal modeling of daily CO and PM10 concentrations across Taiwan from 1998 to 2009 using the QBME method. Results show that the QBME method can effectively improve estimation accuracy in terms of lower mean absolute errors and standard deviations over space and time, especially for pollutants with strong nonhomogeneous variances across space. In addition, the epistemic framework can allow researchers to assimilate the site-specific secondary information where the observations are absent because of the common preferential sampling issues of environmental data. The proposed QBME method provides a practical and powerful framework for the spatiotemporal modeling of ambient pollutants.
A Study of Wind Statistics Through Auto-Regressive and Moving-Average (ARMA) Modeling
尹彰; 周宗仁
2001-01-01
Statistical properties of winds near the Taichung Harbour are investigated. The 26 years′incomplete data of wind speeds, measured on an hourly basis, are used as reference. The possibility of imputation using simulated results of the Auto-Regressive (AR), Moving-Average (MA), and/or Auto-Regressive and Moving-Average (ARMA) models is studied. Predictions of the 25-year extreme wind speeds based upon the augmented data are compared with the original series. Based upon the results, predictions of the 50- and 100-year extreme wind speeds are then made.
US public opinion regarding proposed limits on resident physician work hours
Lurie Peter
2010-06-01
Full Text Available Abstract Background In both Europe and the US, resident physician work hour reduction has been a source of controversy within academic medicine. In 2008, the Institute of Medicine (IOM recommended a reduction in resident physician work hours. We sought to assess the American public perspective on this issue. Methods We conducted a national survey of 1,200 representative members of the public via random digit telephone dialing in order to describe US public opinion on resident physician work hour regulation, particularly with reference to the IOM recommendations. Results Respondents estimated that resident physicians currently work 12.9-h shifts (95% CI 12.5 to 13.3 h and 58.3-h work weeks (95% CI 57.3 to 59.3 h. They believed the maximum shift duration should be 10.9 h (95% CI 10.6 to 11.3 h and the maximum work week should be 50 h (95% CI 49.4 to 50.8 h, with 1% approving of shifts lasting >24 h (95% CI 0.6% to 2%. A total of 81% (95% CI 79% to 84% believed reducing resident physician work hours would be very or somewhat effective in reducing medical errors, and 68% (95% CI 65% to 71% favored the IOM proposal that resident physicians not work more than 16 h over an alternative IOM proposal permitting 30-h shifts with ≥5 h protected sleep time. In all, 81% believed patients should be informed if a treating resident physician had been working for >24 h and 80% (95% CI 78% to 83% would then want a different doctor. Conclusions The American public overwhelmingly favors discontinuation of the 30-h shifts without protected sleep routinely worked by US resident physicians and strongly supports implementation of restrictions on resident physician work hours that are as strict, or stricter, than those proposed by the IOM. Strong support exists to restrict resident physicians' work to 16 or fewer consecutive hours, similar to current limits in New Zealand, the UK and the rest of Europe.
Barış AFŞAR
2012-05-01
Full Text Available OBJECTIVE: Autonomic nervous system dysfunction (ASD has been widely observed in patients with type 2 diabetes. 24-hour ambulatory blood pressure (ABP and heart rate measurements have been found to associate with ASD in patient with Type 2 diabetes. Since albumin excretion is also related with ASD in type 2 diabetes; in the current study, the relationships between daytime, nighttime and 24- hour heart rates with 24 hour urinary albumin excretion (UAE and 24-hour urinary protein excretion (UPE were analyzed in patients with newly diagnosed type 2 diabetes. MATERIAL and METHODS: All patients underwent following procedures: history taking, physical examination, BP measurement, 12 lead electrocardiographic evaluations, routine urine analysis, biochemical analysis, 24-hour urine collection to measure UAE, UPE and creatinine clearance. 24-hour ABP and heart rate monitoring were performed for all patients. RESULTS: In total 80 patients with newly diagnosed type 2 diabetes were included. Stepwise linear regression revealed that logarithmically converted 24-hour UAE were independently related with 24- hour ambulatory SBP, (P:0.001 and heart rate (night (P<0.0001. Stepwise linear regression revealed that logarithmically converted 24-hour UPE were independently related with age (P:0.032, with averaged fasting blood glucose (P:0.023, with 24-hour ambulatory SBP, (P:0.002 and with heart rate (night (P:0.001. CONCLUSION: Nighttime heart rate, but not daytime and 24-hour heart rate was related with both 24-hour UAE and UPE in patients with Type 2 diabetes.
Work hours, social value of leisure and globalisation
Hansen, Jørgen Drud; Molana, Hassan; Montagna, Catia
2012-01-01
and raises the wage rate; there is a U-shaped (inverse-U-shaped) relationship between work hours (wages) and the degree of coordination; coordination is welfare improving; and, the gap between the coordinated and uncoordinated work hours (and the corresponding wage rates) is affected by the extent and nature......We examine how openness interacts with the coordination of consumption-leisure decisions in determining the equilibrium working hours and wage rate when there are leisure externalities (e.g., due to social interactions). The latter are modelled by allowing a worker’s marginal utility of leisure...
Work Hours, Social Value of Leisure and Globalisation
Hansen, Jørgen Drud; Nielsen, Jørgen Ulff-Møller; Montagna, Catia
and raises the wage rate; there is a U-shaped (inverse-U-shaped) relationship between work hours (wages) and the degree of coordination; coordination is welfare improving; and, the gap between the coordinated and uncoordinated work hours (and the corresponding wage rates) is affected by the extent and nature......We examine how openness interacts with the coordination of consumption-leisure decisions in determining the equilibrium working hours and wage rate when there are leisure externalities (e.g., due to social interactions). The latter are modelled by allowing a worker's marginal utility of leisure...
Work Hours, Social Value of Leisure and Globalisation
Hansen, Jørgen Drud; Molana, Hassan; Montagna, Catia
2010-01-01
and raises the wage rate; there is a U-shaped (inverse-U-shaped) relationship between work hours (wages) and the degree of coordination; coordination is welfare improving; and, the gap between the coordinated and uncoordinated work hours (and the corresponding wage rates) is affected by the extent and nature......We examine how openness interacts with the coordination of consumption-leisure decisions in determining the equilibrium working hours and wage rate when there are leisure externalities (e.g., due to social interactions). The latter are modelled by allowing a worker's marginal utility of leisure...
Interpreting Sky-Averaged 21-cm Measurements
Mirocha, Jordan
2015-01-01
Within the first ~billion years after the Big Bang, the intergalactic medium (IGM) underwent a remarkable transformation, from a uniform sea of cold neutral hydrogen gas to a fully ionized, metal-enriched plasma. Three milestones during this epoch of reionization -- the emergence of the first stars, black holes (BHs), and full-fledged galaxies -- are expected to manifest themselves as extrema in sky-averaged ("global") measurements of the redshifted 21-cm background. However, interpreting these measurements will be complicated by the presence of strong foregrounds and non-trivialities in the radiative transfer (RT) modeling required to make robust predictions.I have developed numerical models that efficiently solve the frequency-dependent radiative transfer equation, which has led to two advances in studies of the global 21-cm signal. First, frequency-dependent solutions facilitate studies of how the global 21-cm signal may be used to constrain the detailed spectral properties of the first stars, BHs, and galaxies, rather than just the timing of their formation. And second, the speed of these calculations allows one to search vast expanses of a currently unconstrained parameter space, while simultaneously characterizing the degeneracies between parameters of interest. I find principally that (1) physical properties of the IGM, such as its temperature and ionization state, can be constrained robustly from observations of the global 21-cm signal without invoking models for the astrophysical sources themselves, (2) translating IGM properties to galaxy properties is challenging, in large part due to frequency-dependent effects. For instance, evolution in the characteristic spectrum of accreting BHs can modify the 21-cm absorption signal at levels accessible to first generation instruments, but could easily be confused with evolution in the X-ray luminosity star-formation rate relation. Finally, (3) the independent constraints most likely to aide in the interpretation
Vauchel, Peggy; Arhaliass, Abdellah; Legrand, Jack; Kaas, Raymond; Baron, Regis
2008-01-01
Alginates are natural polysaccharides that are extracted from brown seaweeds and widely used for their rheological properties. The central step in the extraction protocol used in the alginate industry is the alkaline extraction, which requires several hours. In this study, a significant decrease in alginate dynamic viscosity was observed after 2 h of alkaline treatment. Intrinsic viscosity and average molecular weight of alginates from alkaline extractions 1-4 h in duration were determined, i...
Kim, Leonard, E-mail: kimlh@umdnj.edu [Department of Radiation Oncology, Cancer Institute of New Jersey, Robert Wood Johnson Medical School, University of Medicine and Dentistry of New Jersey, New Brunswick, NJ (United States); Narra, Venkat; Yue, Ning [Department of Radiation Oncology, Cancer Institute of New Jersey, Robert Wood Johnson Medical School, University of Medicine and Dentistry of New Jersey, New Brunswick, NJ (United States)
2013-07-01
Recent studies have reported potentially clinically meaningful dose differences when heterogeneity correction is used in breast balloon brachytherapy. In this study, we report on the relationship between heterogeneity-corrected and -uncorrected doses for 2 commonly used plan evaluation metrics: maximum point dose to skin surface and maximum point dose to ribs. Maximum point doses to skin surface and ribs were calculated using TG-43 and Varian Acuros for 20 patients treated with breast balloon brachytherapy. The results were plotted against each other and fit with a zero-intercept line. Max skin dose (Acuros) = max skin dose (TG-43) ⁎ 0.930 (R{sup 2} = 0.995). The average magnitude of difference from this relationship was 1.1% (max 2.8%). Max rib dose (Acuros) = max rib dose (TG-43) ⁎ 0.955 (R{sup 2} = 0.9995). The average magnitude of difference from this relationship was 0.7% (max 1.6%). Heterogeneity-corrected maximum point doses to the skin surface and ribs were proportional to TG-43-calculated doses. The average deviation from proportionality was 1%. The proportional relationship suggests that a different metric other than maximum point dose may be needed to obtain a clinical advantage from heterogeneity correction. Alternatively, if maximum point dose continues to be used in recommended limits while incorporating heterogeneity correction, institutions without this capability may be able to accurately estimate these doses by use of a scaling factor.
Kim, Leonard; Narra, Venkat; Yue, Ning
2013-01-01
Recent studies have reported potentially clinically meaningful dose differences when heterogeneity correction is used in breast balloon brachytherapy. In this study, we report on the relationship between heterogeneity-corrected and -uncorrected doses for 2 commonly used plan evaluation metrics: maximum point dose to skin surface and maximum point dose to ribs. Maximum point doses to skin surface and ribs were calculated using TG-43 and Varian Acuros for 20 patients treated with breast balloon brachytherapy. The results were plotted against each other and fit with a zero-intercept line. Max skin dose (Acuros) = max skin dose (TG-43) * 0.930 (R(2) = 0.995). The average magnitude of difference from this relationship was 1.1% (max 2.8%). Max rib dose (Acuros) = max rib dose (TG-43) * 0.955 (R(2) = 0.9995). The average magnitude of difference from this relationship was 0.7% (max 1.6%). Heterogeneity-corrected maximum point doses to the skin surface and ribs were proportional to TG-43-calculated doses. The average deviation from proportionality was 1%. The proportional relationship suggests that a different metric other than maximum point dose may be needed to obtain a clinical advantage from heterogeneity correction. Alternatively, if maximum point dose continues to be used in recommended limits while incorporating heterogeneity correction, institutions without this capability may be able to accurately estimate these doses by use of a scaling factor.
2016-12-01
NAVAL POSTGRADUATE SCHOOL MONTEREY, CALIFORNIA THESIS Approved for public release. Distribution is unlimited. IMPACT OF THE ...ACQUISITION CORPS MEMBERSHIP REQUIREMENT “24 BUSINESS-CREDIT HOURS” ON THE NAVY ACQUISITION WORKFORCE by Brady R. Peters December 2016...0704–0188 Public reporting burden for this collection of information is estimated to average one hour per response, including the time for reviewing
2010-01-01
In previous studies, of which several were underpowered, the relation between cardiovascular outcome and blood pressure (BP) variability was inconsistent. We followed health outcomes in 8938 subjects (mean age: 53.0 years; 46.8% women) randomly recruited from 11 populations. At baseline, we...... assessed BP variability from the SD and average real variability in 24-hour ambulatory BP recordings. We computed standardized hazard ratios (HRs) while stratifying by cohort and adjusting for 24-hour BP and other risk factors. Over 11.3 years (median), 1242 deaths (487 cardiovascular) occurred, and 1049......, 577, 421, and 457 participants experienced a fatal or nonfatal cardiovascular, cardiac, or coronary event or a stroke. Higher diastolic average real variability in 24-hour ambulatory BP recordings predicted (Por=1.07) with the exception of cardiac and coronary events (HR: or=0.58). Higher systolic...
Matúš Krčmár
2015-09-01
Full Text Available Background: High muscular power output is required in many athletic endeavors in order for success to be achieved. In the scientific community postactivation potentiation and its effect on performance are often discussed. There are many studies where the effect of resistance exercise on motor performance (such as vertical jump performance and running speed has been investigated but only a few of them studied power output. Objective: The purpose of this study was to determine the acute responses to a 2 set loaded half-squat jumps and 2 set loaded back half-squat protocols designed to induce the acute maximum average power output during loaded half-squat jumps. Methods: A randomized cross-over design was used. 11 participants of this study performed 3 trials in randomized order separated by at least 48 hours where maximum average power output was measured. The specific conditioning activities were comprised of 2 sets and 4 repetitions of half-squat jumps, 2 sets and 4 repetitions of back half-squat exercises and a control protocol without an intervention by specific a conditioning activity. Participants were strength trained athletes with different sport specializations (e.g. ice-hockey, volleyball. Mean age of the athletes was 22 ± 1.8 years, body mass 80 ± 7.1 kg and body height 185 ± 6.5 cm. Analysis of variance with repeated measures was used to determine differences between pre- and post-condition in each protocol, as well as between conditioning protocols, and also effect size was used to evaluate practical significance. Results: Maximum average power was significantly enhanced after application of the half-squat jump condition protocol (1496.2 ± 194.5 to 1552 ± 196.1 W, Δ ~ 3.72%, p < .001 and after application of the back half-squat protocol (1500.7 ± 193.2 to 1556 ± 191.2 W, Δ ~ 3.68%, p < .001 after 10 min of rest. Power output after control protocol was