WorldWideScience

Sample records for maximum cluster capacity

  1. Maximum Throughput in a C-RAN Cluster with Limited Fronthaul Capacity

    OpenAIRE

    Duan , Jialong; Lagrange , Xavier; Guilloud , Frédéric

    2016-01-01

    International audience; Centralized/Cloud Radio Access Network (C-RAN) is a promising future mobile network architecture which can ease the cooperation between different cells to manage interference. However, the feasibility of C-RAN is limited by the large bit rate requirement in the fronthaul. This paper study the maximum throughput of different transmission strategies in a C-RAN cluster with transmission power constraints and fronthaul capacity constraints. Both transmission strategies wit...

  2. Maximum Margin Clustering of Hyperspectral Data

    Science.gov (United States)

    Niazmardi, S.; Safari, A.; Homayouni, S.

    2013-09-01

    In recent decades, large margin methods such as Support Vector Machines (SVMs) are supposed to be the state-of-the-art of supervised learning methods for classification of hyperspectral data. However, the results of these algorithms mainly depend on the quality and quantity of available training data. To tackle down the problems associated with the training data, the researcher put effort into extending the capability of large margin algorithms for unsupervised learning. One of the recent proposed algorithms is Maximum Margin Clustering (MMC). The MMC is an unsupervised SVMs algorithm that simultaneously estimates both the labels and the hyperplane parameters. Nevertheless, the optimization of the MMC algorithm is a non-convex problem. Most of the existing MMC methods rely on the reformulating and the relaxing of the non-convex optimization problem as semi-definite programs (SDP), which are computationally very expensive and only can handle small data sets. Moreover, most of these algorithms are two-class classification, which cannot be used for classification of remotely sensed data. In this paper, a new MMC algorithm is used that solve the original non-convex problem using Alternative Optimization method. This algorithm is also extended for multi-class classification and its performance is evaluated. The results of the proposed algorithm show that the algorithm has acceptable results for hyperspectral data clustering.

  3. Maximum-entropy clustering algorithm and its global convergence analysis

    Institute of Scientific and Technical Information of China (English)

    2001-01-01

    Constructing a batch of differentiable entropy functions touniformly approximate an objective function by means of the maximum-entropy principle, a new clustering algorithm, called maximum-entropy clustering algorithm, is proposed based on optimization theory. This algorithm is a soft generalization of the hard C-means algorithm and possesses global convergence. Its relations with other clustering algorithms are discussed.

  4. Algorithms of maximum likelihood data clustering with applications

    Science.gov (United States)

    Giada, Lorenzo; Marsili, Matteo

    2002-12-01

    We address the problem of data clustering by introducing an unsupervised, parameter-free approach based on maximum likelihood principle. Starting from the observation that data sets belonging to the same cluster share a common information, we construct an expression for the likelihood of any possible cluster structure. The likelihood in turn depends only on the Pearson's coefficient of the data. We discuss clustering algorithms that provide a fast and reliable approximation to maximum likelihood configurations. Compared to standard clustering methods, our approach has the advantages that (i) it is parameter free, (ii) the number of clusters need not be fixed in advance and (iii) the interpretation of the results is transparent. In order to test our approach and compare it with standard clustering algorithms, we analyze two very different data sets: time series of financial market returns and gene expression data. We find that different maximization algorithms produce similar cluster structures whereas the outcome of standard algorithms has a much wider variability.

  5. Maximum physical capacity testing in cancer patients undergoing chemotherapy

    DEFF Research Database (Denmark)

    Knutsen, L.; Quist, M; Midtgaard, J

    2006-01-01

    BACKGROUND: Over the past few years there has been a growing interest in the field of physical exercise in rehabilitation of cancer patients, leading to requirements for objective maximum physical capacity measurement (maximum oxygen uptake (VO(2max)) and one-repetition maximum (1RM)) to determin...... early in the treatment process. However, the patients were self-referred and thus highly motivated and as such are not necessarily representative of the whole population of cancer patients treated with chemotherapy....... in performing maximum physical capacity tests as these motivated them through self-perceived competitiveness and set a standard that served to encourage peak performance. CONCLUSION: The positive attitudes in this sample towards maximum physical capacity open the possibility of introducing physical testing...

  6. Maximum Aerobic Capacity of Underground Coal Miners in India

    Directory of Open Access Journals (Sweden)

    Ratnadeep Saha

    2011-01-01

    Full Text Available Miners fitness test was assessed in terms of determination of maximum aerobic capacity by an indirect method following a standard step test protocol before going down to mine by taking into consideration of heart rates (Telemetric recording and oxygen consumption of the subjects (Oxylog-II during exercise at different working rates. Maximal heart rate was derived as 220−age. Coal miners reported a maximum aerobic capacity within a range of 35–38.3 mL/kg/min. It also revealed that oldest miners (50–59 yrs had a lowest maximal oxygen uptake (34.2±3.38 mL/kg/min compared to (42.4±2.03 mL/kg/min compared to (42.4±2.03 mL/kg/min the youngest group (20–29 yrs. It was found to be negatively correlated with age (r=−0.55 and −0.33 for younger and older groups respectively and directly associated with the body weight of the subjects (r=0.57 – 0.68, P≤0.001. Carriers showed maximum cardio respiratory capacity compared to other miners. Indian miners VO2max was found to be lower both compared to their abroad mining counterparts and various other non-mining occupational working groups in India.

  7. Selection of the Maximum Spatial Cluster Size of the Spatial Scan Statistic by Using the Maximum Clustering Set-Proportion Statistic.

    Science.gov (United States)

    Ma, Yue; Yin, Fei; Zhang, Tao; Zhou, Xiaohua Andrew; Li, Xiaosong

    2016-01-01

    Spatial scan statistics are widely used in various fields. The performance of these statistics is influenced by parameters, such as maximum spatial cluster size, and can be improved by parameter selection using performance measures. Current performance measures are based on the presence of clusters and are thus inapplicable to data sets without known clusters. In this work, we propose a novel overall performance measure called maximum clustering set-proportion (MCS-P), which is based on the likelihood of the union of detected clusters and the applied dataset. MCS-P was compared with existing performance measures in a simulation study to select the maximum spatial cluster size. Results of other performance measures, such as sensitivity and misclassification, suggest that the spatial scan statistic achieves accurate results in most scenarios with the maximum spatial cluster sizes selected using MCS-P. Given that previously known clusters are not required in the proposed strategy, selection of the optimal maximum cluster size with MCS-P can improve the performance of the scan statistic in applications without identified clusters.

  8. Parallel implementation of D-Phylo algorithm for maximum likelihood clusters.

    Science.gov (United States)

    Malik, Shamita; Sharma, Dolly; Khatri, Sunil Kumar

    2017-03-01

    This study explains a newly developed parallel algorithm for phylogenetic analysis of DNA sequences. The newly designed D-Phylo is a more advanced algorithm for phylogenetic analysis using maximum likelihood approach. The D-Phylo while misusing the seeking capacity of k -means keeps away from its real constraint of getting stuck at privately conserved motifs. The authors have tested the behaviour of D-Phylo on Amazon Linux Amazon Machine Image(Hardware Virtual Machine)i2.4xlarge, six central processing unit, 122 GiB memory, 8  ×  800 Solid-state drive Elastic Block Store volume, high network performance up to 15 processors for several real-life datasets. Distributing the clusters evenly on all the processors provides us the capacity to accomplish a near direct speed if there should arise an occurrence of huge number of processors.

  9. Heuristics for minimizing the maximum within-clusters distance

    Directory of Open Access Journals (Sweden)

    José Augusto Fioruci

    2012-12-01

    Full Text Available The clustering problem consists in finding patterns in a data set in order to divide it into clusters with high within-cluster similarity. This paper presents the study of a problem, here called MMD problem, which aims at finding a clustering with a predefined number of clusters that minimizes the largest within-cluster distance (diameter among all clusters. There are two main objectives in this paper: to propose heuristics for the MMD and to evaluate the suitability of the best proposed heuristic results according to the real classification of some data sets. Regarding the first objective, the results obtained in the experiments indicate a good performance of the best proposed heuristic that outperformed the Complete Linkage algorithm (the most used method from the literature for this problem. Nevertheless, regarding the suitability of the results according to the real classification of the data sets, the proposed heuristic achieved better quality results than C-Means algorithm, but worse than Complete Linkage.

  10. Optimizing the maximum reported cluster size in the spatial scan statistic for ordinal data.

    Science.gov (United States)

    Kim, Sehwi; Jung, Inkyung

    2017-01-01

    The spatial scan statistic is an important tool for spatial cluster detection. There have been numerous studies on scanning window shapes. However, little research has been done on the maximum scanning window size or maximum reported cluster size. Recently, Han et al. proposed to use the Gini coefficient to optimize the maximum reported cluster size. However, the method has been developed and evaluated only for the Poisson model. We adopt the Gini coefficient to be applicable to the spatial scan statistic for ordinal data to determine the optimal maximum reported cluster size. Through a simulation study and application to a real data example, we evaluate the performance of the proposed approach. With some sophisticated modification, the Gini coefficient can be effectively employed for the ordinal model. The Gini coefficient most often picked the optimal maximum reported cluster sizes that were the same as or smaller than the true cluster sizes with very high accuracy. It seems that we can obtain a more refined collection of clusters by using the Gini coefficient. The Gini coefficient developed specifically for the ordinal model can be useful for optimizing the maximum reported cluster size for ordinal data and helpful for properly and informatively discovering cluster patterns.

  11. Tri-Laboratory Linux Capacity Cluster 2007 SOW

    International Nuclear Information System (INIS)

    Seager, M.

    2007-01-01

    well, the budget demands are extreme and new, more cost effective ways of fielding these systems must be developed. This Tri-Laboratory Linux Capacity Cluster (TLCC) procurement represents the ASC first investment vehicle in these capacity systems. It also represents a new strategy for quickly building, fielding and integrating many Linux clusters of various sizes into classified and unclassified production service through a concept of Scalable Units (SU). The programmatic objective is to dramatically reduce the overall Total Cost of Ownership (TCO) of these 'capacity' systems relative to the best practices in Linux Cluster deployments today. This objective only makes sense in the context of these systems quickly becoming very robust and useful production clusters under the crushing load that will be inflicted on them by the ASC and SSP scientific simulation capacity workload.

  12. The Local Maximum Clustering Method and Its Application in Microarray Gene Expression Data Analysis

    Directory of Open Access Journals (Sweden)

    Chen Yidong

    2004-01-01

    Full Text Available An unsupervised data clustering method, called the local maximum clustering (LMC method, is proposed for identifying clusters in experiment data sets based on research interest. A magnitude property is defined according to research purposes, and data sets are clustered around each local maximum of the magnitude property. By properly defining a magnitude property, this method can overcome many difficulties in microarray data clustering such as reduced projection in similarities, noises, and arbitrary gene distribution. To critically evaluate the performance of this clustering method in comparison with other methods, we designed three model data sets with known cluster distributions and applied the LMC method as well as the hierarchic clustering method, the -mean clustering method, and the self-organized map method to these model data sets. The results show that the LMC method produces the most accurate clustering results. As an example of application, we applied the method to cluster the leukemia samples reported in the microarray study of Golub et al. (1999.

  13. Maximum-likelihood model averaging to profile clustering of site types across discrete linear sequences.

    Directory of Open Access Journals (Sweden)

    Zhang Zhang

    2009-06-01

    Full Text Available A major analytical challenge in computational biology is the detection and description of clusters of specified site types, such as polymorphic or substituted sites within DNA or protein sequences. Progress has been stymied by a lack of suitable methods to detect clusters and to estimate the extent of clustering in discrete linear sequences, particularly when there is no a priori specification of cluster size or cluster count. Here we derive and demonstrate a maximum likelihood method of hierarchical clustering. Our method incorporates a tripartite divide-and-conquer strategy that models sequence heterogeneity, delineates clusters, and yields a profile of the level of clustering associated with each site. The clustering model may be evaluated via model selection using the Akaike Information Criterion, the corrected Akaike Information Criterion, and the Bayesian Information Criterion. Furthermore, model averaging using weighted model likelihoods may be applied to incorporate model uncertainty into the profile of heterogeneity across sites. We evaluated our method by examining its performance on a number of simulated datasets as well as on empirical polymorphism data from diverse natural alleles of the Drosophila alcohol dehydrogenase gene. Our method yielded greater power for the detection of clustered sites across a breadth of parameter ranges, and achieved better accuracy and precision of estimation of clusters, than did the existing empirical cumulative distribution function statistics.

  14. Tri-Laboratory Linux Capacity Cluster 2007 SOW

    Energy Technology Data Exchange (ETDEWEB)

    Seager, M

    2007-03-22

    ). However, given the growing need for 'capability' systems as well, the budget demands are extreme and new, more cost effective ways of fielding these systems must be developed. This Tri-Laboratory Linux Capacity Cluster (TLCC) procurement represents the ASC first investment vehicle in these capacity systems. It also represents a new strategy for quickly building, fielding and integrating many Linux clusters of various sizes into classified and unclassified production service through a concept of Scalable Units (SU). The programmatic objective is to dramatically reduce the overall Total Cost of Ownership (TCO) of these 'capacity' systems relative to the best practices in Linux Cluster deployments today. This objective only makes sense in the context of these systems quickly becoming very robust and useful production clusters under the crushing load that will be inflicted on them by the ASC and SSP scientific simulation capacity workload.

  15. A Novel Automatic Detection System for ECG Arrhythmias Using Maximum Margin Clustering with Immune Evolutionary Algorithm

    Directory of Open Access Journals (Sweden)

    Bohui Zhu

    2013-01-01

    Full Text Available This paper presents a novel maximum margin clustering method with immune evolution (IEMMC for automatic diagnosis of electrocardiogram (ECG arrhythmias. This diagnostic system consists of signal processing, feature extraction, and the IEMMC algorithm for clustering of ECG arrhythmias. First, raw ECG signal is processed by an adaptive ECG filter based on wavelet transforms, and waveform of the ECG signal is detected; then, features are extracted from ECG signal to cluster different types of arrhythmias by the IEMMC algorithm. Three types of performance evaluation indicators are used to assess the effect of the IEMMC method for ECG arrhythmias, such as sensitivity, specificity, and accuracy. Compared with K-means and iterSVR algorithms, the IEMMC algorithm reflects better performance not only in clustering result but also in terms of global search ability and convergence ability, which proves its effectiveness for the detection of ECG arrhythmias.

  16. Weakly and strongly polynomial algorithms for computing the maximum decrease in uniform arc capacities

    Directory of Open Access Journals (Sweden)

    Ghiyasvand Mehdi

    2016-01-01

    Full Text Available In this paper, a new problem on a directed network is presented. Let D be a feasible network such that all arc capacities are equal to U. Given a t > 0, the network D with arc capacities U - t is called the t-network. The goal of the problem is to compute the largest t such that the t-network is feasible. First, we present a weakly polynomial time algorithm to solve this problem, which runs in O(log(nU maximum flow computations, where n is the number of nodes. Then, an O(m2n time approach is presented, where m is the number of arcs. Both the weakly and strongly polynomial algorithms are inspired by McCormick and Ervolina (1994.

  17. The effects of a pilates-aerobic program on maximum exercise capacity of adult women

    Directory of Open Access Journals (Sweden)

    Milena Mikalački

    Full Text Available ABSTRACT Introduction: Physical exercise such as the Pilates method offers clinical benefits on the aging process. Likewise, physiologic parameters may be improved through aerobic exercise. Methods: In order to compare the differences of a Pilates-Aerobic intervention program on physiologic parameters such as the maximum heart rate (HRmax, relative maximal oxygen consumption (relative VO2max and absolute (absolute VOmax, maximum heart rate during maximal oxygen consumption (VO2max-HRmax, maximum minute volume (VE and forced vital capacity (FVC, a total of 64 adult women (active group = 48.1 ± 6.7 years; control group = 47.2 ± 7.4 years participated in the study. The physiological parameters, the maximal speed and total duration of test were measured by maximum exercise capacity testing through Bruce protocol. The HRmax was calculated by a cardio-ergometric software. Pulmonary function tests, maximal speed and total time during the physical test were performed in a treadmill (Medisoft, model 870c. Likewise, the spirometry analyzed the impact on oxygen uptake parameters, including FVC and VE. Results: The VO2max (relative and absolute, VE (all, P<0.001, VO2max-HRmax (P<0.05 and maximal speed of treadmill test (P<0.001 showed significant difference in the active group after a physical exercise interventional program. Conclusion: The present study indicates that the Pilates exercises through a continuous training program might significantly improve the cardiovascular system. Hence, mixing strength and aerobic exercises into a training program is considered the optimal mechanism for healthy aging.

  18. Effects of Plyometric and Cluster Resistance Training on Explosive Power and Maximum Strength in Karate Players

    Directory of Open Access Journals (Sweden)

    Mohsen Aminaei

    2017-08-01

    Full Text Available The purpose of this study was to investigate the effects of plyometric and cluster resistance training on explosive power and maximum strength in karate players. Eighteen women, karate players (age mean ± SD 18.22 ± 3.02 years, mean height 163 ± 0.63cm, and mean body mass 53.25 ± 7.34 kg were selected as volunteer samples. They were divided into two groups with respect to their recorded one repetition maximum squat exercise: [1] plyometric training (PT=9 and [2] Cluster training (CT=9 groups and performed a 9-week resistance training protocol that included three stages; [1] General fitness (2 weeks, [2] Strength (4 weeks and [3] Power (3 weeks. Each group performed strength and power trainings for 7 weeks in stage two and three with owned protocol. The subjects were evaluated three times before stage one and after two and three stages for maximum strength and power. Data was analyzed using two way Repeated Measures (ANOVA at a significance level of (P≤0.05. The statistical analysis showed that training stages on all research variables had a significant impact. The maximum strength of the pre-test, post-test strength and post-test power were in cluster group: 29.05 ± 1.54; 32.89 ± 2.80 and 48.74 ± 4.33w and in plyometric group were 26.98 ± 1.54; 38.48 ± 2.80 and 49.82 ± 4.33w respectively. The explosive power of the pre-test, post-test strength and post-test power in cluster group were 359.32±36.20; 427.91±34.56 and 460.55±36.80w and in plyometric group were 333.90±36.20; 400.33±34.56 and 465.20±36.80w respectively. However, there were not statistically significant differences in research variables between resistance cluster and plyometric training groups after 7 weeks. The results indicated both cluster and plyometric training program seems to improve physical fitness elements at the same levels.

  19. The maximum reservoir capacity of soils for persistent organic pollutants: implications for global cycling

    International Nuclear Information System (INIS)

    Dalla Valle, M.; Jurado, E.; Dachs, J.; Sweetman, A.J.; Jones, K.C.

    2005-01-01

    The concept of maximum reservoir capacity (MRC), the ratio of the capacities of the surface soil and of the atmospheric mixed layer (AML) to hold chemical under equilibrium conditions, is applied to selected persistent organic pollutants (POPs) in the surface 'skin' (1 mm) of soils. MRC is calculated as a function of soil organic matter (SOM) content and temperature-dependent K OA and mapped globally for selected PCB congeners (PCB-28; -153; -180) and HCB, to identify regions with a higher tendency to retain POPs. It is shown to vary over many orders of magnitude, between compounds, locations and time (seasonally/diurnally). The MRC approach emphasises the very large capacity of soils as a storage compartment for POPs. The theoretical MRC concept is compared to reality and its implications for the global cycling of POPs are discussed. Sharp gradients in soil MRC can exist in mountainous areas and between the land and ocean. Exchanges between oceans and land masses via the atmosphere is likely to be an important driver to the global cycling of these compounds, and net ocean-land transfers could occur in some areas. - Major global terrestrial sinks/stores for POPs are identified and the significance of gradients between them discussed

  20. Chaos control of ferroresonance system based on RBF-maximum entropy clustering algorithm

    International Nuclear Information System (INIS)

    Liu Fan; Sun Caixin; Sima Wenxia; Liao Ruijin; Guo Fei

    2006-01-01

    With regards to the ferroresonance overvoltage of neutral grounded power system, a maximum-entropy learning algorithm based on radial basis function neural networks is used to control the chaotic system. The algorithm optimizes the object function to derive learning rule of central vectors, and uses the clustering function of network hidden layers. It improves the regression and learning ability of neural networks. The numerical experiment of ferroresonance system testifies the effectiveness and feasibility of using the algorithm to control chaos in neutral grounded system

  1. Bayesian modeling of the assimilative capacity component of nutrient total maximum daily loads

    Science.gov (United States)

    Faulkner, B. R.

    2008-08-01

    Implementing stream restoration techniques and best management practices to reduce nonpoint source nutrients implies enhancement of the assimilative capacity for the stream system. In this paper, a Bayesian method for evaluating this component of a total maximum daily load (TMDL) load capacity is developed and applied. The joint distribution of nutrient retention metrics from a literature review of 495 measurements was used for Monte Carlo sampling with a process transfer function for nutrient attenuation. Using the resulting histograms of nutrient retention, reference prior distributions were developed for sites in which some of the metrics contributing to the transfer function were measured. Contributing metrics for the prior include stream discharge, cross-sectional area, fraction of storage volume to free stream volume, denitrification rate constant, storage zone mass transfer rate, dispersion coefficient, and others. Confidence of compliance (CC) that any given level of nutrient retention has been achieved is also determined using this approach. The shape of the CC curve is dependent on the metrics measured and serves in part as a measure of the information provided by the metrics to predict nutrient retention. It is also a direct measurement, with a margin of safety, of the fraction of export load that can be reduced through changing retention metrics. For an impaired stream in western Oklahoma, a combination of prior information and measurement of nutrient attenuation was used to illustrate the proposed approach. This method may be considered for TMDL implementation.

  2. Stability of maximum-likelihood-based clustering methods: exploring the backbone of classifications

    International Nuclear Information System (INIS)

    Mungan, Muhittin; Ramasco, José J

    2010-01-01

    Components of complex systems are often classified according to the way they interact with each other. In graph theory such groups are known as clusters or communities. Many different techniques have been recently proposed to detect them, some of which involve inference methods using either Bayesian or maximum likelihood approaches. In this paper, we study a statistical model designed for detecting clusters based on connection similarity. The basic assumption of the model is that the graph was generated by a certain grouping of the nodes and an expectation maximization algorithm is employed to infer that grouping. We show that the method admits further development to yield a stability analysis of the groupings that quantifies the extent to which each node influences its neighbors' group membership. Our approach naturally allows for the identification of the key elements responsible for the grouping and their resilience to changes in the network. Given the generality of the assumptions underlying the statistical model, such nodes are likely to play special roles in the original system. We illustrate this point by analyzing several empirical networks for which further information about the properties of the nodes is available. The search and identification of stabilizing nodes constitutes thus a novel technique to characterize the relevance of nodes in complex networks

  3. Liquid films on shake flask walls explain increasing maximum oxygen transfer capacities with elevating viscosity.

    Science.gov (United States)

    Giese, Heiner; Azizan, Amizon; Kümmel, Anne; Liao, Anping; Peter, Cyril P; Fonseca, João A; Hermann, Robert; Duarte, Tiago M; Büchs, Jochen

    2014-02-01

    In biotechnological screening and production, oxygen supply is a crucial parameter. Even though oxygen transfer is well documented for viscous cultivations in stirred tanks, little is known about the gas/liquid oxygen transfer in shake flask cultures that become increasingly viscous during cultivation. Especially the oxygen transfer into the liquid film, adhering on the shake flask wall, has not yet been described for such cultivations. In this study, the oxygen transfer of chemical and microbial model experiments was measured and the suitability of the widely applied film theory of Higbie was studied. With numerical simulations of Fick's law of diffusion, it was demonstrated that Higbie's film theory does not apply for cultivations which occur at viscosities up to 10 mPa s. For the first time, it was experimentally shown that the maximum oxygen transfer capacity OTRmax increases in shake flasks when viscosity is increased from 1 to 10 mPa s, leading to an improved oxygen supply for microorganisms. Additionally, the OTRmax does not significantly undermatch the OTRmax at waterlike viscosities, even at elevated viscosities of up to 80 mPa s. In this range, a shake flask is a somehow self-regulating system with respect to oxygen supply. This is in contrary to stirred tanks, where the oxygen supply is steadily reduced to only 5% at 80 mPa s. Since, the liquid film formation at shake flask walls inherently promotes the oxygen supply at moderate and at elevated viscosities, these results have significant implications for scale-up. © 2013 Wiley Periodicals, Inc.

  4. It is time to abandon "expected bladder capacity." Systematic review and new models for children's normal maximum voided volumes.

    Science.gov (United States)

    Martínez-García, Roberto; Ubeda-Sansano, Maria Isabel; Díez-Domingo, Javier; Pérez-Hoyos, Santiago; Gil-Salom, Manuel

    2014-09-01

    There is an agreement to use simple formulae (expected bladder capacity and other age based linear formulae) as bladder capacity benchmark. But real normal child's bladder capacity is unknown. To offer a systematic review of children's normal bladder capacity, to measure children's normal maximum voided volumes (MVVs), to construct models of MVVs and to compare them with the usual formulae. Computerized, manual and grey literature were reviewed until February 2013. Epidemiological, observational, transversal, multicenter study. A consecutive sample of healthy children aged 5-14 years, attending Primary Care centres with no urologic abnormality were selected. Participants filled-in a 3-day frequency-volume chart. Variables were MVVs: maximum of 24 hr, nocturnal, and daytime maximum voided volumes. diuresis and its daytime and nighttime fractions; body-measure data; and gender. The consecutive steps method was used in a multivariate regression model. Twelve articles accomplished systematic review's criteria. Five hundred and fourteen cases were analysed. Three models, one for each of the MVVs, were built. All of them were better adjusted to exponential equations. Diuresis (not age) was the most significant factor. There was poor agreement between MVVs and usual formulae. Nocturnal and daytime maximum voided volumes depend on several factors and are different. Nocturnal and daytime maximum voided volumes should be used with different meanings in clinical setting. Diuresis is the main factor for bladder capacity. This is the first model for benchmarking normal MVVs with diuresis as its main factor. Current formulae are not suitable for clinical use. © 2013 Wiley Periodicals, Inc.

  5. On the low SNR capacity of maximum ratio combining over rician fading channels with full channel state information

    KAUST Repository

    Benkhelifa, Fatma

    2013-04-01

    In this letter, we study the ergodic capacity of a maximum ratio combining (MRC) Rician fading channel with full channel state information (CSI) at the transmitter and at the receiver. We focus on the low Signal-to-Noise Ratio (SNR) regime and we show that the capacity scales as L ΩK+L SNRx log(1SNR), where Ω is the expected channel gain per branch, K is the Rician fading factor, and L is the number of diversity branches. We show that one-bit CSI feedback at the transmitter is enough to achieve this capacity using an on-off power control scheme. Our framework can be seen as a generalization of recently established results regarding the fading-channels capacity characterization in the low-SNR regime. © 2012 IEEE.

  6. On the low SNR capacity of maximum ratio combining over rician fading channels with full channel state information

    KAUST Repository

    Benkhelifa, Fatma; Rezki, Zouheir; Alouini, Mohamed-Slim

    2013-01-01

    In this letter, we study the ergodic capacity of a maximum ratio combining (MRC) Rician fading channel with full channel state information (CSI) at the transmitter and at the receiver. We focus on the low Signal-to-Noise Ratio (SNR) regime and we show that the capacity scales as L ΩK+L SNRx log(1SNR), where Ω is the expected channel gain per branch, K is the Rician fading factor, and L is the number of diversity branches. We show that one-bit CSI feedback at the transmitter is enough to achieve this capacity using an on-off power control scheme. Our framework can be seen as a generalization of recently established results regarding the fading-channels capacity characterization in the low-SNR regime. © 2012 IEEE.

  7. Eigenstructures of MIMO Fading Channel Correlation Matrices and Optimum Linear Precoding Designs for Maximum Ergodic Capacity

    Directory of Open Access Journals (Sweden)

    Hamid Reza Bahrami

    2007-01-01

    Full Text Available The ergodic capacity of MIMO frequency-flat and -selective channels depends greatly on the eigenvalue distribution of spatial correlation matrices. Knowing the eigenstructure of correlation matrices at the transmitter is very important to enhance the capacity of the system. This fact becomes of great importance in MIMO wireless systems where because of the fast changing nature of the underlying channel, full channel knowledge is difficult to obtain at the transmitter. In this paper, we first investigate the effect of eigenvalues distribution of spatial correlation matrices on the capacity of frequency-flat and -selective channels. Next, we introduce a practical scheme known as linear precoding that can enhance the ergodic capacity of the channel by changing the eigenstructure of the channel by applying a linear transformation. We derive the structures of precoders using eigenvalue decomposition and linear algebra techniques in both cases and show their similarities from an algebraic point of view. Simulations show the ability of this technique to change the eigenstructure of the channel, and hence enhance the ergodic capacity considerably.

  8. Assessment of Maximum Aerobic Capacity and Anaerobic Threshold of Elite Ballet Dancers.

    Science.gov (United States)

    Wyon, Matthew A; Allen, Nick; Cloak, Ross; Beck, Sarah; Davies, Paul; Clarke, Frances

    2016-09-01

    An athlete's cardiorespiratory profile, maximal aerobic capacity, and anaerobic threshold is affected by training regimen and competition demands. The present study aimed to ascertain whether there are company rank differences in maximal aerobic capacity and anaerobic threshold in elite classical ballet dancers. Seventy-four volunteers (M 34, F 40) were recruited from two full-time professional classical ballet companies. All participants completed a continuous incremental treadmill protocol with a 1-km/hr speed increase at the end of each 1-min stage until termination criteria had been achieved (e.g., voluntary cessation, respiratory exchange ratio ballet companies are probably due to the different rehearsal and performance demands.

  9. Effects of different strength training frequencies on maximum strength, body composition and functional capacity in healthy older individuals.

    Science.gov (United States)

    Turpela, Mari; Häkkinen, Keijo; Haff, Guy Gregory; Walker, Simon

    2017-11-01

    There is controversy in the literature regarding the dose-response relationship of strength training in healthy older participants. The present study determined training frequency effects on maximum strength, muscle mass and functional capacity over 6months following an initial 3-month preparatory strength training period. One-hundred and six 64-75year old volunteers were randomly assigned to one of four groups; performing strength training one (EX1), two (EX2), or three (EX3) times per week and a non-training control (CON) group. Whole-body strength training was performed using 2-5 sets and 4-12 repetitions per exercise and 7-9 exercises per session. Before and after the intervention, maximum dynamic leg press (1-RM) and isometric knee extensor and plantarflexor strength, body composition and quadriceps cross-sectional area, as well as functional capacity (maximum 7.5m forward and backward walking speed, timed-up-and-go test, loaded 10-stair climb test) were measured. All experimental groups increased leg press 1-RM more than CON (EX1: 3±8%, EX2: 6±6%, EX3: 10±8%, CON: -3±6%, Ptraining frequency would induce greater benefit to maximum walking speed (i.e. functional capacity) despite a clear dose-response in dynamic 1-RM strength, at least when predominantly using machine weight-training. It appears that beneficial functional capacity improvements can be achieved through low frequency training (i.e. 1-2 times per week) in previously untrained healthy older participants. Copyright © 2017 Elsevier Inc. All rights reserved.

  10. Measuring galaxy cluster masses with CMB lensing using a Maximum Likelihood estimator: statistical and systematic error budgets for future experiments

    Energy Technology Data Exchange (ETDEWEB)

    Raghunathan, Srinivasan; Patil, Sanjaykumar; Bianchini, Federico; Reichardt, Christian L. [School of Physics, University of Melbourne, 313 David Caro building, Swanston St and Tin Alley, Parkville VIC 3010 (Australia); Baxter, Eric J. [Department of Physics and Astronomy, University of Pennsylvania, 209 S. 33rd Street, Philadelphia, PA 19104 (United States); Bleem, Lindsey E. [Argonne National Laboratory, High-Energy Physics Division, 9700 S. Cass Avenue, Argonne, IL 60439 (United States); Crawford, Thomas M. [Kavli Institute for Cosmological Physics, University of Chicago, 5640 South Ellis Avenue, Chicago, IL 60637 (United States); Holder, Gilbert P. [Department of Astronomy and Department of Physics, University of Illinois, 1002 West Green St., Urbana, IL 61801 (United States); Manzotti, Alessandro, E-mail: srinivasan.raghunathan@unimelb.edu.au, E-mail: s.patil2@student.unimelb.edu.au, E-mail: ebax@sas.upenn.edu, E-mail: federico.bianchini@unimelb.edu.au, E-mail: bleeml@uchicago.edu, E-mail: tcrawfor@kicp.uchicago.edu, E-mail: gholder@illinois.edu, E-mail: manzotti@uchicago.edu, E-mail: christian.reichardt@unimelb.edu.au [Department of Astronomy and Astrophysics, University of Chicago, 5640 South Ellis Avenue, Chicago, IL 60637 (United States)

    2017-08-01

    We develop a Maximum Likelihood estimator (MLE) to measure the masses of galaxy clusters through the impact of gravitational lensing on the temperature and polarization anisotropies of the cosmic microwave background (CMB). We show that, at low noise levels in temperature, this optimal estimator outperforms the standard quadratic estimator by a factor of two. For polarization, we show that the Stokes Q/U maps can be used instead of the traditional E- and B-mode maps without losing information. We test and quantify the bias in the recovered lensing mass for a comprehensive list of potential systematic errors. Using realistic simulations, we examine the cluster mass uncertainties from CMB-cluster lensing as a function of an experiment's beam size and noise level. We predict the cluster mass uncertainties will be 3 - 6% for SPT-3G, AdvACT, and Simons Array experiments with 10,000 clusters and less than 1% for the CMB-S4 experiment with a sample containing 100,000 clusters. The mass constraints from CMB polarization are very sensitive to the experimental beam size and map noise level: for a factor of three reduction in either the beam size or noise level, the lensing signal-to-noise improves by roughly a factor of two.

  11. Assessment of precast beam-column using capacity demand response spectrum subject to design basis earthquake and maximum considered earthquake

    Science.gov (United States)

    Ghani, Kay Dora Abd.; Tukiar, Mohd Azuan; Hamid, Nor Hayati Abdul

    2017-08-01

    Malaysia is surrounded by the tectonic feature of the Sumatera area which consists of two seismically active inter-plate boundaries, namely the Indo-Australian and the Eurasian Plates on the west and the Philippine Plates on the east. Hence, Malaysia experiences tremors from far distant earthquake occurring in Banda Aceh, Nias Island, Padang and other parts of Sumatera Indonesia. In order to predict the safety of precast buildings in Malaysia under near field ground motion the response spectrum analysis could be used for dealing with future earthquake whose specific nature is unknown. This paper aimed to develop of capacity demand response spectrum subject to Design Basis Earthquake (DBE) and Maximum Considered Earthquake (MCE) in order to assess the performance of precast beam column joint. From the capacity-demand response spectrum analysis, it can be concluded that the precast beam-column joints would not survive when subjected to earthquake excitation with surface-wave magnitude, Mw, of more than 5.5 Scale Richter (Type 1 spectra). This means that the beam-column joint which was designed using the current code of practice (BS8110) would be severely damaged when subjected to high earthquake excitation. The capacity-demand response spectrum analysis also shows that the precast beam-column joints in the prototype studied would be severely damaged when subjected to Maximum Considered Earthquake (MCE) with PGA=0.22g having a surface-wave magnitude of more than 5.5 Scale Richter, or Type 1 spectra.

  12. Mathematical description and program documentation for CLASSY, an adaptive maximum likelihood clustering method

    Science.gov (United States)

    Lennington, R. K.; Rassbach, M. E.

    1979-01-01

    Discussed in this report is the clustering algorithm CLASSY, including detailed descriptions of its general structure and mathematical background and of the various major subroutines. The report provides a development of the logic and equations used with specific reference to program variables. Some comments on timing and proposed optimization techniques are included.

  13. Developing the fuzzy c-means clustering algorithm based on maximum entropy for multitarget tracking in a cluttered environment

    Science.gov (United States)

    Chen, Xiao; Li, Yaan; Yu, Jing; Li, Yuxing

    2018-01-01

    For fast and more effective implementation of tracking multiple targets in a cluttered environment, we propose a multiple targets tracking (MTT) algorithm called maximum entropy fuzzy c-means clustering joint probabilistic data association that combines fuzzy c-means clustering and the joint probabilistic data association (PDA) algorithm. The algorithm uses the membership value to express the probability of the target originating from measurement. The membership value is obtained through fuzzy c-means clustering objective function optimized by the maximum entropy principle. When considering the effect of the public measurement, we use a correction factor to adjust the association probability matrix to estimate the state of the target. As this algorithm avoids confirmation matrix splitting, it can solve the high computational load problem of the joint PDA algorithm. The results of simulations and analysis conducted for tracking neighbor parallel targets and cross targets in a different density cluttered environment show that the proposed algorithm can realize MTT quickly and efficiently in a cluttered environment. Further, the performance of the proposed algorithm remains constant with increasing process noise variance. The proposed algorithm has the advantages of efficiency and low computational load, which can ensure optimum performance when tracking multiple targets in a dense cluttered environment.

  14. Maximum capacity model of grid-connected multi-wind farms considering static security constraints in electrical grids

    International Nuclear Information System (INIS)

    Zhou, W; Oodo, S O; He, H; Qiu, G Y

    2013-01-01

    An increasing interest in wind energy and the advance of related technologies have increased the connection of wind power generation into electrical grids. This paper proposes an optimization model for determining the maximum capacity of wind farms in a power system. In this model, generator power output limits, voltage limits and thermal limits of branches in the grid system were considered in order to limit the steady-state security influence of wind generators on the power system. The optimization model was solved by a nonlinear primal-dual interior-point method. An IEEE-30 bus system with two wind farms was tested through simulation studies, plus an analysis conducted to verify the effectiveness of the proposed model. The results indicated that the model is efficient and reasonable.

  15. Maximum capacity model of grid-connected multi-wind farms considering static security constraints in electrical grids

    Science.gov (United States)

    Zhou, W.; Qiu, G. Y.; Oodo, S. O.; He, H.

    2013-03-01

    An increasing interest in wind energy and the advance of related technologies have increased the connection of wind power generation into electrical grids. This paper proposes an optimization model for determining the maximum capacity of wind farms in a power system. In this model, generator power output limits, voltage limits and thermal limits of branches in the grid system were considered in order to limit the steady-state security influence of wind generators on the power system. The optimization model was solved by a nonlinear primal-dual interior-point method. An IEEE-30 bus system with two wind farms was tested through simulation studies, plus an analysis conducted to verify the effectiveness of the proposed model. The results indicated that the model is efficient and reasonable.

  16. Adaptive capacity of geographical clusters: Complexity science and network theory approach

    Science.gov (United States)

    Albino, Vito; Carbonara, Nunzia; Giannoccaro, Ilaria

    This paper deals with the adaptive capacity of geographical clusters (GCs), that is a relevant topic in the literature. To address this topic, GC is considered as a complex adaptive system (CAS). Three theoretical propositions concerning the GC adaptive capacity are formulated by using complexity theory. First, we identify three main properties of CAS s that affect the adaptive capacity, namely the interconnectivity, the heterogeneity, and the level of control, and define how the value of these properties influence the adaptive capacity. Then, we associate these properties with specific GC characteristics so obtaining the key conditions of GCs that give them the adaptive capacity so assuring their competitive advantage. To test these theoretical propositions, a case study on two real GCs is carried out. The considered GCs are modeled as networks where firms are nodes and inter-firms relationships are links. Heterogeneity, interconnectivity, and level of control are considered as network properties and thus measured by using the methods of the network theory.

  17. COMPARISON OF NORMALIZED MAXIMUM AEROBIC CAPACITY AND BODY COMPOSITION OF SUMO WRESTLERS TO ATHLETES IN COMBAT AND OTHER SPORTS

    Directory of Open Access Journals (Sweden)

    Matthew D. Beekley

    2006-07-01

    Full Text Available Sumo wrestling is unique in combat sport, and in all of sport. We examined the maximum aerobic capacity and body composition of sumo wrestlers and compared them to untrained controls. We also compared "aerobic muscle quality", meaning VO2max normalized to predicted skeletal muscle mass (SMM (VO2max /SMM, between sumo wrestlers and controls and among previously published data for male athletes from combat, aerobic, and power sports. Sumo wrestlers, compared to untrained controls, had greater (p < 0.05 body mass (mean ± SD; 117.0 ± 4.9 vs. 56.1 ± 9.8 kg, percent fat (24.0 ± 1.4 vs. 13.3 ± 4.5, fat-free mass (88.9 ± 4.2 vs. 48.4 �� 6.8 kg, predicted SMM (48.2 ± 2.9 vs. 20.6 ± 4.7 kg and absolute VO2max (3.6 ± 1.3 vs. 2.5 ± 0.7 L·min-1. Mean VO2max /SMM (ml·kg SMM-1·min-1 was significantly different (p < 0.05 among aerobic athletes (164.8 ± 18.3, combat athletes (which was not different from untrained controls; 131.4 ± 9.3 and 128.6 ± 13.6, respectively, power athletes (96.5 ± 5.3, and sumo wrestlers (71.4 ± 5.3. There was a strong negative correlation (r = - 0.75 between percent body fat and VO2max /SMM (p < 0.05. We conclude that sumo wrestlers have some of the largest percent body fat and fat-free mass and the lowest "aerobic muscle quality" (VO2max /SMM, both in combat sport and compared to aerobic and power sport athletes. Additionally, it appears from analysis of the relationship between SMM and absolute VO2max for all sports that there is a "ceiling" at which increases in SMM do not result in additional increases in absolute VO2max

  18. GAS SURFACE DENSITY, STAR FORMATION RATE SURFACE DENSITY, AND THE MAXIMUM MASS OF YOUNG STAR CLUSTERS IN A DISK GALAXY. II. THE GRAND-DESIGN GALAXY M51

    International Nuclear Information System (INIS)

    González-Lópezlira, Rosa A.; Pflamm-Altenburg, Jan; Kroupa, Pavel

    2013-01-01

    We analyze the relationship between maximum cluster mass and surface densities of total gas (Σ gas ), molecular gas (Σ H 2 ), neutral gas (Σ H I ), and star formation rate (Σ SFR ) in the grand-design galaxy M51, using published gas data and a catalog of masses, ages, and reddenings of more than 1800 star clusters in its disk, of which 223 are above the cluster mass distribution function completeness limit. By comparing the two-dimensional distribution of cluster masses and gas surface densities, we find for clusters older than 25 Myr that M 3rd ∝Σ H I 0.4±0.2 , whereM 3rd is the median of the five most massive clusters. There is no correlation withΣ gas ,Σ H2 , orΣ SFR . For clusters younger than 10 Myr, M 3rd ∝Σ H I 0.6±0.1 and M 3rd ∝Σ gas 0.5±0.2 ; there is no correlation with either Σ H 2 orΣ SFR . The results could hardly be more different from those found for clusters younger than 25 Myr in M33. For the flocculent galaxy M33, there is no correlation between maximum cluster mass and neutral gas, but we have determined M 3rd ∝Σ gas 3.8±0.3 , M 3rd ∝Σ H 2 1.2±0.1 , and M 3rd ∝Σ SFR 0.9±0.1 . For the older sample in M51, the lack of tight correlations is probably due to the combination of strong azimuthal variations in the surface densities of gas and star formation rate, and the cluster ages. These two facts mean that neither the azimuthal average of the surface densities at a given radius nor the surface densities at the present-day location of a stellar cluster represent the true surface densities at the place and time of cluster formation. In the case of the younger sample, even if the clusters have not yet traveled too far from their birth sites, the poor resolution of the radio data compared to the physical sizes of the clusters results in measuredΣ that are likely quite diluted compared to the actual densities relevant for the formation of the clusters.

  19. Maximum Acceptable Weight of Lift reflects peak lumbosacral extension moments in a Functional Capacity Evaluation test using free style, stoop, and squat lifting

    OpenAIRE

    Kuijer, P.P.F.M.; van Oostrom, S.H.; Duijzer, K.; van Dieen, J.H.

    2012-01-01

    It is unclear whether the maximum acceptable weight of lift (MAWL), a common psychophysical method, reflects joint kinetics when different lifting techniques are employed. In a within-participants study (n = 12), participants performed three lifting techniques - free style, stoop and squat lifting from knee to waist level - using the same dynamic functional capacity evaluation lifting test to assess MAWL and to calculate low back and knee kinetics. We assessed which knee and back kinetic para...

  20. The effects of linear accelerations on the maximum heat transfer capacity of micro pipes with triangular grooves

    International Nuclear Information System (INIS)

    Shokouhmand, H.; Kahrobaian, A.; Tabandeh, N.; Jalilvand, A.

    2002-01-01

    Micro heat pipes are widely used for the thermal control of spacecraft and their electronic components. In this paper the influence of linear accelerations in micro grooves has been studied. A mathematical model for predicating the minimum meniscus radius and the maximum heat transport in triangular groove under the influence of linear acceleration is presented and method for determining the theoretical minimum meniscus radius is developed. It is shown that both, the direction and the magnitude of the acceleration have a great effect upon heat transfer capability of micro heat pipes. The analysis presented here provides a mechanism where by the groove geometry can be optimized with respect to the length of the heat pipe and direction and magnitude of linear acceleration

  1. A bottom-up approach to identifying the maximum operational adaptive capacity of water resource systems to a changing climate

    Science.gov (United States)

    Culley, S.; Noble, S.; Yates, A.; Timbs, M.; Westra, S.; Maier, H. R.; Giuliani, M.; Castelletti, A.

    2016-09-01

    Many water resource systems have been designed assuming that the statistical characteristics of future inflows are similar to those of the historical record. This assumption is no longer valid due to large-scale changes in the global climate, potentially causing declines in water resource system performance, or even complete system failure. Upgrading system infrastructure to cope with climate change can require substantial financial outlay, so it might be preferable to optimize existing system performance when possible. This paper builds on decision scaling theory by proposing a bottom-up approach to designing optimal feedback control policies for a water system exposed to a changing climate. This approach not only describes optimal operational policies for a range of potential climatic changes but also enables an assessment of a system's upper limit of its operational adaptive capacity, beyond which upgrades to infrastructure become unavoidable. The approach is illustrated using the Lake Como system in Northern Italy—a regulated system with a complex relationship between climate and system performance. By optimizing system operation under different hydrometeorological states, it is shown that the system can continue to meet its minimum performance requirements for more than three times as many states as it can under current operations. Importantly, a single management policy, no matter how robust, cannot fully utilize existing infrastructure as effectively as an ensemble of flexible management policies that are updated as the climate changes.

  2. A new approach to hierarchical data analysis: Targeted maximum likelihood estimation for the causal effect of a cluster-level exposure.

    Science.gov (United States)

    Balzer, Laura B; Zheng, Wenjing; van der Laan, Mark J; Petersen, Maya L

    2018-01-01

    We often seek to estimate the impact of an exposure naturally occurring or randomly assigned at the cluster-level. For example, the literature on neighborhood determinants of health continues to grow. Likewise, community randomized trials are applied to learn about real-world implementation, sustainability, and population effects of interventions with proven individual-level efficacy. In these settings, individual-level outcomes are correlated due to shared cluster-level factors, including the exposure, as well as social or biological interactions between individuals. To flexibly and efficiently estimate the effect of a cluster-level exposure, we present two targeted maximum likelihood estimators (TMLEs). The first TMLE is developed under a non-parametric causal model, which allows for arbitrary interactions between individuals within a cluster. These interactions include direct transmission of the outcome (i.e. contagion) and influence of one individual's covariates on another's outcome (i.e. covariate interference). The second TMLE is developed under a causal sub-model assuming the cluster-level and individual-specific covariates are sufficient to control for confounding. Simulations compare the alternative estimators and illustrate the potential gains from pairing individual-level risk factors and outcomes during estimation, while avoiding unwarranted assumptions. Our results suggest that estimation under the sub-model can result in bias and misleading inference in an observational setting. Incorporating working assumptions during estimation is more robust than assuming they hold in the underlying causal model. We illustrate our approach with an application to HIV prevention and treatment.

  3. Liposomal solubilization of new 3-hydroxy-quinolinone derivatives with promising anticancer activity: a screening method to identify maximum incorporation capacity

    DEFF Research Database (Denmark)

    Di Cagno, Massimiliano; Styskala, Jakub; Hlaváč, Jan

    2011-01-01

    Four new 3-hydroxy-quinolinone derivatives with promising anticancer activity could be solubilized using liposomes as vehicle to an extent that allows their in vitro and in vivo testing without use of toxic solvent(s). A screening method to identify the maximum incorporation capacity of hydrophobic......, resulting in a 200-500-fold increase in apparent solubility. Drug-to-lipid ratios in the range of 2-5 µg/mg were obtained. Interestingly, the four quinolinone derivatives have shown different association tendencies with liposomes, probably due to the physicochemical properties of the different group bonded...

  4. Maximum acceptable weight of lift reflects peak lumbosacral extension moments in a functional capacity evaluation test using free style, stoop and squat lifting.

    Science.gov (United States)

    Kuijer, P P F M; van Oostrom, S H; Duijzer, K; van Dieën, J H

    2012-01-01

    It is unclear whether the maximum acceptable weight of lift (MAWL), a common psychophysical method, reflects joint kinetics when different lifting techniques are employed. In a within-participants study (n = 12), participants performed three lifting techniques--free style, stoop and squat lifting from knee to waist level--using the same dynamic functional capacity evaluation lifting test to assess MAWL and to calculate low back and knee kinetics. We assessed which knee and back kinetic parameters increased with the load mass lifted, and whether the magnitudes of the kinetic parameters were consistent across techniques when lifting MAWL. MAWL was significantly different between techniques (p = 0.03). The peak lumbosacral extension moment met both criteria: it had the highest association with the load masses lifted (r > 0.9) and was most consistent between the three techniques when lifting MAWL (ICC = 0.87). In conclusion, MAWL reflects the lumbosacral extension moment across free style, stoop and squat lifting in healthy young males, but the relation between the load mass lifted and lumbosacral extension moment is different between techniques. Tests of maximum acceptable weight of lift (MAWL) from knee to waist height are used to assess work capacity of individuals with low-back disorders. This article shows that the MAWL reflects the lumbosacral extension moment across free style, stoop and squat lifting in healthy young males, but the relation between the load mass lifted and lumbosacral extension moment is different between techniques. This suggests that standardisation of lifting technique used in tests of the MAWL would be indicated if the aim is to assess the capacity of the low back.

  5. Synthesis and characterization of osmium carbonyl cluster compounds with molecular oxygen electroreduction capacity

    Energy Technology Data Exchange (ETDEWEB)

    Castellanos, R.H.; Ocampo, A.L.; Moreira-Acosta, J.; Sebastian, P.J. [CIE-UNAM Solar Energy Laboratory, Morelos (Mexico). Photovoltaic Systems Group, Solar-Hydrogen-Fuel Cell

    2001-12-01

    A transition metal cluster electrocatalyst based on Os{sub x}(CO){sub n} was synthesized by pyrolysis of Os{sub 3} (CO){sub 12} in 1,2-Dichlorobenzene (b.p.{approx_equal}180{sup o}C) under inert atmosphere (N{sub 2}). The electrocatalytic parameters of the oxygen reduction reaction (ORR) for an Os{sub x}(CO){sub n} catalyst were studied with a rotating disk electrode in 0.5 MH{sub 2}SO{sub 4} electrolyte. The diffusion coefficient and solubility of O{sub 2} in 0.5 MH{sub 2}SO{sub 4} were calculated. Koutecky-Levich analysis of the linear voltamperometry data showed that the reaction follows first-order kinetics and the value of the Koutecky-Levich slope indicates a multielectron charge transfer during the ORR. The value of the Tafel slope obtained from the mass transfer corrected Tafel plots is 131 mV/decade. The performance of the catalyst in a H{sub 2}/O{sub 2} PEM fuel cell cathode was evaluated and found to be nearly as good as that of Pt. (author)

  6. clusters

    Indian Academy of Sciences (India)

    2017-09-27

    Sep 27, 2017 ... Author for correspondence (zh4403701@126.com). MS received 15 ... lic clusters using density functional theory (DFT)-GGA of the DMOL3 package. ... In the process of geometric optimization, con- vergence thresholds ..... and Postgraduate Research & Practice Innovation Program of. Jiangsu Province ...

  7. clusters

    Indian Academy of Sciences (India)

    environmental as well as technical problems during fuel gas utilization. ... adsorption on some alloys of Pd, namely PdAu, PdAg ... ried out on small neutral and charged Au24,26,27, Cu,28 ... study of Zanti et al.29 on Pdn (n = 1–9) clusters.

  8. Maximum Acceptable Weight of Lift reflects peak lumbosacral extension moments in a Functional Capacity Evaluation test using free style, stoop, and squat lifting

    NARCIS (Netherlands)

    Kuijer, P.P.F.M.; van Oostrom, S.H.; Duijzer, K.; van Dieen, J.H.

    2012-01-01

    It is unclear whether the maximum acceptable weight of lift (MAWL), a common psychophysical method, reflects joint kinetics when different lifting techniques are employed. In a within-participants study (n = 12), participants performed three lifting techniques - free style, stoop and squat lifting

  9. Maximum acceptable weight of lift reflects peak lumbosacral extension moments in a functional capacity evaluation test using free style, stoop and squat lifting

    NARCIS (Netherlands)

    Kuijer, P. P. F. M.; van Oostrom, S. H.; Duijzer, K.; van Dieën, J. H.

    2012-01-01

    It is unclear whether the maximum acceptable weight of lift (MAWL), a common psychophysical method, reflects joint kinetics when different lifting techniques are employed. In a within-participants study (n = 12), participants performed three lifting techniques - free style, stoop and squat lifting

  10. [Optimization for MSW logistics of new Xicheng and new Dongcheng districts in Beijing based on the maximum capacity of transfer stations].

    Science.gov (United States)

    Yuan, Jing; Li, Guo-xue; Zhang, Hong-yu; Luo, Yi-ming

    2013-09-01

    It is necessary to achieve the optimization for MSW logistics based on the new Xicheng (combining the former Xicheng and the former Xuanwu districts) and the new Dongcheng (combining the former Dongcheng and the former Chongwen districts) districts of Beijing. Based on the analysis of current MSW logistics system, transfer station's processing capacity and the terminal treatment facilities' conditions of the four former districts and other districts, a MSW logistics system was built by GIS methods considering transregional treatment. This article analyzes the MSW material balance of current and new logistics systems. Results show that the optimization scheme could reduce the MSW collection distance of the new Xicheng and the new Dongcheng by 9.3 x 10(5) km x a(-1), reduced by 10% compared with current logistics. Under the new logistics solution, considering transregional treatment, can reduce landfill treatment of untreated MSW about 28.3%. If the construction of three incineration plants finished based on the new logistics, the system's optimal ratio of incineration: biochemical treatment: landfill can reach 3.8 : 4.5 : 1.7 compared with 1 : 4.8 : 4.2, which is the ratio of current MSW logistics. The ratio of the amount of incineration: biochemical treatment: landfill approximately reach 4 : 3 : 3 which is the target for 2015. The research results are benefit in increasing MSW utilization and reduction rate of the new Dongcheng and Xicheng districts and nearby districts.

  11. Effect of a Nutrition Supplement and Physical Activity Program on Pneumonia and Walking Capacity in Chilean Older People: A Factorial Cluster Randomized Trial

    Science.gov (United States)

    Dangour, Alan D.; Albala, Cecilia; Allen, Elizabeth; Grundy, Emily; Walker, Damian G.; Aedo, Cristian; Sanchez, Hugo; Fletcher, Olivia; Elbourne, Diana; Uauy, Ricardo

    2011-01-01

    Background Ageing is associated with increased risk of poor health and functional decline. Uncertainties about the health-related benefits of nutrition and physical activity for older people have precluded their widespread implementation. We investigated the effectiveness and cost-effectiveness of a national nutritional supplementation program and/or a physical activity intervention among older people in Chile. Methods and Findings We conducted a cluster randomized factorial trial among low to middle socioeconomic status adults aged 65–67.9 years living in Santiago, Chile. We randomized 28 clusters (health centers) into the study and recruited 2,799 individuals in 2005 (∼100 per cluster). The interventions were a daily micronutrient-rich nutritional supplement, or two 1-hour physical activity classes per week, or both interventions, or neither, for 24 months. The primary outcomes, assessed blind to allocation, were incidence of pneumonia over 24 months, and physical function assessed by walking capacity 24 months after enrolment. Adherence was good for the nutritional supplement (∼75%), and moderate for the physical activity intervention (∼43%). Over 24 months the incidence rate of pneumonia did not differ between intervention and control clusters (32.5 versus 32.6 per 1,000 person years respectively; risk ratio = 1.00; 95% confidence interval 0.61–1.63; p = 0.99). In intention-to-treat analysis, after 24 months there was a significant difference in walking capacity between the intervention and control clusters (mean difference 33.8 meters; 95% confidence interval 13.9–53.8; p = 0.001). The overall cost of the physical activity intervention over 24 months was US$164/participant; equivalent to US$4.84/extra meter walked. The number of falls and fractures was balanced across physical activity intervention arms and no serious adverse events were reported for either intervention. Conclusions Chile's nutritional supplementation program for older

  12. Effect of a nutrition supplement and physical activity program on pneumonia and walking capacity in Chilean older people: a factorial cluster randomized trial.

    Directory of Open Access Journals (Sweden)

    Alan D Dangour

    2011-04-01

    Full Text Available Ageing is associated with increased risk of poor health and functional decline. Uncertainties about the health-related benefits of nutrition and physical activity for older people have precluded their widespread implementation. We investigated the effectiveness and cost-effectiveness of a national nutritional supplementation program and/or a physical activity intervention among older people in Chile.We conducted a cluster randomized factorial trial among low to middle socioeconomic status adults aged 65-67.9 years living in Santiago, Chile. We randomized 28 clusters (health centers into the study and recruited 2,799 individuals in 2005 (~100 per cluster. The interventions were a daily micronutrient-rich nutritional supplement, or two 1-hour physical activity classes per week, or both interventions, or neither, for 24 months. The primary outcomes, assessed blind to allocation, were incidence of pneumonia over 24 months, and physical function assessed by walking capacity 24 months after enrollment. Adherence was good for the nutritional supplement (~75%, and moderate for the physical activity intervention (~43%. Over 24 months the incidence rate of pneumonia did not differ between intervention and control clusters (32.5 versus 32.6 per 1,000 person years respectively; risk ratio = 1.00; 95% confidence interval 0.61-1.63; p = 0.99. In intention-to-treat analysis, after 24 months there was a significant difference in walking capacity between the intervention and control clusters (mean difference 33.8 meters; 95% confidence interval 13.9-53.8; p = 0.001. The overall cost of the physical activity intervention over 24 months was US$164/participant; equivalent to US$4.84/extra meter walked. The number of falls and fractures was balanced across physical activity intervention arms and no serious adverse events were reported for either intervention.Chile's nutritional supplementation program for older people is not effective in reducing the

  13. Cluster-cluster clustering

    International Nuclear Information System (INIS)

    Barnes, J.; Dekel, A.; Efstathiou, G.; Frenk, C.S.; Yale Univ., New Haven, CT; California Univ., Santa Barbara; Cambridge Univ., England; Sussex Univ., Brighton, England)

    1985-01-01

    The cluster correlation function xi sub c(r) is compared with the particle correlation function, xi(r) in cosmological N-body simulations with a wide range of initial conditions. The experiments include scale-free initial conditions, pancake models with a coherence length in the initial density field, and hybrid models. Three N-body techniques and two cluster-finding algorithms are used. In scale-free models with white noise initial conditions, xi sub c and xi are essentially identical. In scale-free models with more power on large scales, it is found that the amplitude of xi sub c increases with cluster richness; in this case the clusters give a biased estimate of the particle correlations. In the pancake and hybrid models (with n = 0 or 1), xi sub c is steeper than xi, but the cluster correlation length exceeds that of the points by less than a factor of 2, independent of cluster richness. Thus the high amplitude of xi sub c found in studies of rich clusters of galaxies is inconsistent with white noise and pancake models and may indicate a primordial fluctuation spectrum with substantial power on large scales. 30 references

  14. Building the capacity of family day care educators to promote children's social and emotional wellbeing: an exploratory cluster randomised controlled trial

    Directory of Open Access Journals (Sweden)

    Sims Margaret

    2011-11-01

    Full Text Available Abstract Background Childhood mental health problems are highly prevalent, experienced by one in five children living in socioeconomically disadvantaged families. Although childcare settings, including family day care are ideal to promote children's social and emotional wellbeing at a population level in a sustainable way, family day care educators receive limited training in promoting children's mental health. This study is an exploratory wait-list control cluster randomised controlled trial to test the appropriateness, acceptability, cost, and effectiveness of "Thrive," an intervention program to build the capacity of family day care educators to promote children's social and emotional wellbeing. Thrive aims to increase educators' knowledge, confidence and skills in promoting children's social and emotional wellbeing. Methods/Design This study involves one family day care organisation based in a low socioeconomic area of Melbourne. All family day care educators (term used for registered carers who provide care for children for financial reimbursement in the carers own home are eligible to participate in the study. The clusters for randomisation will be the fieldworkers (n = 5 who each supervise 10-15 educators. The intervention group (field workers and educators will participate in a variety of intervention activities over 12 months, including workshops; activity exchanges with other educators; and focused discussion about children's social and emotional wellbeing during field worker visits. The control group will continue with their normal work practice. The intervention will be delivered to the intervention group and then to the control group after a time delay of 15 months post intervention commencement. A baseline survey will be conducted with all consenting educators and field workers (n = ~70 assessing outcomes at the cluster and individual level. The survey will also be administered at one month, six months and 12 months post

  15. Inulin-type fructan degradation capacity of Clostridium cluster IV and XIVa butyrate-producing colon bacteria and their associated metabolic outcomes.

    Science.gov (United States)

    Moens, F; De Vuyst, L

    2017-05-30

    Four selected butyrate-producing colon bacterial strains belonging to Clostridium cluster IV (Butyricicoccus pullicaecorum DSM 23266 T and Faecalibacterium prausnitzii DSM 17677 T ) and XIVa (Eubacterium hallii DSM 17630 and Eubacterium rectale CIP 105953 T ) were studied as to their capacity to degrade inulin-type fructans and concomitant metabolite production. Cultivation of these strains was performed in bottles and fermentors containing a modified medium for colon bacteria, including acetate, supplemented with either fructose, oligofructose, or inulin as the sole energy source. Inulin-type fructan degradation was not a general characteristic among these strains. B. pullicaecorum DSM 23266 T and E. hallii DSM 17630 could only ferment fructose and did not degrade oligofructose or inulin. E. rectale CIP 105953 T and F. prausnitzii DSM 17677 T fermented fructose and could degrade both oligofructose and inulin. All chain length fractions of oligofructose were degraded simultaneously (both strains) and both long and short chain length fractions of inulin were degraded either simultaneously (E. rectale CIP 105953 T ) or consecutively (F. prausnitzii DSM 17677 T ), indicating an extracellular polymer degradation mechanism. B. pullicaecorum DSM 23266 T and E. hallii DSM 17630 produced high concentrations of butyrate, CO 2 , and H 2 from fructose. E. rectale CIP 105953 T produced lactate, butyrate, CO 2 , and H 2 , from fructose, oligofructose, and inulin, whereas F. prausnitzii DSM 17677 T produced butyrate, formate, CO 2 , and traces of lactate from fructose, oligofructose, and inulin. Based on carbon recovery and theoretical metabolite production calculations, an adapted stoichiometrically balanced metabolic pathway for butyrate, formate, lactate, CO 2 , and H 2 production by members of both Clostridium cluster IV and XIVa butyrate-producing bacteria was constructed.

  16. Teor e capacidade máxima de adsorção de arsênio em Latossolos brasileiros Content and maximum capacity of arsenic adsorption in Brazilian Oxisols

    Directory of Open Access Journals (Sweden)

    Mari Lucia Campos

    2007-12-01

    Full Text Available A alta toxicidade do As aos animais e humanos e a possibilidade de existência de grande número de áreas contaminadas tornam imprescindível o conhecimento do teor semitotal em solos ditos não-contaminados e dos processos de adsorção do As em solos de carga variável. O objetivo deste trabalho foi determinar o teor e a capacidade máxima de adsorção de As (CMADS AS em Latossolos. O teor total foi determinado pelo método USEPA 3051A, e a CMADS As, com auxílio de isotermas de Langmuir com base nos valores de adsorção obtidos em dose de As (0, 90, 190, 380, 760 e 1.150 µmol L-1 (relação solo:solução final = 1:100, a pH 5,5 e força iônica de 15 mmol L-1. Os 17 Latossolos apresentaram teor médio total de As de 5,92 mg kg-1 e CMADS As média de 2.013 mg kg-1. O teor de argila e os óxidos de Fe e Al apresentaram influência positiva na CMADS As.In view of the toxicity of As for man and animals and the possibility of existence of a great number of contaminated areas it is imperative to know the total As content in soils considered uncontaminated and about As sorption processes in soils of variable charge. The objective of this work was to determine the total content and maximum capacity of As adsorption (CMADS As in Oxisols. The total content was determined by the USEPA 3051A method. The cmADS As was determined by the Langmuir Isotherms using six solution concentrations (0, 0.09, 0.19, 0.38, 0.76, 1.15 mmol L-1 (1:100 soil: solution ratio, pH values 5.5 and ionic strength 15 mmol L-1. In the 17 Oxisols the average total As content was 5.92 mg kg-1 and mean cmADS As was 2.013 mg kg-1. Clay, and Fe and Al oxides content influenced cmADSs positively.

  17. 'MATRI-SUMAN' a capacity building and text messaging intervention to enhance maternal and child health service utilization among pregnant women from rural Nepal: study protocol for a cluster randomised controlled trial.

    Science.gov (United States)

    Singh, Jitendra Kumar; Kadel, Rajendra; Acharya, Dilaram; Lombard, Daniel; Khanal, Saval; Singh, Shri Prakash

    2018-06-14

    Capacity development of health volunteers and text messaging to pregnant women through mobile phones have shown improved maternal and child health (MCH) outcomes and is associated with increased utilisation of MCH services. However, such interventions are uncommon in Nepal. We aim to carry out an intervention with the hypothesis that capacity building and text messaging intervention will increase the MCH service utilisation. MATRI-SUMAN is a 12-month cluster randomized controlled trial (RCT). The trial involves pregnant women from 52 clusters of six village development committees (VDCs) covering 66,000 populations of Dhanusha district of Nepal. In the intervention clusters, Female Community Health Volunteers (FCHVs) will receive capacity development skills through reinforcement training, supervision and monitoring skills for the promotion of health seeking behaviour among pregnant women and study participants will receive periodic promotional text messaging service about MCH components through mobile phones. A sample of 354with equal numbers in each study arm is estimated using power calculation formula. The primary outcomes of this study are the rate of utilization of skilled birth attendants and consumption of a specified diversified meal. The secondary outcomes are: four antenatal (ANC) visits, weight gain of women during pregnancy, delivery of a baby at the health facility, postnatal care (PNC) visits, positive changes in child feeding practices among mothers, performance of FCHVs in MCH service utilization. The intervention is designed to enhance the capacity of health volunteers for the promotion of health seeking behaviour among pregnant women and text messaging through a mobile phone to expecting mothers to increase MCH service utilization. The trial if proven effective will have policy implications in poor resource settings. ISRCTN60684155, ( https://doi.org/10.1186/ISRCTN60684155 ). The trial was registered retrospectively.

  18. Fator capacidade de fósforo em solos de pernambuco mineralogicamente diferentes e influência do pH na capacidade máxima de adsorção Phosphate capacity factor in mineralogically different soils in Pernambuco and the influence of pH on the maximum capacity of adsorption

    Directory of Open Access Journals (Sweden)

    Fábio Broggi

    2011-02-01

    Full Text Available O Fator Capacidade de Fósforo (FCP é definido pela razão de equilíbrio entre o fator quantidade de P (Q e o fator intensidade (I e representa uma medida da capacidade do solo em manter um determinado nível de P em solução. As características e o teor dos constituintes minerais da fração argila são responsáveis por uma maior ou menor FCP, interferindo nas relações solo-planta. Por outro lado, o pH do solo tem, em alguns casos, mostrado-se com efeito na adsorção e, em outros, com pequena e não consistente alteração na Capacidade Máxima de Adsorção de P (CMAP. Objetivou-se, neste trabalho, determinar o FCP de solos mineralogicamente diferentes em Pernambuco; correlacionar características físicas e químicas dos solos com o FCP; e avaliar o efeito do pH na CMAP. Amostras subsuperficiais de quatro solos, mineralogicamente diferentes, foram caracterizadas química e fisicamente e determinado o FCP. Essas amostras foram corrigidas com CaCO3 e MgCO3 na proporção 4:1 e incubadas por 30 dias, com exceção do Vertissolo. Determinou-se a CMAP antes e após a correção dos solos. O experimento consistiu de um fatorial 4 x 2 (quatro solos com e sem correção, distribuídos em blocos ao acaso, com três repetições. As características dos solos que melhor refletiram o FCP foram o P remanescente (P-rem e a CMAP. Independentemente dos constituintes mineralógicos da fração argila, solos com elevados teores de alumínio apresentaram aumento da CMAP com a correção. A energia de adsorção (EA nos solos corrigidos foi, em média, significativamente menor, independentemente do solo.Phosphate Maximum Capacity (FCP is defined by the ratio of equilibrium between the amount of factor P (Q and factor intensity (I and represents a measure of the soil ability to maintain a certain level of P in solution. The characteristics and content of the constituents of clay minerals are responsible for a greater or lesser FCP, interfering in soil

  19. Clustering of near clusters versus cluster compactness

    International Nuclear Information System (INIS)

    Yu Gao; Yipeng Jing

    1989-01-01

    The clustering properties of near Zwicky clusters are studied by using the two-point angular correlation function. The angular correlation functions for compact and medium compact clusters, for open clusters, and for all near Zwicky clusters are estimated. The results show much stronger clustering for compact and medium compact clusters than for open clusters, and that open clusters have nearly the same clustering strength as galaxies. A detailed study of the compactness-dependence of correlation function strength is worth investigating. (author)

  20. Approximate maximum parsimony and ancestral maximum likelihood.

    Science.gov (United States)

    Alon, Noga; Chor, Benny; Pardi, Fabio; Rapoport, Anat

    2010-01-01

    We explore the maximum parsimony (MP) and ancestral maximum likelihood (AML) criteria in phylogenetic tree reconstruction. Both problems are NP-hard, so we seek approximate solutions. We formulate the two problems as Steiner tree problems under appropriate distances. The gist of our approach is the succinct characterization of Steiner trees for a small number of leaves for the two distances. This enables the use of known Steiner tree approximation algorithms. The approach leads to a 16/9 approximation ratio for AML and asymptotically to a 1.55 approximation ratio for MP.

  1. 基于5A景区最大承载量和游客接待量的旅游供需关系模型研究%Study of Mathematical Model Between Tourism Supply and Demand Based on Relationship of the 5A Level Scenic Spots' Maximum Carrying Capacity and the Amount of the Tourists

    Institute of Scientific and Technical Information of China (English)

    刘静; 刘耀龙; 段锦

    2017-01-01

    旅游承载量是反映旅游供给能力的重要指标,游客接待量是衡量旅游需求状况的常用指标.在分析我国5A级景区最大承载量空间分异特征的基础上,对2014年各省市5A级景区日最大承载量和年游客接待量进行相关性和回归分析,构建基于景区承载量和游客接待量的旅游供需关系模型.结果表明:(1)5A级景区最大承载量和游客接待量之间存在显著的正相关关系;(2)旅游供需关系模型可由复合函数定量表征;(3)模型能够预测旅游供需变化,对旅游业发展具有一定的指示意义.%Tourism carrying capacity is an important index to reflect the tourism supply capacity, and the amount of the tourists is a common index to measure tourism demand. On the basis of the analysis of the spatial distribution characteristics of 5A level scenic spots' maximum carrying capacity, the paper carries out correlation and regression analysis about the 5A level scenic' maximum carrying capacity and the amount of the tourists. Sample data is calculated or statistical results of 31 provinces in 2014. Tourism supply and demand model is constructed based on the relationship between the 5A level scenic spots' maximum carrying capacity and the amount of the tourists. Results indicate: (1)There is a significant positive correlation between the 5A level scenic spot'maximum carrying capacity and the amount of tourists; (2)The model of tourism supply and demand relations can be quantitatively characterized by complex functions;(3)The model can predict the change of tourism supply and demand, which have referential meaning to the development of tourism.

  2. Maximum permissible dose

    International Nuclear Information System (INIS)

    Anon.

    1979-01-01

    This chapter presents a historic overview of the establishment of radiation guidelines by various national and international agencies. The use of maximum permissible dose and maximum permissible body burden limits to derive working standards is discussed

  3. Spanning Tree Based Attribute Clustering

    DEFF Research Database (Denmark)

    Zeng, Yifeng; Jorge, Cordero Hernandez

    2009-01-01

    Attribute clustering has been previously employed to detect statistical dependence between subsets of variables. We propose a novel attribute clustering algorithm motivated by research of complex networks, called the Star Discovery algorithm. The algorithm partitions and indirectly discards...... inconsistent edges from a maximum spanning tree by starting appropriate initial modes, therefore generating stable clusters. It discovers sound clusters through simple graph operations and achieves significant computational savings. We compare the Star Discovery algorithm against earlier attribute clustering...

  4. Progressive Exponential Clustering-Based Steganography

    Directory of Open Access Journals (Sweden)

    Li Yue

    2010-01-01

    Full Text Available Cluster indexing-based steganography is an important branch of data-hiding techniques. Such schemes normally achieve good balance between high embedding capacity and low embedding distortion. However, most cluster indexing-based steganographic schemes utilise less efficient clustering algorithms for embedding data, which causes redundancy and leaves room for increasing the embedding capacity further. In this paper, a new clustering algorithm, called progressive exponential clustering (PEC, is applied to increase the embedding capacity by avoiding redundancy. Meanwhile, a cluster expansion algorithm is also developed in order to further increase the capacity without sacrificing imperceptibility.

  5. Maximum Acceleration Recording Circuit

    Science.gov (United States)

    Bozeman, Richard J., Jr.

    1995-01-01

    Coarsely digitized maximum levels recorded in blown fuses. Circuit feeds power to accelerometer and makes nonvolatile record of maximum level to which output of accelerometer rises during measurement interval. In comparison with inertia-type single-preset-trip-point mechanical maximum-acceleration-recording devices, circuit weighs less, occupies less space, and records accelerations within narrower bands of uncertainty. In comparison with prior electronic data-acquisition systems designed for same purpose, circuit simpler, less bulky, consumes less power, costs and analysis of data recorded in magnetic or electronic memory devices. Circuit used, for example, to record accelerations to which commodities subjected during transportation on trucks.

  6. Maximum Quantum Entropy Method

    OpenAIRE

    Sim, Jae-Hoon; Han, Myung Joon

    2018-01-01

    Maximum entropy method for analytic continuation is extended by introducing quantum relative entropy. This new method is formulated in terms of matrix-valued functions and therefore invariant under arbitrary unitary transformation of input matrix. As a result, the continuation of off-diagonal elements becomes straightforward. Without introducing any further ambiguity, the Bayesian probabilistic interpretation is maintained just as in the conventional maximum entropy method. The applications o...

  7. Maximum power demand cost

    International Nuclear Information System (INIS)

    Biondi, L.

    1998-01-01

    The charging for a service is a supplier's remuneration for the expenses incurred in providing it. There are currently two charges for electricity: consumption and maximum demand. While no problem arises about the former, the issue is more complicated for the latter and the analysis in this article tends to show that the annual charge for maximum demand arbitrarily discriminates among consumer groups, to the disadvantage of some [it

  8. Genomes: At the edge of chaos with maximum information capacity

    Science.gov (United States)

    Kong, Sing-Guan; Chen, Hong-Da; Torda, Andrew; Lee, H. C.

    2016-12-01

    We propose an order index, ϕ, which quantifies the notion of “life at the edge of chaos” when applied to genome sequences. It maps genomes to a number from 0 (random and of infinite length) to 1 (fully ordered) and applies regardless of sequence length and base composition. The 786 complete genomic sequences in GenBank were found to have ϕ values in a very narrow range, 0.037 ± 0.027. We show this implies that genomes are halfway towards being completely random, namely, at the edge of chaos. We argue that this narrow range represents the neighborhood of a fixed-point in the space of sequences, and genomes are driven there by the dynamics of a robust, predominantly neutral evolution process.

  9. A modeling using the maximum growth capacity of Hantzschia ...

    African Journals Online (AJOL)

    Homa lagoon found in the Izmir bay, Aegean Sea, is important because it is the last lagoon in which fishing activities are carried out. Hantzschia amphioxys species, on which this study was carried on, are benthic diatoms and they are isolated from Homa Lagoon. The Monod equation expressing the nutrient dependency of ...

  10. A modeling using the maximum growth capacity of Hantzschia ...

    African Journals Online (AJOL)

    CREA

    2014-03-05

    Mar 5, 2014 ... primary production (Perrissinotto et al., 2000; Montani et al., 2003). It is reported that diatoms have an important role in global biogeochemical cycle of nitrogen and carbon (Serodio, 2003). Furthermore, coastal waters are characterized by abundance of diatoms (Nelson et al.,. 1995). In coastal waters, wave, ...

  11. The logistic model-generated carrying capacities, maximum ...

    African Journals Online (AJOL)

    This paper deals with the derivation of logistic models for cattle, sheep and goats in a commercial ranching system in Machakos District, Kenya, a savannah ecosystem with average annual rainfall of 589.3 ± 159.3mm and an area of 10 117ha. It involves modelling livestock population dynamics as discrete-time logistic ...

  12. Maximum likely scale estimation

    DEFF Research Database (Denmark)

    Loog, Marco; Pedersen, Kim Steenstrup; Markussen, Bo

    2005-01-01

    A maximum likelihood local scale estimation principle is presented. An actual implementation of the estimation principle uses second order moments of multiple measurements at a fixed location in the image. These measurements consist of Gaussian derivatives possibly taken at several scales and/or ...

  13. Robust Maximum Association Estimators

    NARCIS (Netherlands)

    A. Alfons (Andreas); C. Croux (Christophe); P. Filzmoser (Peter)

    2017-01-01

    textabstractThe maximum association between two multivariate variables X and Y is defined as the maximal value that a bivariate association measure between one-dimensional projections αX and αY can attain. Taking the Pearson correlation as projection index results in the first canonical correlation

  14. Maximum power point tracking

    International Nuclear Information System (INIS)

    Enslin, J.H.R.

    1990-01-01

    A well engineered renewable remote energy system, utilizing the principal of Maximum Power Point Tracking can be m ore cost effective, has a higher reliability and can improve the quality of life in remote areas. This paper reports that a high-efficient power electronic converter, for converting the output voltage of a solar panel, or wind generator, to the required DC battery bus voltage has been realized. The converter is controlled to track the maximum power point of the input source under varying input and output parameters. Maximum power point tracking for relative small systems is achieved by maximization of the output current in a battery charging regulator, using an optimized hill-climbing, inexpensive microprocessor based algorithm. Through practical field measurements it is shown that a minimum input source saving of 15% on 3-5 kWh/day systems can easily be achieved. A total cost saving of at least 10-15% on the capital cost of these systems are achievable for relative small rating Remote Area Power Supply systems. The advantages at larger temperature variations and larger power rated systems are much higher. Other advantages include optimal sizing and system monitor and control

  15. Maximum entropy methods

    International Nuclear Information System (INIS)

    Ponman, T.J.

    1984-01-01

    For some years now two different expressions have been in use for maximum entropy image restoration and there has been some controversy over which one is appropriate for a given problem. Here two further entropies are presented and it is argued that there is no single correct algorithm. The properties of the four different methods are compared using simple 1D simulations with a view to showing how they can be used together to gain as much information as possible about the original object. (orig.)

  16. The last glacial maximum

    Science.gov (United States)

    Clark, P.U.; Dyke, A.S.; Shakun, J.D.; Carlson, A.E.; Clark, J.; Wohlfarth, B.; Mitrovica, J.X.; Hostetler, S.W.; McCabe, A.M.

    2009-01-01

    We used 5704 14C, 10Be, and 3He ages that span the interval from 10,000 to 50,000 years ago (10 to 50 ka) to constrain the timing of the Last Glacial Maximum (LGM) in terms of global ice-sheet and mountain-glacier extent. Growth of the ice sheets to their maximum positions occurred between 33.0 and 26.5 ka in response to climate forcing from decreases in northern summer insolation, tropical Pacific sea surface temperatures, and atmospheric CO2. Nearly all ice sheets were at their LGM positions from 26.5 ka to 19 to 20 ka, corresponding to minima in these forcings. The onset of Northern Hemisphere deglaciation 19 to 20 ka was induced by an increase in northern summer insolation, providing the source for an abrupt rise in sea level. The onset of deglaciation of the West Antarctic Ice Sheet occurred between 14 and 15 ka, consistent with evidence that this was the primary source for an abrupt rise in sea level ???14.5 ka.

  17. Cluster headache

    Science.gov (United States)

    Histamine headache; Headache - histamine; Migrainous neuralgia; Headache - cluster; Horton's headache; Vascular headache - cluster ... Doctors do not know exactly what causes cluster headaches. They ... (chemical in the body released during an allergic response) or ...

  18. Maximum Entropy Fundamentals

    Directory of Open Access Journals (Sweden)

    F. Topsøe

    2001-09-01

    Full Text Available Abstract: In its modern formulation, the Maximum Entropy Principle was promoted by E.T. Jaynes, starting in the mid-fifties. The principle dictates that one should look for a distribution, consistent with available information, which maximizes the entropy. However, this principle focuses only on distributions and it appears advantageous to bring information theoretical thinking more prominently into play by also focusing on the "observer" and on coding. This view was brought forward by the second named author in the late seventies and is the view we will follow-up on here. It leads to the consideration of a certain game, the Code Length Game and, via standard game theoretical thinking, to a principle of Game Theoretical Equilibrium. This principle is more basic than the Maximum Entropy Principle in the sense that the search for one type of optimal strategies in the Code Length Game translates directly into the search for distributions with maximum entropy. In the present paper we offer a self-contained and comprehensive treatment of fundamentals of both principles mentioned, based on a study of the Code Length Game. Though new concepts and results are presented, the reading should be instructional and accessible to a rather wide audience, at least if certain mathematical details are left aside at a rst reading. The most frequently studied instance of entropy maximization pertains to the Mean Energy Model which involves a moment constraint related to a given function, here taken to represent "energy". This type of application is very well known from the literature with hundreds of applications pertaining to several different elds and will also here serve as important illustration of the theory. But our approach reaches further, especially regarding the study of continuity properties of the entropy function, and this leads to new results which allow a discussion of models with so-called entropy loss. These results have tempted us to speculate over

  19. Probable maximum flood control

    International Nuclear Information System (INIS)

    DeGabriele, C.E.; Wu, C.L.

    1991-11-01

    This study proposes preliminary design concepts to protect the waste-handling facilities and all shaft and ramp entries to the underground from the probable maximum flood (PMF) in the current design configuration for the proposed Nevada Nuclear Waste Storage Investigation (NNWSI) repository protection provisions were furnished by the United States Bureau of Reclamation (USSR) or developed from USSR data. Proposed flood protection provisions include site grading, drainage channels, and diversion dikes. Figures are provided to show these proposed flood protection provisions at each area investigated. These areas are the central surface facilities (including the waste-handling building and waste treatment building), tuff ramp portal, waste ramp portal, men-and-materials shaft, emplacement exhaust shaft, and exploratory shafts facility

  20. Introduction to maximum entropy

    International Nuclear Information System (INIS)

    Sivia, D.S.

    1988-01-01

    The maximum entropy (MaxEnt) principle has been successfully used in image reconstruction in a wide variety of fields. We review the need for such methods in data analysis and show, by use of a very simple example, why MaxEnt is to be preferred over other regularizing functions. This leads to a more general interpretation of the MaxEnt method, and its use is illustrated with several different examples. Practical difficulties with non-linear problems still remain, this being highlighted by the notorious phase problem in crystallography. We conclude with an example from neutron scattering, using data from a filter difference spectrometer to contrast MaxEnt with a conventional deconvolution. 12 refs., 8 figs., 1 tab

  1. Solar maximum observatory

    International Nuclear Information System (INIS)

    Rust, D.M.

    1984-01-01

    The successful retrieval and repair of the Solar Maximum Mission (SMM) satellite by Shuttle astronauts in April 1984 permitted continuance of solar flare observations that began in 1980. The SMM carries a soft X ray polychromator, gamma ray, UV and hard X ray imaging spectrometers, a coronagraph/polarimeter and particle counters. The data gathered thus far indicated that electrical potentials of 25 MeV develop in flares within 2 sec of onset. X ray data show that flares are composed of compressed magnetic loops that have come too close together. Other data have been taken on mass ejection, impacts of electron beams and conduction fronts with the chromosphere and changes in the solar radiant flux due to sunspots. 13 references

  2. Introduction to maximum entropy

    International Nuclear Information System (INIS)

    Sivia, D.S.

    1989-01-01

    The maximum entropy (MaxEnt) principle has been successfully used in image reconstruction in a wide variety of fields. The author reviews the need for such methods in data analysis and shows, by use of a very simple example, why MaxEnt is to be preferred over other regularizing functions. This leads to a more general interpretation of the MaxEnt method, and its use is illustrated with several different examples. Practical difficulties with non-linear problems still remain, this being highlighted by the notorious phase problem in crystallography. He concludes with an example from neutron scattering, using data from a filter difference spectrometer to contrast MaxEnt with a conventional deconvolution. 12 refs., 8 figs., 1 tab

  3. Functional Maximum Autocorrelation Factors

    DEFF Research Database (Denmark)

    Larsen, Rasmus; Nielsen, Allan Aasbjerg

    2005-01-01

    MAF outperforms the functional PCA in concentrating the interesting' spectra/shape variation in one end of the eigenvalue spectrum and allows for easier interpretation of effects. Conclusions. Functional MAF analysis is a useful methods for extracting low dimensional models of temporally or spatially......Purpose. We aim at data where samples of an underlying function are observed in a spatial or temporal layout. Examples of underlying functions are reflectance spectra and biological shapes. We apply functional models based on smoothing splines and generalize the functional PCA in......\\verb+~+\\$\\backslash\\$cite{ramsay97} to functional maximum autocorrelation factors (MAF)\\verb+~+\\$\\backslash\\$cite{switzer85,larsen2001d}. We apply the method to biological shapes as well as reflectance spectra. {\\$\\backslash\\$bf Methods}. MAF seeks linear combination of the original variables that maximize autocorrelation between...

  4. Regularized maximum correntropy machine

    KAUST Repository

    Wang, Jim Jing-Yan; Wang, Yunji; Jing, Bing-Yi; Gao, Xin

    2015-01-01

    In this paper we investigate the usage of regularized correntropy framework for learning of classifiers from noisy labels. The class label predictors learned by minimizing transitional loss functions are sensitive to the noisy and outlying labels of training samples, because the transitional loss functions are equally applied to all the samples. To solve this problem, we propose to learn the class label predictors by maximizing the correntropy between the predicted labels and the true labels of the training samples, under the regularized Maximum Correntropy Criteria (MCC) framework. Moreover, we regularize the predictor parameter to control the complexity of the predictor. The learning problem is formulated by an objective function considering the parameter regularization and MCC simultaneously. By optimizing the objective function alternately, we develop a novel predictor learning algorithm. The experiments on two challenging pattern classification tasks show that it significantly outperforms the machines with transitional loss functions.

  5. Regularized maximum correntropy machine

    KAUST Repository

    Wang, Jim Jing-Yan

    2015-02-12

    In this paper we investigate the usage of regularized correntropy framework for learning of classifiers from noisy labels. The class label predictors learned by minimizing transitional loss functions are sensitive to the noisy and outlying labels of training samples, because the transitional loss functions are equally applied to all the samples. To solve this problem, we propose to learn the class label predictors by maximizing the correntropy between the predicted labels and the true labels of the training samples, under the regularized Maximum Correntropy Criteria (MCC) framework. Moreover, we regularize the predictor parameter to control the complexity of the predictor. The learning problem is formulated by an objective function considering the parameter regularization and MCC simultaneously. By optimizing the objective function alternately, we develop a novel predictor learning algorithm. The experiments on two challenging pattern classification tasks show that it significantly outperforms the machines with transitional loss functions.

  6. Economic feasibility of district heating delivery from industrial excess heat: A case study of a Swedish petrochemical cluster

    International Nuclear Information System (INIS)

    Morandin, Matteo; Hackl, Roman; Harvey, Simon

    2014-01-01

    The present work discusses the potential and the economic feasibility of DH (district heating) delivery using industrial excess heat from a petrochemical cluster at the Swedish West Coast. Pinch Analysis was used for estimating the DH capacity targets and for estimating the cost of heat exchanger installation. A discounted cash flow rate of return of 10% was used as a criterion for identifying the minimum yearly DH delivery that should be guaranteed for a given DH capacity at different DH sales prices. The study was conducted for the current scenario in which no heat recovery is achieved between the cluster plants and for a possible future scenario in which 50% of the fuel currently used for heating purposes is saved by increasing the heat recovery at the site. The competition between excess heat export and local energy efficiency measures is also discussed in terms of CO 2 emission consequences. The maximum capacity of DH delivery amounts today to around 235 MW, which reduces to 110 MW in the future scenario of increased site heat recovery. The results of our analysis show that feasible conditions exist that make DH delivery profitable in the entire capacity range. - Highlights: • Pinch Analysis targeting approach and short-cut cost accounting procedure. • Economic analysis for different DH delivery capacities up to maximum targets. • Sensitivity analysis of heat sales prices. • Parallel plants and cluster wide heat collection systems considered. • Competition between cluster internal heat recovery and excess heat export is discussed

  7. Protocol for a cluster-randomised controlled trial evaluating the impact of a preschool-based capacity building intervention on intimate partner violence and substance misuse in Sri Lanka.

    Science.gov (United States)

    Lokuge, Kamalini; Wallace, Polly; Subasinghe, Kalini; Thurber, Katherine; De Silva, Tissa; Clarke, Naomi; Waas, Dulshika; Liyanage, Nisansala; Attygalle, Udena; Carron-Arthur, Bradley; Rodrigo, Kalyana; Banks, Emily; D'Este, Cate; Rajapakse, Thilini

    2018-05-02

    Past research has identified links between intimate partner violence (IPV) and alcohol misuse and poverty in Sri Lanka. Services that address substance misuse are amongst the few interventions shown to reduce IPV in settings similar to Sri Lanka. This paper describes the protocol for a study examining the impact of a preschool-based capacity building intervention on the prevalence of IPV and substance misuse in parents with children attending preschools, including uptake of available government services. The study is a cluster randomised controlled trial. Government-managed preschools (n = 34) in Galle and Colombo municipalities  will be randomly assigned to an intervention (n = 17) or control group (n = 17). Parents with children attending these preschools will be recruited to participate. The study intervention will build the capacity of selected community volunteers (parents) and preschool teachers in the provision of information and support to families affected by IPV and substance misuse. This intervention is directed at improving uptake, access and coordination of existing services. Data will be collected from all parents, and teachers in the intervention group, pre-intervention and 10 months post-intervention. The primary outcome for this study is experience of IPV amongst mothers of preschool-attending children. Secondary outcomes are substance misuse amongst fathers, measured via the locally adapted Alcohol Use Disorders Identification Test and Drug Abuse Screening Test; and awareness and uptake of services for these issues measured through locally-relevant tools. Demographic information and satisfaction with the intervention will also be assessed. By intervening through preschools we aim to support high-risk families early enough to arrest the cycle of violence that results in children themselves becoming victims and perpetrators of such violence. The innovative project design will reach the most vulnerable sections of the community and will

  8. Solar maximum mission

    International Nuclear Information System (INIS)

    Ryan, J.

    1981-01-01

    By understanding the sun, astrophysicists hope to expand this knowledge to understanding other stars. To study the sun, NASA launched a satellite on February 14, 1980. The project is named the Solar Maximum Mission (SMM). The satellite conducted detailed observations of the sun in collaboration with other satellites and ground-based optical and radio observations until its failure 10 months into the mission. The main objective of the SMM was to investigate one aspect of solar activity: solar flares. A brief description of the flare mechanism is given. The SMM satellite was valuable in providing information on where and how a solar flare occurs. A sequence of photographs of a solar flare taken from SMM satellite shows how a solar flare develops in a particular layer of the solar atmosphere. Two flares especially suitable for detailed observations by a joint effort occurred on April 30 and May 21 of 1980. These flares and observations of the flares are discussed. Also discussed are significant discoveries made by individual experiments

  9. Station Capacity

    DEFF Research Database (Denmark)

    Landex, Alex

    2011-01-01

    the probability of conflicts and the minimum headway times into account. The last method analyzes how optimal platform tracks are used by examining the arrival and departure pattern of the trains. The developed methods can either be used separately to analyze specific characteristics of the capacity of a station......Stations are often limiting the capacity of railway networks. This is due to extra need of tracks when trains stand still, trains turning around, and conflicting train routes. Although stations are often the capacity bottlenecks, most capacity analysis methods focus on open line capacity. Therefore...... for platform tracks and the probability that arriving trains will not get a platform track immediately at arrival. The third method is a scalable method that analyzes the conflicts in the switch zone(s). In its simplest stage, the method just analyzes the track layout while the more advanced stages also take...

  10. Weighted Clustering

    DEFF Research Database (Denmark)

    Ackerman, Margareta; Ben-David, Shai; Branzei, Simina

    2012-01-01

    We investigate a natural generalization of the classical clustering problem, considering clustering tasks in which different instances may have different weights.We conduct the first extensive theoretical analysis on the influence of weighted data on standard clustering algorithms in both...... the partitional and hierarchical settings, characterizing the conditions under which algorithms react to weights. Extending a recent framework for clustering algorithm selection, we propose intuitive properties that would allow users to choose between clustering algorithms in the weighted setting and classify...

  11. Cluster management.

    Science.gov (United States)

    Katz, R

    1992-11-01

    Cluster management is a management model that fosters decentralization of management, develops leadership potential of staff, and creates ownership of unit-based goals. Unlike shared governance models, there is no formal structure created by committees and it is less threatening for managers. There are two parts to the cluster management model. One is the formation of cluster groups, consisting of all staff and facilitated by a cluster leader. The cluster groups function for communication and problem-solving. The second part of the cluster management model is the creation of task forces. These task forces are designed to work on short-term goals, usually in response to solving one of the unit's goals. Sometimes the task forces are used for quality improvement or system problems. Clusters are groups of not more than five or six staff members, facilitated by a cluster leader. A cluster is made up of individuals who work the same shift. For example, people with job titles who work days would be in a cluster. There would be registered nurses, licensed practical nurses, nursing assistants, and unit clerks in the cluster. The cluster leader is chosen by the manager based on certain criteria and is trained for this specialized role. The concept of cluster management, criteria for choosing leaders, training for leaders, using cluster groups to solve quality improvement issues, and the learning process necessary for manager support are described.

  12. Dual capacity reciprocating compressor

    Science.gov (United States)

    Wolfe, Robert W.

    1984-01-01

    A multi-cylinder compressor 10 particularly useful in connection with northern climate heat pumps and in which different capacities are available in accordance with reversing motor 16 rotation is provided with an eccentric cam 38 on a crank pin 34 under a fraction of the connecting rods, and arranged for rotation upon the crank pin between opposite positions 180.degree. apart so that with cam rotation on the crank pin such that the crank throw is at its normal maximum value all pistons pump at full capacity, and with rotation of the crank shaft in the opposite direction the cam moves to a circumferential position on the crank pin such that the overall crank throw is zero. Pistons 24 whose connecting rods 30 ride on a crank pin 36 without a cam pump their normal rate with either crank rotational direction. Thus a small clearance volume is provided for any piston that moves when in either capacity mode of operation.

  13. Isotopic clusters

    International Nuclear Information System (INIS)

    Geraedts, J.M.P.

    1983-01-01

    Spectra of isotopically mixed clusters (dimers of SF 6 ) are calculated as well as transition frequencies. The result leads to speculations about the suitability of the laser-cluster fragmentation process for isotope separation. (Auth.)

  14. Cluster Headache

    Science.gov (United States)

    ... a role. Unlike migraine and tension headache, cluster headache generally isn't associated with triggers, such as foods, hormonal changes or stress. Once a cluster period begins, however, drinking alcohol ...

  15. Cluster Headache

    OpenAIRE

    Pearce, Iris

    1985-01-01

    Cluster headache is the most severe primary headache with recurrent pain attacks described as worse than giving birth. The aim of this paper was to make an overview of current knowledge on cluster headache with a focus on pathophysiology and treatment. This paper presents hypotheses of cluster headache pathophysiology, current treatment options and possible future therapy approaches. For years, the hypothalamus was regarded as the key structure in cluster headache, but is now thought to be pa...

  16. Categorias Cluster

    OpenAIRE

    Queiroz, Dayane Andrade

    2015-01-01

    Neste trabalho apresentamos as categorias cluster, que foram introduzidas por Aslak Bakke Buan, Robert Marsh, Markus Reineke, Idun Reiten e Gordana Todorov, com o objetivo de categoriíicar as algebras cluster criadas em 2002 por Sergey Fomin e Andrei Zelevinsky. Os autores acima, em [4], mostraram que existe uma estreita relação entre algebras cluster e categorias cluster para quivers cujo grafo subjacente é um diagrama de Dynkin. Para isto desenvolveram uma teoria tilting na estrutura triang...

  17. Meaningful Clusters

    Energy Technology Data Exchange (ETDEWEB)

    Sanfilippo, Antonio P.; Calapristi, Augustin J.; Crow, Vernon L.; Hetzler, Elizabeth G.; Turner, Alan E.

    2004-05-26

    We present an approach to the disambiguation of cluster labels that capitalizes on the notion of semantic similarity to assign WordNet senses to cluster labels. The approach provides interesting insights on how document clustering can provide the basis for developing a novel approach to word sense disambiguation.

  18. Horticultural cluster

    OpenAIRE

    SHERSTIUK S.V.; POSYLAYEVA K.I.

    2013-01-01

    In the article there are the theoretical and methodological approaches to the nature and existence of the cluster. The cluster differences from other kinds of cooperative and integration associations. Was develop by scientific-practical recommendations for forming a competitive horticultur cluster.

  19. Carrying Capacity

    DEFF Research Database (Denmark)

    Schroll, Henning; Andersen, Jan; Kjærgård, Bente

    2012-01-01

    A spatial planning act was introduced inIndonesia 1992 and renewed in 2008. It emphasised the planning role of decentralised authorities. The spatial planning act covers both spatial and environmental issues. It defines the concept of carrying capacity and includes definitions of supportive....../cities. Four different sectors (water, food production, waste, and forests) were selected as core areas for decentralised spatial planning. Indicators for SCC and ACC were identified and assessed with regard to relevance and quantifiability. For each of the indicators selected, a legal threshold or guiding...... was introduced inIndonesia 1992 and renewed in 2008. It emphasised the planning role of decentralised authorities. The spatial planning act covers both spatial and environmental issues. It defines the concept of carrying capacity and includes definitions of supportive carrying capacity (SCC) and assimilative...

  20. Cluster Matters

    DEFF Research Database (Denmark)

    Gulati, Mukesh; Lund-Thomsen, Peter; Suresh, Sangeetha

    2018-01-01

    sell their products successfully in international markets, but there is also an increasingly large consumer base within India. Indeed, Indian industrial clusters have contributed to a substantial part of this growth process, and there are several hundred registered clusters within the country...... of this handbook, which focuses on the role of CSR in MSMEs. Hence we contribute to the literature on CSR in industrial clusters and specifically CSR in Indian industrial clusters by investigating the drivers of CSR in India’s industrial clusters....

  1. Using Evaluation for CBNRM Capacity Development (Southeast Asia)

    International Development Research Centre (IDRC) Digital Library (Canada)

    Using Evaluation for CBNRM Capacity Development (Southeast Asia) ... for evaluating both the process and outcome of capacity development efforts in CBNRM. ... Cluster Case Studies Planning Workshop, Beijing, September 21-23, 2006; ...

  2. Credal Networks under Maximum Entropy

    OpenAIRE

    Lukasiewicz, Thomas

    2013-01-01

    We apply the principle of maximum entropy to select a unique joint probability distribution from the set of all joint probability distributions specified by a credal network. In detail, we start by showing that the unique joint distribution of a Bayesian tree coincides with the maximum entropy model of its conditional distributions. This result, however, does not hold anymore for general Bayesian networks. We thus present a new kind of maximum entropy models, which are computed sequentially. ...

  3. Data Clustering

    Science.gov (United States)

    Wagstaff, Kiri L.

    2012-03-01

    On obtaining a new data set, the researcher is immediately faced with the challenge of obtaining a high-level understanding from the observations. What does a typical item look like? What are the dominant trends? How many distinct groups are included in the data set, and how is each one characterized? Which observable values are common, and which rarely occur? Which items stand out as anomalies or outliers from the rest of the data? This challenge is exacerbated by the steady growth in data set size [11] as new instruments push into new frontiers of parameter space, via improvements in temporal, spatial, and spectral resolution, or by the desire to "fuse" observations from different modalities and instruments into a larger-picture understanding of the same underlying phenomenon. Data clustering algorithms provide a variety of solutions for this task. They can generate summaries, locate outliers, compress data, identify dense or sparse regions of feature space, and build data models. It is useful to note up front that "clusters" in this context refer to groups of items within some descriptive feature space, not (necessarily) to "galaxy clusters" which are dense regions in physical space. The goal of this chapter is to survey a variety of data clustering methods, with an eye toward their applicability to astronomical data analysis. In addition to improving the individual researcher’s understanding of a given data set, clustering has led directly to scientific advances, such as the discovery of new subclasses of stars [14] and gamma-ray bursts (GRBs) [38]. All clustering algorithms seek to identify groups within a data set that reflect some observed, quantifiable structure. Clustering is traditionally an unsupervised approach to data analysis, in the sense that it operates without any direct guidance about which items should be assigned to which clusters. There has been a recent trend in the clustering literature toward supporting semisupervised or constrained

  4. Exerting Capacity.

    Science.gov (United States)

    Leger, J Michael; Phillips, Carolyn A

    2017-05-01

    Patient safety has been at the forefront of nursing research since the release of the Institute of Medicine's report estimating the number of preventable adverse events in hospital settings; yet no research to date has incorporated the perspectives of bedside nurses using classical grounded theory (CGT) methodology. This CGT study explored the perceptions of bedside registered nurses regarding patient safety in adult acute care hospitals. Data analysis used three techniques unique to CGT-the constant comparative method, coding, and memoing-to explore the values, realities, and beliefs of bedside nurses about patient safety. The analysis resulted in a substantive theory, Exerting Capacity, which explained how bedside nurses balance the demands of keeping their patients safe. Exerting Capacity has implications for health care organization leaders, nursing leaders, and bedside nurses; it also has indications for future research into the concept of patient safety.

  5. Cluster evolution

    International Nuclear Information System (INIS)

    Schaeffer, R.

    1987-01-01

    The galaxy and cluster luminosity functions are constructed from a model of the mass distribution based on hierarchical clustering at an epoch where the matter distribution is non-linear. These luminosity functions are seen to reproduce the present distribution of objects as can be inferred from the observations. They can be used to deduce the redshift dependence of the cluster distribution and to extrapolate the observations towards the past. The predicted evolution of the cluster distribution is quite strong, although somewhat less rapid than predicted by the linear theory

  6. Capacity Building

    International Nuclear Information System (INIS)

    Molloy, Brian; Mallick, Shahid

    2014-01-01

    Outcomes & Recommendations: • Significant increase needed in the nuclear workforce both to replace soon-to-retire current generation and to staff large numbers of new units planned • Key message, was the importance of an integrated approach to workforce development. • IAEA and other International Organisations were asked to continue to work on Knowledge Management, Networks and E&T activities • IAEA requested to conduct Global Survey of HR needs – survey initiated but only 50% of operating countries (30% of capacity) took part, so results inconclusive

  7. Cluster Analysis of Maize Inbred Lines

    Directory of Open Access Journals (Sweden)

    Jiban Shrestha

    2016-12-01

    Full Text Available The determination of diversity among inbred lines is important for heterosis breeding. Sixty maize inbred lines were evaluated for their eight agro morphological traits during winter season of 2011 to analyze their genetic diversity. Clustering was done by average linkage method. The inbred lines were grouped into six clusters. Inbred lines grouped into Clusters II had taller plants with maximum number of leaves. The cluster III was characterized with shorter plants with minimum number of leaves. The inbred lines categorized into cluster V had early flowering whereas the group into cluster VI had late flowering time. The inbred lines grouped into the cluster III were characterized by higher value of anthesis silking interval (ASI and those of cluster VI had lower value of ASI. These results showed that the inbred lines having widely divergent clusters can be utilized in hybrid breeding programme.

  8. Spatial clustering of pixels of a multispectral image

    Science.gov (United States)

    Conger, James Lynn

    2014-08-19

    A method and system for clustering the pixels of a multispectral image is provided. A clustering system computes a maximum spectral similarity score for each pixel that indicates the similarity between that pixel and the most similar neighboring. To determine the maximum similarity score for a pixel, the clustering system generates a similarity score between that pixel and each of its neighboring pixels and then selects the similarity score that represents the highest similarity as the maximum similarity score. The clustering system may apply a filtering criterion based on the maximum similarity score so that pixels with similarity scores below a minimum threshold are not clustered. The clustering system changes the current pixel values of the pixels in a cluster based on an averaging of the original pixel values of the pixels in the cluster.

  9. Stopping of hypervelocity clusters in solids

    International Nuclear Information System (INIS)

    Anders, Christian; Ziegenhain, Gerolf; Urbassek, Herbert M; Bringa, Eduardo M

    2011-01-01

    Using molecular-dynamics simulations, we study the processes underlying the stopping of energetic clusters upon impact in matter. We investigate self-bombardment of both a metallic (Cu) and a van-der-Waals bonded (frozen Ar) target. Clusters with sizes up to N = 10 4 atoms and with energies per atom of E/N = 0.1-1600 eV atom -1 were studied. We find that the stopping force exerted on a cluster follows an N 2/3 -dependence with cluster size N; thus large clusters experience less stopping than equi-velocity atoms. In the course of being stopped, the cluster is strongly deformed and attains a roughly pancake shape. Due to the cluster inertia, maximum deformation occurs later than the maximum stopping force. The time scale of projectile stopping is set by t 0 , the time the cluster needs to cover its own diameter before impacting the target; it thus depends on both cluster size and velocity. The time when the cluster experiences its maximum stopping force is around (0.7-0.8)t 0 . We find that the cluster is deformed with huge strain rates of around 1/2t 0 ; this amounts to 10 11 -10 13 s -1 for the cases studied here. (paper)

  10. Clustering Dycom

    KAUST Repository

    Minku, Leandro L.

    2017-10-06

    Background: Software Effort Estimation (SEE) can be formulated as an online learning problem, where new projects are completed over time and may become available for training. In this scenario, a Cross-Company (CC) SEE approach called Dycom can drastically reduce the number of Within-Company (WC) projects needed for training, saving the high cost of collecting such training projects. However, Dycom relies on splitting CC projects into different subsets in order to create its CC models. Such splitting can have a significant impact on Dycom\\'s predictive performance. Aims: This paper investigates whether clustering methods can be used to help finding good CC splits for Dycom. Method: Dycom is extended to use clustering methods for creating the CC subsets. Three different clustering methods are investigated, namely Hierarchical Clustering, K-Means, and Expectation-Maximisation. Clustering Dycom is compared against the original Dycom with CC subsets of different sizes, based on four SEE databases. A baseline WC model is also included in the analysis. Results: Clustering Dycom with K-Means can potentially help to split the CC projects, managing to achieve similar or better predictive performance than Dycom. However, K-Means still requires the number of CC subsets to be pre-defined, and a poor choice can negatively affect predictive performance. EM enables Dycom to automatically set the number of CC subsets while still maintaining or improving predictive performance with respect to the baseline WC model. Clustering Dycom with Hierarchical Clustering did not offer significant advantage in terms of predictive performance. Conclusion: Clustering methods can be an effective way to automatically generate Dycom\\'s CC subsets.

  11. Clustering analysis

    International Nuclear Information System (INIS)

    Romli

    1997-01-01

    Cluster analysis is the name of group of multivariate techniques whose principal purpose is to distinguish similar entities from the characteristics they process.To study this analysis, there are several algorithms that can be used. Therefore, this topic focuses to discuss the algorithms, such as, similarity measures, and hierarchical clustering which includes single linkage, complete linkage and average linkage method. also, non-hierarchical clustering method, which is popular name K -mean method ' will be discussed. Finally, this paper will be described the advantages and disadvantages of every methods

  12. Cluster analysis

    CERN Document Server

    Everitt, Brian S; Leese, Morven; Stahl, Daniel

    2011-01-01

    Cluster analysis comprises a range of methods for classifying multivariate data into subgroups. By organizing multivariate data into such subgroups, clustering can help reveal the characteristics of any structure or patterns present. These techniques have proven useful in a wide range of areas such as medicine, psychology, market research and bioinformatics.This fifth edition of the highly successful Cluster Analysis includes coverage of the latest developments in the field and a new chapter dealing with finite mixture models for structured data.Real life examples are used throughout to demons

  13. Cluster editing

    DEFF Research Database (Denmark)

    Böcker, S.; Baumbach, Jan

    2013-01-01

    . The problem has been the inspiration for numerous algorithms in bioinformatics, aiming at clustering entities such as genes, proteins, phenotypes, or patients. In this paper, we review exact and heuristic methods that have been proposed for the Cluster Editing problem, and also applications......The Cluster Editing problem asks to transform a graph into a disjoint union of cliques using a minimum number of edge modifications. Although the problem has been proven NP-complete several times, it has nevertheless attracted much research both from the theoretical and the applied side...

  14. Occupational Clusters.

    Science.gov (United States)

    Pottawattamie County School System, Council Bluffs, IA.

    The 15 occupational clusters (transportation, fine arts and humanities, communications and media, personal service occupations, construction, hospitality and recreation, health occupations, marine science occupations, consumer and homemaking-related occupations, agribusiness and natural resources, environment, public service, business and office…

  15. Fuzzy Clustering

    DEFF Research Database (Denmark)

    Berks, G.; Keyserlingk, Diedrich Graf von; Jantzen, Jan

    2000-01-01

    A symptom is a condition indicating the presence of a disease, especially, when regarded as an aid in diagnosis.Symptoms are the smallest units indicating the existence of a disease. A syndrome on the other hand is an aggregate, set or cluster of concurrent symptoms which together indicate...... and clustering are the basic concerns in medicine. Classification depends on definitions of the classes and their required degree of participant of the elements in the cases' symptoms. In medicine imprecise conditions are the rule and therefore fuzzy methods are much more suitable than crisp ones. Fuzzy c......-mean clustering is an easy and well improved tool, which has been applied in many medical fields. We used c-mean fuzzy clustering after feature extraction from an aphasia database. Factor analysis was applied on a correlation matrix of 26 symptoms of language disorders and led to five factors. The factors...

  16. Maximum Entropy in Drug Discovery

    Directory of Open Access Journals (Sweden)

    Chih-Yuan Tseng

    2014-07-01

    Full Text Available Drug discovery applies multidisciplinary approaches either experimentally, computationally or both ways to identify lead compounds to treat various diseases. While conventional approaches have yielded many US Food and Drug Administration (FDA-approved drugs, researchers continue investigating and designing better approaches to increase the success rate in the discovery process. In this article, we provide an overview of the current strategies and point out where and how the method of maximum entropy has been introduced in this area. The maximum entropy principle has its root in thermodynamics, yet since Jaynes’ pioneering work in the 1950s, the maximum entropy principle has not only been used as a physics law, but also as a reasoning tool that allows us to process information in hand with the least bias. Its applicability in various disciplines has been abundantly demonstrated. We give several examples of applications of maximum entropy in different stages of drug discovery. Finally, we discuss a promising new direction in drug discovery that is likely to hinge on the ways of utilizing maximum entropy.

  17. Quantum reading capacity

    International Nuclear Information System (INIS)

    Pirandola, Stefano; Braunstein, Samuel L; Lupo, Cosmo; Mancini, Stefano; Giovannetti, Vittorio

    2011-01-01

    The readout of a classical memory can be modelled as a problem of quantum channel discrimination, where a decoder retrieves information by distinguishing the different quantum channels encoded in each cell of the memory (Pirandola 2011 Phys. Rev. Lett. 106 090504). In the case of optical memories, such as CDs and DVDs, this discrimination involves lossy bosonic channels and can be remarkably boosted by the use of nonclassical light (quantum reading). Here we generalize these concepts by extending the model of memory from single-cell to multi-cell encoding. In general, information is stored in a block of cells by using a channel-codeword, i.e. a sequence of channels chosen according to a classical code. Correspondingly, the readout of data is realized by a process of ‘parallel’ channel discrimination, where the entire block of cells is probed simultaneously and decoded via an optimal collective measurement. In the limit of a large block we define the quantum reading capacity of the memory, quantifying the maximum number of readable bits per cell. This notion of capacity is nontrivial when we suitably constrain the physical resources of the decoder. For optical memories (encoding bosonic channels), such a constraint is energetic and corresponds to fixing the mean total number of photons per cell. In this case, we are able to prove a separation between the quantum reading capacity and the maximum information rate achievable by classical transmitters, i.e. arbitrary classical mixtures of coherent states. In fact, we can easily construct nonclassical transmitters that are able to outperform any classical transmitter, thus showing that the advantages of quantum reading persist in the optimal multi-cell scenario. (paper)

  18. Cluster generator

    Science.gov (United States)

    Donchev, Todor I [Urbana, IL; Petrov, Ivan G [Champaign, IL

    2011-05-31

    Described herein is an apparatus and a method for producing atom clusters based on a gas discharge within a hollow cathode. The hollow cathode includes one or more walls. The one or more walls define a sputtering chamber within the hollow cathode and include a material to be sputtered. A hollow anode is positioned at an end of the sputtering chamber, and atom clusters are formed when a gas discharge is generated between the hollow anode and the hollow cathode.

  19. Cluster Bulleticity

    OpenAIRE

    Massey, Richard; Kitching, Thomas; Nagai, Daisuke

    2010-01-01

    The unique properties of dark matter are revealed during collisions between clusters of galaxies, such as the bullet cluster (1E 0657−56) and baby bullet (MACS J0025−12). These systems provide evidence for an additional, invisible mass in the separation between the distributions of their total mass, measured via gravitational lensing, and their ordinary ‘baryonic’ matter, measured via its X-ray emission. Unfortunately, the information available from these systems is limited by their rarity. C...

  20. Cluster headache

    OpenAIRE

    Leroux, Elizabeth; Ducros, Anne

    2008-01-01

    Abstract Cluster headache (CH) is a primary headache disease characterized by recurrent short-lasting attacks (15 to 180 minutes) of excruciating unilateral periorbital pain accompanied by ipsilateral autonomic signs (lacrimation, nasal congestion, ptosis, miosis, lid edema, redness of the eye). It affects young adults, predominantly males. Prevalence is estimated at 0.5–1.0/1,000. CH has a circannual and circadian periodicity, attacks being clustered (hence the name) in bouts that can occur ...

  1. Efficient heuristics for maximum common substructure search.

    Science.gov (United States)

    Englert, Péter; Kovács, Péter

    2015-05-26

    Maximum common substructure search is a computationally hard optimization problem with diverse applications in the field of cheminformatics, including similarity search, lead optimization, molecule alignment, and clustering. Most of these applications have strict constraints on running time, so heuristic methods are often preferred. However, the development of an algorithm that is both fast enough and accurate enough for most practical purposes is still a challenge. Moreover, in some applications, the quality of a common substructure depends not only on its size but also on various topological features of the one-to-one atom correspondence it defines. Two state-of-the-art heuristic algorithms for finding maximum common substructures have been implemented at ChemAxon Ltd., and effective heuristics have been developed to improve both their efficiency and the relevance of the atom mappings they provide. The implementations have been thoroughly evaluated and compared with existing solutions (KCOMBU and Indigo). The heuristics have been found to greatly improve the performance and applicability of the algorithms. The purpose of this paper is to introduce the applied methods and present the experimental results.

  2. The capacity limitations of orientation summary statistics

    Science.gov (United States)

    Attarha, Mouna; Moore, Cathleen M.

    2015-01-01

    The simultaneous–sequential method was used to test the processing capacity of establishing mean orientation summaries. Four clusters of oriented Gabor patches were presented in the peripheral visual field. One of the clusters had a mean orientation that was tilted either left or right while the mean orientations of the other three clusters were roughly vertical. All four clusters were presented at the same time in the simultaneous condition whereas the clusters appeared in temporal subsets of two in the sequential condition. Performance was lower when the means of all four clusters had to be processed concurrently than when only two had to be processed in the same amount of time. The advantage for establishing fewer summaries at a given time indicates that the processing of mean orientation engages limited-capacity processes (Experiment 1). This limitation cannot be attributed to crowding, low target-distractor discriminability, or a limited-capacity comparison process (Experiments 2 and 3). In contrast to the limitations of establishing multiple summary representations, establishing a single summary representation unfolds without interference (Experiment 4). When interpreted in the context of recent work on the capacity of summary statistics, these findings encourage reevaluation of the view that early visual perception consists of summary statistic representations that unfold independently across multiple areas of the visual field. PMID:25810160

  3. Capacity competition in electricity markets

    International Nuclear Information System (INIS)

    Crampes, Claude; Creti, Anna

    2005-01-01

    The article analyzed a two-stage game where capacity constrained electricity generators first choose how much capacity they make available and then compete in a uniform-rice auction. It is studied how capacity withholding can be used strategically to enforce market power and how uniform auctions in the price game change the results of capacity constrained competition models. The uniform auction procedure gives strong incentives to capacity restriction. At equilibrium, however, power shortage never occurs. Though auctions in electricity markets have already been studied by several economists, yet an important feature of spot trading is the capacity availability decision. In fact, for technical reasons, such as equipment maintenance or failures, the installed capacity may not work at maximum operating level and the spot market rules oblige generators to announce which plants they are willing to use and simultaneously their offer prices. Beside technical reasons, the so-called 'capacity declarations' also offer a strategic instrument for firms: by restricting capacity, operators can benefit from scarcity rents. Assessing whether generators withhold capacity is an intriguing issue for real electricity markets, though proving it is a difficult task. Several theoretical papers show that generators are able to keep wholesale prices high as compared to their generation costs. In our model, a generator is not obliged to declare all installed capacity as available, but decides on the amount of MW of electricity that is available. Hence the available capacity is an endogenous variable while the installed one is exogenous. The distinction between installed capacities and 'available' capacities allows to explain clearly whether generators exert market power by declaring unavailable some production units. Although we find multiple sub game perfect equilibria that cannot be eliminated by Pareto-dominance, all the outcomes are characterized by market price at the highest

  4. Maximum Profit Configurations of Commercial Engines

    Directory of Open Access Journals (Sweden)

    Yiran Chen

    2011-06-01

    Full Text Available An investigation of commercial engines with finite capacity low- and high-price economic subsystems and a generalized commodity transfer law [n ∝ Δ (P m] in commodity flow processes, in which effects of the price elasticities of supply and demand are introduced, is presented in this paper. Optimal cycle configurations of commercial engines for maximum profit are obtained by applying optimal control theory. In some special cases, the eventual state—market equilibrium—is solely determined by the initial conditions and the inherent characteristics of two subsystems; while the different ways of transfer affect the model in respects of the specific forms of the paths of prices and the instantaneous commodity flow, i.e., the optimal configuration.

  5. Maximum stellar iron core mass

    Indian Academy of Sciences (India)

    60, No. 3. — journal of. March 2003 physics pp. 415–422. Maximum stellar iron core mass. F W GIACOBBE. Chicago Research Center/American Air Liquide ... iron core compression due to the weight of non-ferrous matter overlying the iron cores within large .... thermal equilibrium velocities will tend to be non-relativistic.

  6. Maximum entropy beam diagnostic tomography

    International Nuclear Information System (INIS)

    Mottershead, C.T.

    1985-01-01

    This paper reviews the formalism of maximum entropy beam diagnostic tomography as applied to the Fusion Materials Irradiation Test (FMIT) prototype accelerator. The same formalism has also been used with streak camera data to produce an ultrahigh speed movie of the beam profile of the Experimental Test Accelerator (ETA) at Livermore. 11 refs., 4 figs

  7. Maximum entropy beam diagnostic tomography

    International Nuclear Information System (INIS)

    Mottershead, C.T.

    1985-01-01

    This paper reviews the formalism of maximum entropy beam diagnostic tomography as applied to the Fusion Materials Irradiation Test (FMIT) prototype accelerator. The same formalism has also been used with streak camera data to produce an ultrahigh speed movie of the beam profile of the Experimental Test Accelerator (ETA) at Livermore

  8. A portable storage maximum thermometer

    International Nuclear Information System (INIS)

    Fayart, Gerard.

    1976-01-01

    A clinical thermometer storing the voltage corresponding to the maximum temperature in an analog memory is described. End of the measurement is shown by a lamp switch out. The measurement time is shortened by means of a low thermal inertia platinum probe. This portable thermometer is fitted with cell test and calibration system [fr

  9. Neutron spectra unfolding with maximum entropy and maximum likelihood

    International Nuclear Information System (INIS)

    Itoh, Shikoh; Tsunoda, Toshiharu

    1989-01-01

    A new unfolding theory has been established on the basis of the maximum entropy principle and the maximum likelihood method. This theory correctly embodies the Poisson statistics of neutron detection, and always brings a positive solution over the whole energy range. Moreover, the theory unifies both problems of overdetermined and of underdetermined. For the latter, the ambiguity in assigning a prior probability, i.e. the initial guess in the Bayesian sense, has become extinct by virtue of the principle. An approximate expression of the covariance matrix for the resultant spectra is also presented. An efficient algorithm to solve the nonlinear system, which appears in the present study, has been established. Results of computer simulation showed the effectiveness of the present theory. (author)

  10. On Maximum Entropy and Inference

    Directory of Open Access Journals (Sweden)

    Luigi Gresele

    2017-11-01

    Full Text Available Maximum entropy is a powerful concept that entails a sharp separation between relevant and irrelevant variables. It is typically invoked in inference, once an assumption is made on what the relevant variables are, in order to estimate a model from data, that affords predictions on all other (dependent variables. Conversely, maximum entropy can be invoked to retrieve the relevant variables (sufficient statistics directly from the data, once a model is identified by Bayesian model selection. We explore this approach in the case of spin models with interactions of arbitrary order, and we discuss how relevant interactions can be inferred. In this perspective, the dimensionality of the inference problem is not set by the number of parameters in the model, but by the frequency distribution of the data. We illustrate the method showing its ability to recover the correct model in a few prototype cases and discuss its application on a real dataset.

  11. Simulations of floor cooling system capacity

    International Nuclear Information System (INIS)

    Odyjas, Andrzej; Górka, Andrzej

    2013-01-01

    Floor cooling system capacity depends on its physical and operative parameters. Using numerical simulations, it appears that cooling capacity of the system largely depends on the type of cooling loads occurring in the room. In the case of convective cooling loads capacity of the system is small. However, when radiation flux falls directly on the floor the system significantly increases productivity. The article describes the results of numerical simulations which allow to determine system capacity in steady thermal conditions, depending on the type of physical parameters of the system and the type of cooling load occurring in the room. Moreover, the paper sets out the limits of system capacity while maintaining a minimum temperature of the floor surface equal to 20 °C. The results are helpful for designing system capacity in different type of cooling loads and show maximum system capacity in acceptable thermal comfort condition. -- Highlights: ► We have developed numerical model for simulation of floor cooling system. ► We have described floor system capacity depending on its physical parameters. ► We have described floor system capacity depending on type of cooling loads. ► The most important in the obtained cooling capacities is the type of cooling loads. ► The paper sets out the possible maximum cooling floor system capacity

  12. Maximum Water Hammer Sensitivity Analysis

    OpenAIRE

    Jalil Emadi; Abbas Solemani

    2011-01-01

    Pressure waves and Water Hammer occur in a pumping system when valves are closed or opened suddenly or in the case of sudden failure of pumps. Determination of maximum water hammer is considered one of the most important technical and economical items of which engineers and designers of pumping stations and conveyance pipelines should take care. Hammer Software is a recent application used to simulate water hammer. The present study focuses on determining significance of ...

  13. Maximum Gene-Support Tree

    Directory of Open Access Journals (Sweden)

    Yunfeng Shan

    2008-01-01

    Full Text Available Genomes and genes diversify during evolution; however, it is unclear to what extent genes still retain the relationship among species. Model species for molecular phylogenetic studies include yeasts and viruses whose genomes were sequenced as well as plants that have the fossil-supported true phylogenetic trees available. In this study, we generated single gene trees of seven yeast species as well as single gene trees of nine baculovirus species using all the orthologous genes among the species compared. Homologous genes among seven known plants were used for validation of the finding. Four algorithms—maximum parsimony (MP, minimum evolution (ME, maximum likelihood (ML, and neighbor-joining (NJ—were used. Trees were reconstructed before and after weighting the DNA and protein sequence lengths among genes. Rarely a gene can always generate the “true tree” by all the four algorithms. However, the most frequent gene tree, termed “maximum gene-support tree” (MGS tree, or WMGS tree for the weighted one, in yeasts, baculoviruses, or plants was consistently found to be the “true tree” among the species. The results provide insights into the overall degree of divergence of orthologous genes of the genomes analyzed and suggest the following: 1 The true tree relationship among the species studied is still maintained by the largest group of orthologous genes; 2 There are usually more orthologous genes with higher similarities between genetically closer species than between genetically more distant ones; and 3 The maximum gene-support tree reflects the phylogenetic relationship among species in comparison.

  14. LCLS Maximum Credible Beam Power

    International Nuclear Information System (INIS)

    Clendenin, J.

    2005-01-01

    The maximum credible beam power is defined as the highest credible average beam power that the accelerator can deliver to the point in question, given the laws of physics, the beam line design, and assuming all protection devices have failed. For a new accelerator project, the official maximum credible beam power is determined by project staff in consultation with the Radiation Physics Department, after examining the arguments and evidence presented by the appropriate accelerator physicist(s) and beam line engineers. The definitive parameter becomes part of the project's safety envelope. This technical note will first review the studies that were done for the Gun Test Facility (GTF) at SSRL, where a photoinjector similar to the one proposed for the LCLS is being tested. In Section 3 the maximum charge out of the gun for a single rf pulse is calculated. In Section 4, PARMELA simulations are used to track the beam from the gun to the end of the photoinjector. Finally in Section 5 the beam through the matching section and injected into Linac-1 is discussed

  15. A study of the dynamical state of the hot plasma in galaxy clusters using XMM-Newton data and numerical simulation

    International Nuclear Information System (INIS)

    Solovyeva, Lilia

    2008-01-01

    Cluster of galaxies are the largest and youngest objects in the Universe and these objects are very interesting for study the cosmology. In this moment with the capacity of the instruments (XMM-Newton, Chandra) and with numerical simulations it is possible to study the dynamical state of gas in the cluster during their formation. And plus, now, we have the possibility to study the cluster in different wavelengths (optics, radio, X-ray). Our study helps us to understand the physics processes in clusters. In our work we studied the galaxy cluster around the maximum core collapse. We used the X-ray data, how the first indicators of dynamical state of gas. After with the help of numerical simulation and optics data we performed the completed analysis with the proposition of merger scenario possible. We performed the detailed analysis of two clusters (CL0016+16 and A548b), these clusters presents the signature of major merger and also we studied the cluster from numerical simulation (Cluster 6) around the major merger. (author) [fr

  16. Clustering Dycom

    KAUST Repository

    Minku, Leandro L.; Hou, Siqing

    2017-01-01

    baseline WC model is also included in the analysis. Results: Clustering Dycom with K-Means can potentially help to split the CC projects, managing to achieve similar or better predictive performance than Dycom. However, K-Means still requires the number

  17. Generic maximum likely scale selection

    DEFF Research Database (Denmark)

    Pedersen, Kim Steenstrup; Loog, Marco; Markussen, Bo

    2007-01-01

    in this work is on applying this selection principle under a Brownian image model. This image model provides a simple scale invariant prior for natural images and we provide illustrative examples of the behavior of our scale estimation on such images. In these illustrative examples, estimation is based......The fundamental problem of local scale selection is addressed by means of a novel principle, which is based on maximum likelihood estimation. The principle is generally applicable to a broad variety of image models and descriptors, and provides a generic scale estimation methodology. The focus...

  18. Extreme Maximum Land Surface Temperatures.

    Science.gov (United States)

    Garratt, J. R.

    1992-09-01

    There are numerous reports in the literature of observations of land surface temperatures. Some of these, almost all made in situ, reveal maximum values in the 50°-70°C range, with a few, made in desert regions, near 80°C. Consideration of a simplified form of the surface energy balance equation, utilizing likely upper values of absorbed shortwave flux (1000 W m2) and screen air temperature (55°C), that surface temperatures in the vicinity of 90°-100°C may occur for dry, darkish soils of low thermal conductivity (0.1-0.2 W m1 K1). Numerical simulations confirm this and suggest that temperature gradients in the first few centimeters of soil may reach 0.5°-1°C mm1 under these extreme conditions. The study bears upon the intrinsic interest of identifying extreme maximum temperatures and yields interesting information regarding the comfort zone of animals (including man).

  19. Cluster forcing

    DEFF Research Database (Denmark)

    Christensen, Thomas Budde

    The cluster theory attributed to Michael Porter has significantly influenced industrial policies in countries across Europe and North America since the beginning of the 1990s. Institutions such as the EU, OECD and the World Bank and governments in countries such as the UK, France, The Netherlands...... or management. Both the Accelerate Wales and the Accelerate Cluster programmes target this issue by trying to establish networks between companies that can be used to supply knowledge from research institutions to manufacturing companies. The paper concludes that public sector interventions can make...... businesses. The universities were not considered by the participating companies to be important parts of the local business environment and inputs from universities did not appear to be an important source to access knowledge about new product development or new techniques in production, distribution...

  20. Regional Innovation Clusters

    Data.gov (United States)

    Small Business Administration — The Regional Innovation Clusters serve a diverse group of sectors and geographies. Three of the initial pilot clusters, termed Advanced Defense Technology clusters,...

  1. 33 CFR 183.41 - Persons capacity: Outboard boats.

    Science.gov (United States)

    2010-07-01

    ... tank weight from table 4 of subpart H of this part; or (2) For boats with a maximum persons capacity less than 550 pounds, the maximum persons capacity determined in the following manner: (i) Float the... control weight, battery weight, and full portable fuel tank weight, if any, shown in table 4 of subpart H...

  2. Cluster analysis

    OpenAIRE

    Mucha, Hans-Joachim; Sofyan, Hizir

    2000-01-01

    As an explorative technique, duster analysis provides a description or a reduction in the dimension of the data. It classifies a set of observations into two or more mutually exclusive unknown groups based on combinations of many variables. Its aim is to construct groups in such a way that the profiles of objects in the same groups are relatively homogenous whereas the profiles of objects in different groups are relatively heterogeneous. Clustering is distinct from classification techniques, ...

  3. Clustering algorithms for Stokes space modulation format recognition

    DEFF Research Database (Denmark)

    Boada, Ricard; Borkowski, Robert; Tafur Monroy, Idelfonso

    2015-01-01

    influences the performance of the detection process, particularly at low signal-to-noise ratios. This paper reports on an extensive study of six different clustering algorithms: k-means, expectation maximization, density-based DBSCAN and OPTICS, spectral clustering and maximum likelihood clustering, used...

  4. System for memorizing maximum values

    Science.gov (United States)

    Bozeman, Richard J., Jr.

    1992-08-01

    The invention discloses a system capable of memorizing maximum sensed values. The system includes conditioning circuitry which receives the analog output signal from a sensor transducer. The conditioning circuitry rectifies and filters the analog signal and provides an input signal to a digital driver, which may be either linear or logarithmic. The driver converts the analog signal to discrete digital values, which in turn triggers an output signal on one of a plurality of driver output lines n. The particular output lines selected is dependent on the converted digital value. A microfuse memory device connects across the driver output lines, with n segments. Each segment is associated with one driver output line, and includes a microfuse that is blown when a signal appears on the associated driver output line.

  5. Remarks on the maximum luminosity

    Science.gov (United States)

    Cardoso, Vitor; Ikeda, Taishi; Moore, Christopher J.; Yoo, Chul-Moon

    2018-04-01

    The quest for fundamental limitations on physical processes is old and venerable. Here, we investigate the maximum possible power, or luminosity, that any event can produce. We show, via full nonlinear simulations of Einstein's equations, that there exist initial conditions which give rise to arbitrarily large luminosities. However, the requirement that there is no past horizon in the spacetime seems to limit the luminosity to below the Planck value, LP=c5/G . Numerical relativity simulations of critical collapse yield the largest luminosities observed to date, ≈ 0.2 LP . We also present an analytic solution to the Einstein equations which seems to give an unboundedly large luminosity; this will guide future numerical efforts to investigate super-Planckian luminosities.

  6. Maximum mutual information regularized classification

    KAUST Repository

    Wang, Jim Jing-Yan

    2014-09-07

    In this paper, a novel pattern classification approach is proposed by regularizing the classifier learning to maximize mutual information between the classification response and the true class label. We argue that, with the learned classifier, the uncertainty of the true class label of a data sample should be reduced by knowing its classification response as much as possible. The reduced uncertainty is measured by the mutual information between the classification response and the true class label. To this end, when learning a linear classifier, we propose to maximize the mutual information between classification responses and true class labels of training samples, besides minimizing the classification error and reducing the classifier complexity. An objective function is constructed by modeling mutual information with entropy estimation, and it is optimized by a gradient descend method in an iterative algorithm. Experiments on two real world pattern classification problems show the significant improvements achieved by maximum mutual information regularization.

  7. Scintillation counter, maximum gamma aspect

    International Nuclear Information System (INIS)

    Thumim, A.D.

    1975-01-01

    A scintillation counter, particularly for counting gamma ray photons, includes a massive lead radiation shield surrounding a sample-receiving zone. The shield is disassembleable into a plurality of segments to allow facile installation and removal of a photomultiplier tube assembly, the segments being so constructed as to prevent straight-line access of external radiation through the shield into radiation-responsive areas. Provisions are made for accurately aligning the photomultiplier tube with respect to one or more sample-transmitting bores extending through the shield to the sample receiving zone. A sample elevator, used in transporting samples into the zone, is designed to provide a maximum gamma-receiving aspect to maximize the gamma detecting efficiency. (U.S.)

  8. Maximum mutual information regularized classification

    KAUST Repository

    Wang, Jim Jing-Yan; Wang, Yi; Zhao, Shiguang; Gao, Xin

    2014-01-01

    In this paper, a novel pattern classification approach is proposed by regularizing the classifier learning to maximize mutual information between the classification response and the true class label. We argue that, with the learned classifier, the uncertainty of the true class label of a data sample should be reduced by knowing its classification response as much as possible. The reduced uncertainty is measured by the mutual information between the classification response and the true class label. To this end, when learning a linear classifier, we propose to maximize the mutual information between classification responses and true class labels of training samples, besides minimizing the classification error and reducing the classifier complexity. An objective function is constructed by modeling mutual information with entropy estimation, and it is optimized by a gradient descend method in an iterative algorithm. Experiments on two real world pattern classification problems show the significant improvements achieved by maximum mutual information regularization.

  9. Maximum entropy and Bayesian methods

    International Nuclear Information System (INIS)

    Smith, C.R.; Erickson, G.J.; Neudorfer, P.O.

    1992-01-01

    Bayesian probability theory and Maximum Entropy methods are at the core of a new view of scientific inference. These 'new' ideas, along with the revolution in computational methods afforded by modern computers allow astronomers, electrical engineers, image processors of any type, NMR chemists and physicists, and anyone at all who has to deal with incomplete and noisy data, to take advantage of methods that, in the past, have been applied only in some areas of theoretical physics. The title workshops have been the focus of a group of researchers from many different fields, and this diversity is evident in this book. There are tutorial and theoretical papers, and applications in a very wide variety of fields. Almost any instance of dealing with incomplete and noisy data can be usefully treated by these methods, and many areas of theoretical research are being enhanced by the thoughtful application of Bayes' theorem. Contributions contained in this volume present a state-of-the-art overview that will be influential and useful for many years to come

  10. 49 CFR 237.71 - Determination of bridge load capacities.

    Science.gov (United States)

    2010-10-01

    ... capacity shall be determined. (g) Bridge load capacity may be expressed in terms of numerical values related to a standard system of bridge loads, but shall in any case be stated in terms of weight and...) Bridge load capacity may be expressed in terms of both normal and maximum load conditions. Operation of...

  11. Genetic Diversity and Relationships of Neolamarckia cadamba (Roxb. Bosser progenies through cluster analysis

    Directory of Open Access Journals (Sweden)

    M. Preethi Shree

    2018-04-01

    Full Text Available Genetic diversity analysis was conducted for biometric attributes in 20 progenies of Neolamarckia cadamba. The application of D2 clustering technique in Neolamarckia cadamba genetic resources resolved the 20 progenies into five clusters. The maximum intra cluster distance was shown by the cluster II. The maximum inter cluster distance was recorded between cluster III and V which indicated the presence of wider genetic distance between Neolamarckia cadamba progenies. Among the growth attributes, volume (36.84 % contributed maximum towards genetic divergence followed by bole height, basal diameter, tree height, number of branches in Neolamarckia cadamba progenies.

  12. Maximum entropy principal for transportation

    International Nuclear Information System (INIS)

    Bilich, F.; Da Silva, R.

    2008-01-01

    In this work we deal with modeling of the transportation phenomenon for use in the transportation planning process and policy-impact studies. The model developed is based on the dependence concept, i.e., the notion that the probability of a trip starting at origin i is dependent on the probability of a trip ending at destination j given that the factors (such as travel time, cost, etc.) which affect travel between origin i and destination j assume some specific values. The derivation of the solution of the model employs the maximum entropy principle combining a priori multinomial distribution with a trip utility concept. This model is utilized to forecast trip distributions under a variety of policy changes and scenarios. The dependence coefficients are obtained from a regression equation where the functional form is derived based on conditional probability and perception of factors from experimental psychology. The dependence coefficients encode all the information that was previously encoded in the form of constraints. In addition, the dependence coefficients encode information that cannot be expressed in the form of constraints for practical reasons, namely, computational tractability. The equivalence between the standard formulation (i.e., objective function with constraints) and the dependence formulation (i.e., without constraints) is demonstrated. The parameters of the dependence-based trip-distribution model are estimated, and the model is also validated using commercial air travel data in the U.S. In addition, policy impact analyses (such as allowance of supersonic flights inside the U.S. and user surcharge at noise-impacted airports) on air travel are performed.

  13. Nuclear clustering - a cluster core model study

    International Nuclear Information System (INIS)

    Paul Selvi, G.; Nandhini, N.; Balasubramaniam, M.

    2015-01-01

    Nuclear clustering, similar to other clustering phenomenon in nature is a much warranted study, since it would help us in understanding the nature of binding of the nucleons inside the nucleus, closed shell behaviour when the system is highly deformed, dynamics and structure at extremes. Several models account for the clustering phenomenon of nuclei. We present in this work, a cluster core model study of nuclear clustering in light mass nuclei

  14. Last Glacial Maximum Salinity Reconstruction

    Science.gov (United States)

    Homola, K.; Spivack, A. J.

    2016-12-01

    It has been previously demonstrated that salinity can be reconstructed from sediment porewater. The goal of our study is to reconstruct high precision salinity during the Last Glacial Maximum (LGM). Salinity is usually determined at high precision via conductivity, which requires a larger volume of water than can be extracted from a sediment core, or via chloride titration, which yields lower than ideal precision. It has been demonstrated for water column samples that high precision density measurements can be used to determine salinity at the precision of a conductivity measurement using the equation of state of seawater. However, water column seawater has a relatively constant composition, in contrast to porewater, where variations from standard seawater composition occur. These deviations, which affect the equation of state, must be corrected for through precise measurements of each ion's concentration and knowledge of apparent partial molar density in seawater. We have developed a density-based method for determining porewater salinity that requires only 5 mL of sample, achieving density precisions of 10-6 g/mL. We have applied this method to porewater samples extracted from long cores collected along a N-S transect across the western North Atlantic (R/V Knorr cruise KN223). Density was determined to a precision of 2.3x10-6 g/mL, which translates to salinity uncertainty of 0.002 gms/kg if the effect of differences in composition is well constrained. Concentrations of anions (Cl-, and SO4-2) and cations (Na+, Mg+, Ca+2, and K+) were measured. To correct salinities at the precision required to unravel LGM Meridional Overturning Circulation, our ion precisions must be better than 0.1% for SO4-/Cl- and Mg+/Na+, and 0.4% for Ca+/Na+, and K+/Na+. Alkalinity, pH and Dissolved Inorganic Carbon of the porewater were determined to precisions better than 4% when ratioed to Cl-, and used to calculate HCO3-, and CO3-2. Apparent partial molar densities in seawater were

  15. Maximum Parsimony on Phylogenetic networks

    Science.gov (United States)

    2012-01-01

    Background Phylogenetic networks are generalizations of phylogenetic trees, that are used to model evolutionary events in various contexts. Several different methods and criteria have been introduced for reconstructing phylogenetic trees. Maximum Parsimony is a character-based approach that infers a phylogenetic tree by minimizing the total number of evolutionary steps required to explain a given set of data assigned on the leaves. Exact solutions for optimizing parsimony scores on phylogenetic trees have been introduced in the past. Results In this paper, we define the parsimony score on networks as the sum of the substitution costs along all the edges of the network; and show that certain well-known algorithms that calculate the optimum parsimony score on trees, such as Sankoff and Fitch algorithms extend naturally for networks, barring conflicting assignments at the reticulate vertices. We provide heuristics for finding the optimum parsimony scores on networks. Our algorithms can be applied for any cost matrix that may contain unequal substitution costs of transforming between different characters along different edges of the network. We analyzed this for experimental data on 10 leaves or fewer with at most 2 reticulations and found that for almost all networks, the bounds returned by the heuristics matched with the exhaustively determined optimum parsimony scores. Conclusion The parsimony score we define here does not directly reflect the cost of the best tree in the network that displays the evolution of the character. However, when searching for the most parsimonious network that describes a collection of characters, it becomes necessary to add additional cost considerations to prefer simpler structures, such as trees over networks. The parsimony score on a network that we describe here takes into account the substitution costs along the additional edges incident on each reticulate vertex, in addition to the substitution costs along the other edges which are

  16. Heavy hitters via cluster-preserving clustering

    DEFF Research Database (Denmark)

    Larsen, Kasper Green; Nelson, Jelani; Nguyen, Huy L.

    2016-01-01

    In the turnstile lp heavy hitters problem with parameter ε, one must maintain a high-dimensional vector xεRn subject to updates of the form update (i,Δ) causing the change xi≤ ← xi + Δ, where iε[n], ΔεR. Upon receiving a query, the goal is to report every "heavy hitter" iε[n] with |xi| ≥ε......|x|p as part of a list L⊆[n] of size O(1/εp), i.e. proportional to the maximum possible number of heavy hitters. For any pε(0,2] the COUNTSKETCH of [CCFC04] solves lp heavy hitters using O(ε-plog n) words of space with O(log n) update time, O(nlog n) query time to output L, and whose output after any query......, providing correctness whp. In fact, a simpler version of our algorithm for p = 1 in the strict turnstile model answers queries even faster than the "dyadic trick" by roughly a log n factor, dominating it in all regards. Our main innovation is an efficient reduction from the heavy hitters to a clustering...

  17. Probing dark energy via galaxy cluster outskirts

    Science.gov (United States)

    Morandi, Andrea; Sun, Ming

    2016-04-01

    We present a Bayesian approach to combine Planck data and the X-ray physical properties of the intracluster medium in the virialization region of a sample of 320 galaxy clusters (0.056 definition of cluster boundary radius is more tenable, namely based on a fixed overdensity with respect to the critical density of the Universe. This novel cosmological test has the capacity to provide a generational leap forward in our understanding of the equation of state of dark energy.

  18. Cluster headache

    Directory of Open Access Journals (Sweden)

    Ducros Anne

    2008-07-01

    Full Text Available Abstract Cluster headache (CH is a primary headache disease characterized by recurrent short-lasting attacks (15 to 180 minutes of excruciating unilateral periorbital pain accompanied by ipsilateral autonomic signs (lacrimation, nasal congestion, ptosis, miosis, lid edema, redness of the eye. It affects young adults, predominantly males. Prevalence is estimated at 0.5–1.0/1,000. CH has a circannual and circadian periodicity, attacks being clustered (hence the name in bouts that can occur during specific months of the year. Alcohol is the only dietary trigger of CH, strong odors (mainly solvents and cigarette smoke and napping may also trigger CH attacks. During bouts, attacks may happen at precise hours, especially during the night. During the attacks, patients tend to be restless. CH may be episodic or chronic, depending on the presence of remission periods. CH is associated with trigeminovascular activation and neuroendocrine and vegetative disturbances, however, the precise cautive mechanisms remain unknown. Involvement of the hypothalamus (a structure regulating endocrine function and sleep-wake rhythms has been confirmed, explaining, at least in part, the cyclic aspects of CH. The disease is familial in about 10% of cases. Genetic factors play a role in CH susceptibility, and a causative role has been suggested for the hypocretin receptor gene. Diagnosis is clinical. Differential diagnoses include other primary headache diseases such as migraine, paroxysmal hemicrania and SUNCT syndrome. At present, there is no curative treatment. There are efficient treatments to shorten the painful attacks (acute treatments and to reduce the number of daily attacks (prophylactic treatments. Acute treatment is based on subcutaneous administration of sumatriptan and high-flow oxygen. Verapamil, lithium, methysergide, prednisone, greater occipital nerve blocks and topiramate may be used for prophylaxis. In refractory cases, deep-brain stimulation of the

  19. STUDY ON MAXIMUM SPECIFIC SLUDGE ACIVITY OF DIFFERENT ANAEROBIC GRANULAR SLUDGE BY BATCH TESTS

    Institute of Scientific and Technical Information of China (English)

    2001-01-01

    The maximum specific sludge activity of granular sludge from large-scale UASB, IC and Biobed anaerobic reactors were investigated by batch tests. The limitation factors related to maximum specific sludge activity (diffusion, substrate sort, substrate concentration and granular size) were studied. The general principle and procedure for the precise measurement of maximum specific sludge activity were suggested. The potential capacity of loading rate of the IC and Biobed anaerobic reactors were analyzed and compared by use of the batch tests results.

  20. Two-dimensional maximum entropy image restoration

    International Nuclear Information System (INIS)

    Brolley, J.E.; Lazarus, R.B.; Suydam, B.R.; Trussell, H.J.

    1977-07-01

    An optical check problem was constructed to test P LOG P maximum entropy restoration of an extremely distorted image. Useful recovery of the original image was obtained. Comparison with maximum a posteriori restoration is made. 7 figures

  1. Brightest Cluster Galaxies in REXCESS Clusters

    Science.gov (United States)

    Haarsma, Deborah B.; Leisman, L.; Bruch, S.; Donahue, M.

    2009-01-01

    Most galaxy clusters contain a Brightest Cluster Galaxy (BCG) which is larger than the other cluster ellipticals and has a more extended profile. In the hierarchical model, the BCG forms through many galaxy mergers in the crowded center of the cluster, and thus its properties give insight into the assembly of the cluster as a whole. In this project, we are working with the Representative XMM-Newton Cluster Structure Survey (REXCESS) team (Boehringer et al 2007) to study BCGs in 33 X-ray luminous galaxy clusters, 0.055 < z < 0.183. We are imaging the BCGs in R band at the Southern Observatory for Astrophysical Research (SOAR) in Chile. In this poster, we discuss our methods and give preliminary measurements of the BCG magnitudes, morphology, and stellar mass. We compare these BCG properties with the properties of their host clusters, particularly of the X-ray emitting gas.

  2. Cluster infall in the concordance LCDM model

    OpenAIRE

    Pivato, Maximiliano C.; Padilla, Nelson D.; Lambas, Diego G.

    2005-01-01

    We perform statistical analyses of the infall of dark-matter onto clusters in numerical simulations within the concordance LCDM model. By studying the infall profile around clusters of different mass, we find a linear relation between the maximum infall velocity and mass which reach 900km/s for the most massive groups. The maximum infall velocity and the group mass follow a suitable power law fit of the form, V_{inf}^{max} = (M/m_0)^{gamma}. By comparing the measured infall velocity to the li...

  3. Partitional clustering algorithms

    CERN Document Server

    2015-01-01

    This book summarizes the state-of-the-art in partitional clustering. Clustering, the unsupervised classification of patterns into groups, is one of the most important tasks in exploratory data analysis. Primary goals of clustering include gaining insight into, classifying, and compressing data. Clustering has a long and rich history that spans a variety of scientific disciplines including anthropology, biology, medicine, psychology, statistics, mathematics, engineering, and computer science. As a result, numerous clustering algorithms have been proposed since the early 1950s. Among these algorithms, partitional (nonhierarchical) ones have found many applications, especially in engineering and computer science. This book provides coverage of consensus clustering, constrained clustering, large scale and/or high dimensional clustering, cluster validity, cluster visualization, and applications of clustering. Examines clustering as it applies to large and/or high-dimensional data sets commonly encountered in reali...

  4. Receiver function estimated by maximum entropy deconvolution

    Institute of Scientific and Technical Information of China (English)

    吴庆举; 田小波; 张乃铃; 李卫平; 曾融生

    2003-01-01

    Maximum entropy deconvolution is presented to estimate receiver function, with the maximum entropy as the rule to determine auto-correlation and cross-correlation functions. The Toeplitz equation and Levinson algorithm are used to calculate the iterative formula of error-predicting filter, and receiver function is then estimated. During extrapolation, reflective coefficient is always less than 1, which keeps maximum entropy deconvolution stable. The maximum entropy of the data outside window increases the resolution of receiver function. Both synthetic and real seismograms show that maximum entropy deconvolution is an effective method to measure receiver function in time-domain.

  5. Capacity in the energy underwriting market - an overview

    International Nuclear Information System (INIS)

    Mulhall, L.

    1992-01-01

    It is important to clarify the different definitions of capacity, and in the insurance business this work has two distinct interpretations. The capacity of a Lloyd's syndicate is described by its ability to write premium income. Of more interest is the maximum ability and willingness of an underwriter, whether Lloyds's or Company, to commit itself to catastrophe loss policies. Or put more simply, what is the size of their line? It is this capacity for covering high value catastrophic exposure or Target Risks that will be discussed using the definition that capacity in this case is: ''The ability of insurance markets to cover any single maximum loss''. (Author)

  6. Maximum Power from a Solar Panel

    Directory of Open Access Journals (Sweden)

    Michael Miller

    2010-01-01

    Full Text Available Solar energy has become a promising alternative to conventional fossil fuel sources. Solar panels are used to collect solar radiation and convert it into electricity. One of the techniques used to maximize the effectiveness of this energy alternative is to maximize the power output of the solar collector. In this project the maximum power is calculated by determining the voltage and the current of maximum power. These quantities are determined by finding the maximum value for the equation for power using differentiation. After the maximum values are found for each time of day, each individual quantity, voltage of maximum power, current of maximum power, and maximum power is plotted as a function of the time of day.

  7. Total iron binding capacity

    Science.gov (United States)

    ... page: //medlineplus.gov/ency/article/003489.htm Total iron binding capacity To use the sharing features on this page, please enable JavaScript. Total iron binding capacity (TIBC) is a blood test to ...

  8. Air capacity for Sydney

    OpenAIRE

    Forsyth, Peter

    2013-01-01

    Like most large cities, Sydney has an airport problem. Demand is increasing faster than supply, and additional capacity will be needed if costly rationing, and delays, are to be avoided. However, compared to many cities, the problems facing Sydney are modest. At the moment, demand is only just exceeding capacity. There is a good chance that the available capacity will be rationed efficiently. Options for expanding capacity are being evaluated well. There may be problems in the future- poor op...

  9. Responsibility and Capacities

    DEFF Research Database (Denmark)

    Ryberg, Jesper

    2014-01-01

    That responsible moral agency presupposes certain mental capacities, constitutes a widely accepted view among theorists. Moreover, it is often assumed that degrees in the development of the relevant capacities co-vary with degrees of responsibility. In this article it is argued that, the move from...... the view that responsibility requires certain mental capacities to the position that degrees of responsibility co-vary with degrees of the development of the mental capacities, is premature....

  10. CDMA systems capacity engineering

    CERN Document Server

    Kim, Kiseon

    2004-01-01

    This new hands-on resource tackles capacity planning and engineering issues that are crucial to optimizing wireless communication systems performance. Going beyond the system physical level and investigating CDMA system capacity at the service level, this volume is the single-source for engineering and analyzing systems capacity and resources.

  11. PAYMENT CAPACITY SENSITIVITY FACTORS

    Directory of Open Access Journals (Sweden)

    Daniel BRÎNDESCU – OLARIU

    2014-11-01

    The results of the study facilitate the determination and classification of the main sensitivity factors for the payment capacity at sample level, the establishment of general benchmarks for the payment capacity (as no such benchmarks currently exist in the Romanian literature and the identification of the mechanisms through which the variation of different factors impacts the payment capacity.

  12. Intercalating graphene with clusters of Fe3O4 nanocrystals for electrochemical supercapacitors

    Science.gov (United States)

    Ke, Qingqing; Tang, Chunhua; Liu, Yanqiong; Liu, Huajun; Wang, John

    2014-04-01

    A hierarchical nanostructure consisting of graphene sheets intercalated by clusters of Fe3O4 nanocystals is developed for high-performance supercapacitor electrode. Here we show that the negatively charged graphene oxide (GO) and positively charged Fe3O4 clusters enable a strong electrostatic interaction, generating a hierarchical 3D nanostructure, which gives rise to the intercalated composites through a rational hydrothermal process. The electrocapacitive behavior of the resultant composites is systematically investigated by cyclic voltammeter and galvanostatic charge-discharge techniques, where a positive synergistic effect between graphene and Fe3O4 clusters is identified. A maximum specific capacitance of 169 F g-1 is achieved in the Fe3O4 clusters decorated with effectively reduced graphene oxide (Fe3O4-rGO-12h), which is much higher than those of rGO (101 F g-1) and Fe3O4 (68 F g-1) at the current density of 1 Ag-1. Moreover, this intercalated hierarchical nanostructure demonstrates a good capacitance retention, retaining over 88% of the initial capacity after 1000 cycles.

  13. Diversity among galaxy clusters

    International Nuclear Information System (INIS)

    Struble, M.F.; Rood, H.J.

    1988-01-01

    The classification of galaxy clusters is discussed. Consideration is given to the classification scheme of Abell (1950's), Zwicky (1950's), Morgan, Matthews, and Schmidt (1964), and Morgan-Bautz (1970). Galaxies can be classified based on morphology, chemical composition, spatial distribution, and motion. The correlation between a galaxy's environment and morphology is examined. The classification scheme of Rood-Sastry (1971), which is based on clusters's morphology and galaxy population, is described. The six types of clusters they define include: (1) a cD-cluster dominated by a single large galaxy, (2) a cluster dominated by a binary, (3) a core-halo cluster, (4) a cluster dominated by several bright galaxies, (5) a cluster appearing flattened, and (6) an irregularly shaped cluster. Attention is also given to the evolution of cluster structures, which is related to initial density and cluster motion

  14. Ionic cluster size distributions of swollen nafion/sulfated beta-cyclodextrin membranes characterized by nuclear magnetic resonance cryoporometry.

    Science.gov (United States)

    Jeon, Jae-Deok; Kwak, Seung-Yeop

    2007-08-16

    Nafion/sb-CD membranes were prepared by mixing 5 wt% Nafion solution with H+-form sulfated beta-cyclodextrin (sb-CD), and their water uptakes, ion exchange capacities (IECs), and ionic cluster size distributions were measured. Gravimetric and thermogravimetric measurements showed that the water uptake of the membranes increased with increases in their sb-CD content. The IECs of the membrane were measured with acid-base titration and found to increase with increases in the sb-CD content, reaching 0.96 mequiv/g for NC5 ("NCx" denotes a Nafion/sb-CD composite membrane containing x wt% of sb-CD). The cluster-correlation peaks and ionic cluster size distributions of the water-swollen membranes were determined using small-angle X-ray scattering (SAXS) and 1H nuclear magnetic resonance (NMR) cryoporometry, respectively. The SAXS experiments confirmed that increases in the sb-CD content of the membranes shifted the maximum SAXS peaks to lower angles, indicating an increase in the cluster correlation peak. NMR cryoporometry is based on the theory of the melting point depression, Delta Tm, of a liquid confined within a pore, which is dependent on the pore diameter. The melting point depression was determined by analyzing the variation of the NMR signal intensity with temperature. Our analysis of the intensity-temperature (IT) curves showed that the ionic cluster size distribution gradually became broader with increases in the membrane sb-CD content due to the increased water content, indicating an increase in the ionic cluster size. This result indicates that the presence of sb-CD with its many sulfonic acid sites in the Nafion membranes results in increases in the ionic cluster size as well as in the water uptake and the IEC. We conclude that NMR cryoporometry provides a method for determining the ionic cluster size on the nanometer scale in an aqueous environment, which cannot be obtained using other methods.

  15. Effective capacity of multiple antenna channels: Correlation and keyhole

    KAUST Repository

    Zhong, Caijun; Ratnarajah, Tharm; Wong, Kaikit; Alouini, Mohamed-Slim

    2012-01-01

    In this study, the authors derive the effective capacity limits for multiple antenna channels which quantify the maximum achievable rate with consideration of link-layer delay-bound violation probability. Both correlated multiple-input single

  16. Capacity Expansion and Reliability Evaluation on the Networks Flows with Continuous Stochastic Functional Capacity

    Directory of Open Access Journals (Sweden)

    F. Hamzezadeh

    2014-01-01

    Full Text Available In many systems such as computer network, fuel distribution, and transportation system, it is necessary to change the capacity of some arcs in order to increase maximum flow value from source s to sink t, while the capacity change incurs minimum cost. In real-time networks, some factors cause loss of arc’s flow. For example, in some flow distribution systems, evaporation, erosion or sediment in pipes waste the flow. Here we define a real capacity, or the so-called functional capacity, which is the operational capacity of an arc. In other words, the functional capacity of an arc equals the possible maximum flow that may pass through the arc. Increasing the functional arcs capacities incurs some cost. There is a certain resource available to cover the costs. First, we construct a mathematical model to minimize the total cost of expanding the functional capacities to the required levels. Then, we consider the loss of flow on each arc as a stochastic variable and compute the system reliability.

  17. Maximum permissible voltage of YBCO coated conductors

    Energy Technology Data Exchange (ETDEWEB)

    Wen, J.; Lin, B.; Sheng, J.; Xu, J.; Jin, Z. [Department of Electrical Engineering, Shanghai Jiao Tong University, Shanghai (China); Hong, Z., E-mail: zhiyong.hong@sjtu.edu.cn [Department of Electrical Engineering, Shanghai Jiao Tong University, Shanghai (China); Wang, D.; Zhou, H.; Shen, X.; Shen, C. [Qingpu Power Supply Company, State Grid Shanghai Municipal Electric Power Company, Shanghai (China)

    2014-06-15

    Highlights: • We examine three kinds of tapes’ maximum permissible voltage. • We examine the relationship between quenching duration and maximum permissible voltage. • Continuous I{sub c} degradations under repetitive quenching where tapes reaching maximum permissible voltage. • The relationship between maximum permissible voltage and resistance, temperature. - Abstract: Superconducting fault current limiter (SFCL) could reduce short circuit currents in electrical power system. One of the most important thing in developing SFCL is to find out the maximum permissible voltage of each limiting element. The maximum permissible voltage is defined as the maximum voltage per unit length at which the YBCO coated conductors (CC) do not suffer from critical current (I{sub c}) degradation or burnout. In this research, the time of quenching process is changed and voltage is raised until the I{sub c} degradation or burnout happens. YBCO coated conductors test in the experiment are from American superconductor (AMSC) and Shanghai Jiao Tong University (SJTU). Along with the quenching duration increasing, the maximum permissible voltage of CC decreases. When quenching duration is 100 ms, the maximum permissible of SJTU CC, 12 mm AMSC CC and 4 mm AMSC CC are 0.72 V/cm, 0.52 V/cm and 1.2 V/cm respectively. Based on the results of samples, the whole length of CCs used in the design of a SFCL can be determined.

  18. REDUCTION CAPACITY OF SALTSTONE AND SALTSTONE COMPONENTS

    Energy Technology Data Exchange (ETDEWEB)

    Roberts, K.; Kaplan, D.

    2009-11-30

    The duration that saltstone retains its ability to immobilize some key radionuclides, such as technetium (Tc), plutonium (Pu), and neptunium (Np), depends on its capacity to maintain a low redox status (or low oxidation state). The reduction capacity is a measure of the mass of reductants present in the saltstone; the reductants are the active ingredients that immobilize Tc, Pu, and Np. Once reductants are exhausted, the saltstone loses its ability to immobilize these radionuclides. The reduction capacity values reported here are based on the Ce(IV)/Fe(II) system. The Portland cement (198 {micro}eq/g) and especially the fly ash (299 {micro}eq/g) had a measurable amount of reduction capacity, but the blast furnace slag (820 {micro}eq/g) not surprisingly accounted for most of the reduction capacity. The blast furnace slag contains ferrous iron and sulfides which are strong reducing and precipitating species for a large number of solids. Three saltstone samples containing 45% slag or one sample containing 90% slag had essentially the same reduction capacity as pure slag. There appears to be some critical concentration between 10% and 45% slag in the Saltstone formulation that is needed to create the maximum reduction capacity. Values from this work supported those previously reported, namely that the reduction capacity of SRS saltstone is about 820 {micro}eq/g; this value is recommended for estimating the longevity that the Saltstone Disposal Facility will retain its ability to immobilize radionuclides.

  19. What Makes Clusters Decline?

    DEFF Research Database (Denmark)

    Østergaard, Christian Richter; Park, Eun Kyung

    2015-01-01

    Most studies on regional clusters focus on identifying factors and processes that make clusters grow. However, sometimes technologies and market conditions suddenly shift, and clusters decline. This paper analyses the process of decline of the wireless communication cluster in Denmark. The longit...... but being quick to withdraw in times of crisis....

  20. Clustering of correlated networks

    OpenAIRE

    Dorogovtsev, S. N.

    2003-01-01

    We obtain the clustering coefficient, the degree-dependent local clustering, and the mean clustering of networks with arbitrary correlations between the degrees of the nearest-neighbor vertices. The resulting formulas allow one to determine the nature of the clustering of a network.

  1. Performance of heat engines with non-zero heat capacity

    International Nuclear Information System (INIS)

    Odes, Ron; Kribus, Abraham

    2013-01-01

    Highlights: ► Finite heat capacity is a second irreversibility mechanism in addition to thermal resistance. ► Heat capacity introduces thermal transients and reverse heat flow. ► Engine maximum power and efficiency are lower for finite heat capacity. ► Implementing the optimal engine cycle requires active control. - Abstract: The performance of a heat engine is analyzed subject to two types of irreversibility: a non-zero heat capacity, together with the more common finite heat transfer rate between the engine and the external heat reservoirs. The heat capacity represents an engine body that undergoes significant temperature variations during the engine cycle. An option to cut off the heat exchange between the engine and the external surrounding for part of the engine cycle is also explored. A variational approach was taken to find the engine’s internal temperature profile (which defines the internal thermodynamic cycle) that would produce maximum power. The maximum power is shown to be lower than the case of zero heat capacity, due to a loss of heat that is stored in the engine body and then lost, bypassing the thermodynamic cycle. The maximum efficiency and the efficiency at maximum power are also lower than the zero heat capacity case. Similar to the Curzon–Ahlborn analysis, power can be traded for increased efficiency, but for high heat capacity, the range of efficiency that is available for such a trade is diminished. Isolating the engine during part of the cycle reduces maximum power, but the efficiency at maximum power and the maximum efficiency are improved, due to better exploitation of heat stored in the engine body. This might be useful for real engines that are limited by the internal energy change during a single engine cycle or by the operating frequency, leading to a broader power–efficiency curve.

  2. Bioinformatics and Astrophysics Cluster (BinAc)

    Science.gov (United States)

    Krüger, Jens; Lutz, Volker; Bartusch, Felix; Dilling, Werner; Gorska, Anna; Schäfer, Christoph; Walter, Thomas

    2017-09-01

    BinAC provides central high performance computing capacities for bioinformaticians and astrophysicists from the state of Baden-Württemberg. The bwForCluster BinAC is part of the implementation concept for scientific computing for the universities in Baden-Württemberg. Community specific support is offered through the bwHPC-C5 project.

  3. Relevant Subspace Clustering

    DEFF Research Database (Denmark)

    Müller, Emmanuel; Assent, Ira; Günnemann, Stephan

    2009-01-01

    Subspace clustering aims at detecting clusters in any subspace projection of a high dimensional space. As the number of possible subspace projections is exponential in the number of dimensions, the result is often tremendously large. Recent approaches fail to reduce results to relevant subspace...... clusters. Their results are typically highly redundant, i.e. many clusters are detected multiple times in several projections. In this work, we propose a novel model for relevant subspace clustering (RESCU). We present a global optimization which detects the most interesting non-redundant subspace clusters...... achieves top clustering quality while competing approaches show greatly varying performance....

  4. Cluster ion beam facilities

    International Nuclear Information System (INIS)

    Popok, V.N.; Prasalovich, S.V.; Odzhaev, V.B.; Campbell, E.E.B.

    2001-01-01

    A brief state-of-the-art review in the field of cluster-surface interactions is presented. Ionised cluster beams could become a powerful and versatile tool for the modification and processing of surfaces as an alternative to ion implantation and ion assisted deposition. The main effects of cluster-surface collisions and possible applications of cluster ion beams are discussed. The outlooks of the Cluster Implantation and Deposition Apparatus (CIDA) being developed in Guteborg University are shown

  5. Manipulation of Microbubble Clusters Using Focused Ultrasound

    Science.gov (United States)

    Matsuzaki, Hironobu; Osaki, Taichi; Kawaguchi, Kei; Unga, Johan; Ichiyanagi, Mitsuhisa; Azuma, Takashi; Suzuki, Ryo; Maruyama, Kazuo; Takagi, Shu

    2017-11-01

    In recent years, microbubbles (MBs) are expected to be utilized for the ultrasound drug delivery system (DDS). For the MB-DDS, it is important to establish a method of controlling bubbles and bubble clusters using ultrasound field. The objective of this study is to clarify behaviors of bubble clusters with various physical conditions. MBs in the ultrasound field are subjected to the primary Bjerknes force. The force traps MBs at the focal region of the focused ultrasound field. The trapped MBs form a bubble cluster at the region. A bubble cluster continues growing with absorbing surrounding bubbles until it reaches a maximum size beyond which it disappears from the focal region. In the present study, two kinds of MBs are used for the experiment. One is Sonazoid with average diameter of 2.6 um and resonant frequency of 5 MHz. The other is developed by Teikyo Univ., with average diameter of 1.5 um and presumed resonant frequency of 4 MHz. The bubble cluster's behaviors are analyzed using the high-speed camera. Sonazoid clusters have larger critical size than the other in every frequency, and its cluster size is inversely proportional to the ultrasound frequency, while Teikyo-bubble clusters have different tendency. These results are discussed in the presentation.

  6. Revealing the Maximum Strength in Nanotwinned Copper

    DEFF Research Database (Denmark)

    Lu, L.; Chen, X.; Huang, Xiaoxu

    2009-01-01

    boundary–related processes. We investigated the maximum strength of nanotwinned copper samples with different twin thicknesses. We found that the strength increases with decreasing twin thickness, reaching a maximum at 15 nanometers, followed by a softening at smaller values that is accompanied by enhanced...

  7. Modelling maximum canopy conductance and transpiration in ...

    African Journals Online (AJOL)

    There is much current interest in predicting the maximum amount of water that can be transpired by Eucalyptus trees. It is possible that industrial waste water may be applied as irrigation water to eucalypts and it is important to predict the maximum transpiration rates of these plantations in an attempt to dispose of this ...

  8. PREFACE: Nuclear Cluster Conference; Cluster'07

    Science.gov (United States)

    Freer, Martin

    2008-05-01

    The Cluster Conference is a long-running conference series dating back to the 1960's, the first being initiated by Wildermuth in Bochum, Germany, in 1969. The most recent meeting was held in Nara, Japan, in 2003, and in 2007 the 9th Cluster Conference was held in Stratford-upon-Avon, UK. As the name suggests the town of Stratford lies upon the River Avon, and shortly before the conference, due to unprecedented rainfall in the area (approximately 10 cm within half a day), lay in the River Avon! Stratford is the birthplace of the `Bard of Avon' William Shakespeare, and this formed an intriguing conference backdrop. The meeting was attended by some 90 delegates and the programme contained 65 70 oral presentations, and was opened by a historical perspective presented by Professor Brink (Oxford) and closed by Professor Horiuchi (RCNP) with an overview of the conference and future perspectives. In between, the conference covered aspects of clustering in exotic nuclei (both neutron and proton-rich), molecular structures in which valence neutrons are exchanged between cluster cores, condensates in nuclei, neutron-clusters, superheavy nuclei, clusters in nuclear astrophysical processes and exotic cluster decays such as 2p and ternary cluster decay. The field of nuclear clustering has become strongly influenced by the physics of radioactive beam facilities (reflected in the programme), and by the excitement that clustering may have an important impact on the structure of nuclei at the neutron drip-line. It was clear that since Nara the field had progressed substantially and that new themes had emerged and others had crystallized. Two particular topics resonated strongly condensates and nuclear molecules. These topics are thus likely to be central in the next cluster conference which will be held in 2011 in the Hungarian city of Debrechen. Martin Freer Participants and Cluster'07

  9. Capacity planning and management

    OpenAIRE

    Boydell, Briony

    2011-01-01

    After reading this chapter you should be able to: • Define and measure capacity and appreciate the factors that influence it. • Assess the difficulties of matching capacity to demand. • Evaluate and apply the different strategies for matching capacity with demand in the short, medium and long term. • Analyse the impact of constraints and bottlenecks on a process and consider the Theory of Constraints. • Outline the different strategies available for both manufacturing and service operations. ...

  10. The French capacity mechanism

    International Nuclear Information System (INIS)

    2014-01-01

    The French capacity mechanism has been design to ensure security of supply in the context of the energy transition. This energy transition challenges the electricity market design with several features: peak load growth, the development of renewables, demand response,... To ensure security of supply in this context, a capacity mechanism is being implemented in France. It is a market wide capacity obligation on electricity suppliers, based on market principles. Suppliers are responsible for forecasting their obligation, which corresponds to their contribution to winter peak load, and must procure enough capacity certificates to meet their obligations. Capacity certificates are granted to capacities through a certification process, which assesses their contribution to security of supply on the basis of availability commitments. This certification process is technology neutral and performance based, associated with controls and penalties in case of non compliance. Demand Side is fully integrated in the market, either through the reduction of suppliers' capacity obligation or direct participation after certification. In addition to the expected benefits in terms of security of supply, the French capacity market will foster the development of demand response. The participation of foreign capacities will require adaptations which are scheduled in a road-map, and could pave the way for further European integration of energy policies. (authors)

  11. Lung function profiles and aerobic capacity of adult cigarette and ...

    African Journals Online (AJOL)

    Lung function profiles and aerobic capacity of adult cigarette and hookah smokers after 12 weeks intermittent training. ... All subjects performed 30 min of interval exercise (2 min of work followed by 1 min of rest) three times a week for 12 weeks at an intensity estimated at 70% of the subject's maximum aerobic capacity ...

  12. Light dependence of carboxylation capacity for C3 photosynthesis models

    Science.gov (United States)

    Photosynthesis at high light is often modelled by assuming limitation by the maximum capacity of Rubisco carboxylation at low carbon dioxide concentrations, by electron transport capacity at higher concentrations, and sometimes by triose-phosphate utilization rate at the highest concentrations. Pho...

  13. Carrying capacity of Chaetoceros gracilis in Homa Lagoon and the ...

    African Journals Online (AJOL)

    The possibility for nutrient limitation to affect C. gracilis was assessed from two different ecosystems (Izmir Bay and Homa Lagoon). Our goal was to determine the growth rate of all nutrients and the maximum levels of the C. gracilis phytoplankton biomass (the maximum biomass carrying capacity) on the extent of its full ...

  14. Management of cluster headache

    DEFF Research Database (Denmark)

    Tfelt-Hansen, Peer C; Jensen, Rigmor H

    2012-01-01

    The prevalence of cluster headache is 0.1% and cluster headache is often not diagnosed or misdiagnosed as migraine or sinusitis. In cluster headache there is often a considerable diagnostic delay - an average of 7 years in a population-based survey. Cluster headache is characterized by very severe...... or severe orbital or periorbital pain with a duration of 15-180 minutes. The cluster headache attacks are accompanied by characteristic associated unilateral symptoms such as tearing, nasal congestion and/or rhinorrhoea, eyelid oedema, miosis and/or ptosis. In addition, there is a sense of restlessness...... and agitation. Patients may have up to eight attacks per day. Episodic cluster headache (ECH) occurs in clusters of weeks to months duration, whereas chronic cluster headache (CCH) attacks occur for more than 1 year without remissions. Management of cluster headache is divided into acute attack treatment...

  15. Symmetries of cluster configurations

    International Nuclear Information System (INIS)

    Kramer, P.

    1975-01-01

    A deeper understanding of clustering phenomena in nuclei must encompass at least two interrelated aspects of the subject: (A) Given a system of A nucleons with two-body interactions, what are the relevant and persistent modes of clustering involved. What is the nature of the correlated nucleon groups which form the clusters, and what is their mutual interaction. (B) Given the cluster modes and their interaction, what systematic patterns of nuclear structure and reactions emerge from it. Are there, for example, families of states which share the same ''cluster parents''. Which cluster modes are compatible or exclude each other. What quantum numbers could characterize cluster configurations. There is no doubt that we can learn a good deal from the experimentalists who have discovered many of the features relevant to aspect (B). Symmetries specific to cluster configurations which can throw some light on both aspects of clustering are discussed

  16. Heat capacity characterization at phase transition temperature of Agl superionic

    International Nuclear Information System (INIS)

    Widowati, Arie

    2000-01-01

    The phase transition of Agl superionic conductor was investigated by calorometric. A single phase transition was found at (153±5) o C which corresponds to the α - β transition. Calorimetric measurement showed an anomalously high heat capacity with a large discontinues change in the Arrhenius plot, was found above the transition temperature of β - α phase. The maximum heat capacity was found to be ±19.7 cal/gmol. Key words : superionic conductor, thermal capacity

  17. Heat Capacity Analysis Report

    International Nuclear Information System (INIS)

    Findikakis, A.

    2004-01-01

    The purpose of this report is to provide heat capacity values for the host and surrounding rock layers for the waste repository at Yucca Mountain. The heat capacity representations provided by this analysis are used in unsaturated zone (UZ) flow, transport, and coupled processes numerical modeling activities, and in thermal analyses as part of the design of the repository to support the license application. Among the reports that use the heat capacity values estimated in this report are the ''Multiscale Thermohydrologic Model'' report, the ''Drift Degradation Analysis'' report, the ''Ventilation Model and Analysis Report, the Igneous Intrusion Impacts on Waste Packages and Waste Forms'' report, the ''Dike/Drift Interactions report, the Drift-Scale Coupled Processes (DST and TH Seepage) Models'' report, and the ''In-Drift Natural Convection and Condensation'' report. The specific objective of this study is to determine the rock-grain and rock-mass heat capacities for the geologic stratigraphy identified in the ''Mineralogic Model (MM3.0) Report'' (BSC 2004 [DIRS 170031], Table 1-1). This report provides estimates of the heat capacity for all stratigraphic layers except the Paleozoic, for which the mineralogic abundance data required to estimate the heat capacity are not available. The temperature range of interest in this analysis is 25 C to 325 C. This interval is broken into three separate temperature sub-intervals: 25 C to 95 C, 95 C to 114 C, and 114 C to 325 C, which correspond to the preboiling, trans-boiling, and postboiling regimes. Heat capacity is defined as the amount of energy required to raise the temperature of a unit mass of material by one degree (Nimick and Connolly 1991 [DIRS 100690], p. 5). The rock-grain heat capacity is defined as the heat capacity of the rock solids (minerals), and does not include the effect of water that exists in the rock pores. By comparison, the rock-mass heat capacity considers the heat capacity of both solids and pore

  18. Heat capacity and solid solubility of iron in scandium

    International Nuclear Information System (INIS)

    Tsang, T.-W.E.

    1981-01-01

    The maximum solid solubility of iron in scandium was determined to be between 50 and 85 at.ppm in the as-cast condition. As the concentration of iron increases, it segregates along the grain boundary, as is evident from optical metallography and electron microprobe examinations. Annealing also causes the iron dissolved in scandium to separate out and cluster along the grain boundary. Heat capacity measurements show an anomaly in the C/T versus T 2 plots for iron concentrations of 19 at.ppm or greater. For iron dissolved in solid scandium the excess entropy due to the iron impurity is in agreement with the theoretical prediction of ck ln(2S + 1) for an impurity-conduction electron (Kondo) interaction, but is 4 - 8 times larger than the theoretical prediction when iron segregates along the grain boundary. Furthermore, our results suggest that most of the previously reported low temperature physical properties of scandium are probably in error because of either iron impurity-conduction electron interactions or Fe-Fe interactions in the precipitated second-phase Sc-Fe compound. (Auth.)

  19. Benefits of the maximum tolerated dose (MTD) and maximum tolerated concentration (MTC) concept in aquatic toxicology

    International Nuclear Information System (INIS)

    Hutchinson, Thomas H.; Boegi, Christian; Winter, Matthew J.; Owens, J. Willie

    2009-01-01

    There is increasing recognition of the need to identify specific sublethal effects of chemicals, such as reproductive toxicity, and specific modes of actions of the chemicals, such as interference with the endocrine system. To achieve these aims requires criteria which provide a basis to interpret study findings so as to separate these specific toxicities and modes of action from not only acute lethality per se but also from severe inanition and malaise that non-specifically compromise reproductive capacity and the response of endocrine endpoints. Mammalian toxicologists have recognized that very high dose levels are sometimes required to elicit both specific adverse effects and present the potential of non-specific 'systemic toxicity'. Mammalian toxicologists have developed the concept of a maximum tolerated dose (MTD) beyond which a specific toxicity or action cannot be attributed to a test substance due to the compromised state of the organism. Ecotoxicologists are now confronted by a similar challenge and must develop an analogous concept of a MTD and the respective criteria. As examples of this conundrum, we note recent developments in efforts to validate protocols for fish reproductive toxicity and endocrine screens (e.g. some chemicals originally selected as 'negatives' elicited decreases in fecundity or changes in endpoints intended to be biomarkers for endocrine modes of action). Unless analogous criteria can be developed, the potentially confounding effects of systemic toxicity may then undermine the reliable assessment of specific reproductive effects or biomarkers such as vitellogenin or spiggin. The same issue confronts other areas of aquatic toxicology (e.g., genotoxicity) and the use of aquatic animals for preclinical assessments of drugs (e.g., use of zebrafish for drug safety assessment). We propose that there are benefits to adopting the concept of an MTD for toxicology and pharmacology studies using fish and other aquatic organisms and the

  20. MXLKID: a maximum likelihood parameter identifier

    International Nuclear Information System (INIS)

    Gavel, D.T.

    1980-07-01

    MXLKID (MaXimum LiKelihood IDentifier) is a computer program designed to identify unknown parameters in a nonlinear dynamic system. Using noisy measurement data from the system, the maximum likelihood identifier computes a likelihood function (LF). Identification of system parameters is accomplished by maximizing the LF with respect to the parameters. The main body of this report briefly summarizes the maximum likelihood technique and gives instructions and examples for running the MXLKID program. MXLKID is implemented LRLTRAN on the CDC7600 computer at LLNL. A detailed mathematical description of the algorithm is given in the appendices. 24 figures, 6 tables

  1. Cluster Decline and Resilience

    DEFF Research Database (Denmark)

    Østergaard, Christian Richter; Park, Eun Kyung

    Most studies on regional clusters focus on identifying factors and processes that make clusters grow. However, sometimes technologies and market conditions suddenly shift, and clusters decline. This paper analyses the process of decline of the wireless communication cluster in Denmark, 1963......-2011. Our longitudinal study reveals that technological lock-in and exit of key firms have contributed to impairment of the cluster’s resilience in adapting to disruptions. Entrepreneurship has a positive effect on cluster resilience, while multinational companies have contradicting effects by bringing...... in new resources to the cluster but being quick to withdraw in times of crisis....

  2. Prospects for OPEC capacity

    International Nuclear Information System (INIS)

    Adelman, M.A.

    1995-01-01

    OPEC capacity is not exogenous, but responds to demand. Price increases have not been caused by capacity shortages. OPEC nations find it hard to set aside even very small portions of their revenues for oil investment, despite its extreme profitability. Foreign investors face high risks. Production sharing makes their after-tax return even more unstable. (author)

  3. Uncertainty in adaptive capacity

    International Nuclear Information System (INIS)

    Neil Adger, W.; Vincent, K.

    2005-01-01

    The capacity to adapt is a critical element of the process of adaptation: it is the vector of resources that represent the asset base from which adaptation actions can be made. Adaptive capacity can in theory be identified and measured at various scales, from the individual to the nation. The assessment of uncertainty within such measures comes from the contested knowledge domain and theories surrounding the nature of the determinants of adaptive capacity and the human action of adaptation. While generic adaptive capacity at the national level, for example, is often postulated as being dependent on health, governance and political rights, and literacy, and economic well-being, the determinants of these variables at national levels are not widely understood. We outline the nature of this uncertainty for the major elements of adaptive capacity and illustrate these issues with the example of a social vulnerability index for countries in Africa. (authors)

  4. Comprehensive cluster analysis with Transitivity Clustering.

    Science.gov (United States)

    Wittkop, Tobias; Emig, Dorothea; Truss, Anke; Albrecht, Mario; Böcker, Sebastian; Baumbach, Jan

    2011-03-01

    Transitivity Clustering is a method for the partitioning of biological data into groups of similar objects, such as genes, for instance. It provides integrated access to various functions addressing each step of a typical cluster analysis. To facilitate this, Transitivity Clustering is accessible online and offers three user-friendly interfaces: a powerful stand-alone version, a web interface, and a collection of Cytoscape plug-ins. In this paper, we describe three major workflows: (i) protein (super)family detection with Cytoscape, (ii) protein homology detection with incomplete gold standards and (iii) clustering of gene expression data. This protocol guides the user through the most important features of Transitivity Clustering and takes ∼1 h to complete.

  5. Maximum neutron flux in thermal reactors

    International Nuclear Information System (INIS)

    Strugar, P.V.

    1968-12-01

    Direct approach to the problem is to calculate spatial distribution of fuel concentration if the reactor core directly using the condition of maximum neutron flux and comply with thermal limitations. This paper proved that the problem can be solved by applying the variational calculus, i.e. by using the maximum principle of Pontryagin. Mathematical model of reactor core is based on the two-group neutron diffusion theory with some simplifications which make it appropriate from maximum principle point of view. Here applied theory of maximum principle are suitable for application. The solution of optimum distribution of fuel concentration in the reactor core is obtained in explicit analytical form. The reactor critical dimensions are roots of a system of nonlinear equations and verification of optimum conditions can be done only for specific examples

  6. Maximum allowable load on wheeled mobile manipulators

    International Nuclear Information System (INIS)

    Habibnejad Korayem, M.; Ghariblu, H.

    2003-01-01

    This paper develops a computational technique for finding the maximum allowable load of mobile manipulator during a given trajectory. The maximum allowable loads which can be achieved by a mobile manipulator during a given trajectory are limited by the number of factors; probably the dynamic properties of mobile base and mounted manipulator, their actuator limitations and additional constraints applied to resolving the redundancy are the most important factors. To resolve extra D.O.F introduced by the base mobility, additional constraint functions are proposed directly in the task space of mobile manipulator. Finally, in two numerical examples involving a two-link planar manipulator mounted on a differentially driven mobile base, application of the method to determining maximum allowable load is verified. The simulation results demonstrates the maximum allowable load on a desired trajectory has not a unique value and directly depends on the additional constraint functions which applies to resolve the motion redundancy

  7. Maximum phytoplankton concentrations in the sea

    DEFF Research Database (Denmark)

    Jackson, G.A.; Kiørboe, Thomas

    2008-01-01

    A simplification of plankton dynamics using coagulation theory provides predictions of the maximum algal concentration sustainable in aquatic systems. These predictions have previously been tested successfully against results from iron fertilization experiments. We extend the test to data collect...

  8. Estimation of Maximum Allowable PV Connection to LV Residential Power Networks

    DEFF Research Database (Denmark)

    Demirok, Erhan; Sera, Dezso; Teodorescu, Remus

    2011-01-01

    Maximum photovoltaic (PV) hosting capacity of low voltage (LV) power networks is mainly restricted by either thermal limits of network components or grid voltage quality resulted from high penetration of distributed PV systems. This maximum hosting capacity may be lower than the available solar...... potential of geographic area due to power network limitations even though all rooftops are fully occupied with PV modules. Therefore, it becomes more of an issue to know what exactly limits higher PV penetration level and which solutions should be engaged efficiently such as over sizing distribution...

  9. A polynomial time algorithm for solving the maximum flow problem in directed networks

    International Nuclear Information System (INIS)

    Tlas, M.

    2015-01-01

    An efficient polynomial time algorithm for solving maximum flow problems has been proposed in this paper. The algorithm is basically based on the binary representation of capacities; it solves the maximum flow problem as a sequence of O(m) shortest path problems on residual networks with nodes and m arcs. It runs in O(m"2r) time, where is the smallest integer greater than or equal to log B , and B is the largest arc capacity of the network. A numerical example has been illustrated using this proposed algorithm.(author)

  10. Evaluating Maximum Photovoltaic Integration in District Distribution Systems Considering Optimal Inverter Dispatch and Cloud Shading Conditions

    DEFF Research Database (Denmark)

    Ding, Tao; Kou, Yu; Yang, Yongheng

    2017-01-01

    . However, the intermittency of solar PV energy (e.g., due to passing clouds) may affect the PV generation in the district distribution network. To address this issue, the voltage magnitude constraints under the cloud shading conditions should be taken into account in the optimization model, which can......As photovoltaic (PV) integration increases in distribution systems, to investigate the maximum allowable PV integration capacity for a district distribution system becomes necessary in the planning phase, an optimization model is thus proposed to evaluate the maximum PV integration capacity while...

  11. Maximum-Likelihood Detection Of Noncoherent CPM

    Science.gov (United States)

    Divsalar, Dariush; Simon, Marvin K.

    1993-01-01

    Simplified detectors proposed for use in maximum-likelihood-sequence detection of symbols in alphabet of size M transmitted by uncoded, full-response continuous phase modulation over radio channel with additive white Gaussian noise. Structures of receivers derived from particular interpretation of maximum-likelihood metrics. Receivers include front ends, structures of which depends only on M, analogous to those in receivers of coherent CPM. Parts of receivers following front ends have structures, complexity of which would depend on N.

  12. LMC clusters: young

    International Nuclear Information System (INIS)

    Freeman, K.C.

    1980-01-01

    The young globular clusters of the LMC have ages of 10 7 -10 8 y. Their masses and structure are similar to those of the smaller galactic globular clusters. Their stellar mass functions (in the mass range 6 solar masses to 1.2 solar masses) vary greatly from cluster to cluster, although the clusters are similar in total mass, age, structure and chemical composition. It would be very interesting to know why these clusters are forming now in the LMC and not in the Galaxy. The author considers the 'young globular' or 'blue populous' clusters of the LMC. The ages of these objects are 10 7 to 10 8 y, and their masses are 10 4 to 10 5 solar masses, so they are populous enough to be really useful for studying the evolution of massive stars. The author concentrates on the structure and stellar content of these young clusters. (Auth.)

  13. Star clusters and associations

    International Nuclear Information System (INIS)

    Ruprecht, J.; Palous, J.

    1983-01-01

    All 33 papers presented at the symposium were inputted to INIS. They dealt with open clusters, globular clusters, stellar associations and moving groups, and local kinematics and galactic structures. (E.S.)

  14. Cluster beam injection

    International Nuclear Information System (INIS)

    Bottiglioni, F.; Coutant, J.; Fois, M.

    1978-01-01

    Areas of possible applications of cluster injection are discussed. The deposition inside the plasma of molecules, issued from the dissociation of the injected clusters, has been computed. Some empirical scaling laws for the penetration are given

  15. Interstellar Extinction in 20 Open Star Clusters

    Science.gov (United States)

    Rangwal, Geeta; Yadav, R. K. S.; Durgapal, Alok K.; Bisht, D.

    2017-12-01

    The interstellar extinction law in 20 open star clusters namely, Berkeley 7, Collinder 69, Hogg 10, NGC 2362, Czernik 43, NGC 6530, NGC 6871, Bochum 10, Haffner 18, IC 4996, NGC 2384, NGC 6193, NGC 6618, NGC 7160, Collinder 232, Haffner 19, NGC 2401, NGC 6231, NGC 6823, and NGC 7380 have been studied in the optical and near-IR wavelength ranges. The difference between maximum and minimum values of E(B - V) indicates the presence of non-uniform extinction in all the clusters except Collinder 69, NGC 2362, and NGC 2384. The colour excess ratios are consistent with a normal extinction law for the clusters NGC 6823, Haffner 18, Haffner 19, NGC 7160, NGC 6193, NGC 2401, NGC 2384, NGC 6871, NGC 7380, Berkeley 7, Collinder 69, and IC 4996. We have found that the differential colour-excess ΔE(B - V), which may be due to the occurrence of dust and gas inside the clusters, decreases with the age of the clusters. A spatial variation of colour excess is found in NGC 6193 in the sense that it decreases from east to west in the cluster region. For the clusters Berkeley 7, NGC 7380, and NGC 6871, a dependence of colour excess E(B - V) with spectral class and luminosity is observed. Eight stars in Collinder 232, four stars in NGC 6530, and one star in NGC 6231 have excess flux in near-IR. This indicates that these stars may have circumstellar material around them.

  16. Pullout capacity of batter pile in sand.

    Science.gov (United States)

    Nazir, Ashraf; Nasr, Ahmed

    2013-03-01

    Many offshore structures are subjected to overturning moments due to wind load, wave pressure, and ship impacts. Also most of retaining walls are subjected to horizontal forces and bending moments, these forces are due to earth pressure. For foundations in such structures, usually a combination of vertical and batter piles is used. Little information is available in the literature about estimating the capacity of piles under uplift. In cases where these supporting piles are not vertical, the behavior under axial pullout is not well established. In order to delineate the significant variables affecting the ultimate uplift shaft resistance of batter pile in dry sand, a testing program comprising 62 pullout tests was conducted. The tests are conducted on model steel pile installed in loose, medium, and dense sand to an embedded depth ratio, L/d, vary from 7.5 to 30 and with various batter angles of 0°, 10°, 20°, and 30°. Results indicate that the pullout capacity of a batter pile constructed in dense and/or medium density sand increases with the increase of batter angle attains maximum value and then decreases, the maximum value of Pα occurs at batter angle approximately equal to 20°, and it is about 21-31% more than the vertical pile capacity, while the pullout capacity for batter pile that constructed in loose sand decreases with the increase of pile inclination. The results also indicated that the circular pile is more resistant to pullout forces than the square and rectangular pile shape. The rough model piles tested is experienced 18-75% increase in capacity compared with the smooth model piles. The suggested relations for the pullout capacity of batter pile regarding the vertical pile capacity are well predicted.

  17. Pullout capacity of batter pile in sand

    Directory of Open Access Journals (Sweden)

    Ashraf Nazir

    2013-03-01

    Full Text Available Many offshore structures are subjected to overturning moments due to wind load, wave pressure, and ship impacts. Also most of retaining walls are subjected to horizontal forces and bending moments, these forces are due to earth pressure. For foundations in such structures, usually a combination of vertical and batter piles is used. Little information is available in the literature about estimating the capacity of piles under uplift. In cases where these supporting piles are not vertical, the behavior under axial pullout is not well established. In order to delineate the significant variables affecting the ultimate uplift shaft resistance of batter pile in dry sand, a testing program comprising 62 pullout tests was conducted. The tests are conducted on model steel pile installed in loose, medium, and dense sand to an embedded depth ratio, L/d, vary from 7.5 to 30 and with various batter angles of 0°, 10°, 20°, and 30°. Results indicate that the pullout capacity of a batter pile constructed in dense and/or medium density sand increases with the increase of batter angle attains maximum value and then decreases, the maximum value of Pα occurs at batter angle approximately equal to 20°, and it is about 21–31% more than the vertical pile capacity, while the pullout capacity for batter pile that constructed in loose sand decreases with the increase of pile inclination. The results also indicated that the circular pile is more resistant to pullout forces than the square and rectangular pile shape. The rough model piles tested is experienced 18–75% increase in capacity compared with the smooth model piles. The suggested relations for the pullout capacity of batter pile regarding the vertical pile capacity are well predicted.

  18. OPEC future capacity expansions

    International Nuclear Information System (INIS)

    Sandrea, I.

    2005-01-01

    This conference presentation examined OPEC future capacity expansions including highlights from 2000-2004 from the supply perspective and actions by OPEC; OPEC spare capacity in 2005/2006; medium-term capacity expansion and investments; long-term scenarios, challenges and opportunities; and upstream policies in member countries. Highlights from the supply perspective included worst than expected non-OPEC supply response; non-OPEC supply affected by a number of accidents and strikes; geopolitical tensions; and higher than expected demand for OPEC crude. OPEC's actions included closer relationship with other producers and consumers; capacity expansions in 2004 and 2005/2006; and OPEC kept the market well supplied with crude in 2004. The presentation also provided data using graphical charts on OPEC net capacity additions until 2005/2006; OPEC production versus spare capacity from 2003 to 2005; OPEC production and capacity to 2010; and change in required OPEC production from 2005-2020. Medium term expansion to 2010 includes over 60 projects. Medium-term risks such as project execution, financing, costs, demand, reserves, depletion, integration of Iraq, and geopolitical tensions were also discussed. The presentation concluded that in the long term, large uncertainties remain; the peak of world supply is not imminent; and continued and enhanced cooperation is essential to market stability. tabs., figs

  19. An alternative methodological approach to value analysis of regions, municipal corporations and clusters

    Directory of Open Access Journals (Sweden)

    Mojmír Sabolovič

    2011-01-01

    Full Text Available The paper deals with theoretical conception of value analysis of regions, municipal corporations and clusters. The subject of this paper is heterodox approach to sensitivity analysis of finite set of variables based on non-additive measure. For dynamic analysis of trajectory of general value are sufficient robust models based on maximum entropy principle. Findings concern explanation of proper fuzzy integral – Choquet integral. The fuzzy measure is represented by theory of capacities (Choquet, 1953 on powerset. In fine, the conception of the New integral for capacities (Lehler, 2005 is discussed. Value analysis and transmission constitutes remarkable aspect of performance evaluation of regions, municipal corporations and clusters. In the light of high ratio of soft variables, social behavior, intangible assets and human capital within those types of subjects the fuzzy integral introduce useful tool for modeling. The New integral afterwards concerns considerable characteristic of people behavior – risk averse articulated concave function and non-additive operator. Results comprehended tools enabling observation of synergy, redundancy and inhibition of value variables as consequence of non-additive measure. In fine, results induced issues for future research.

  20. Clustering at high redshifts

    International Nuclear Information System (INIS)

    Shaver, P.A.

    1986-01-01

    Evidence for clustering of and with high-redshift QSOs is discussed. QSOs of different redshifts show no clustering, but QSOs of similar redshifts appear to be clustered on a scale comparable to that of galaxies at the present epoch. In addition, spectroscopic studies of close pairs of QSOs indicate that QSOs are surrounded by a relatively high density of absorbing matter, possibly clusters of galaxies

  1. FlowMax: A Computational Tool for Maximum Likelihood Deconvolution of CFSE Time Courses.

    Directory of Open Access Journals (Sweden)

    Maxim Nikolaievich Shokhirev

    Full Text Available The immune response is a concerted dynamic multi-cellular process. Upon infection, the dynamics of lymphocyte populations are an aggregate of molecular processes that determine the activation, division, and longevity of individual cells. The timing of these single-cell processes is remarkably widely distributed with some cells undergoing their third division while others undergo their first. High cell-to-cell variability and technical noise pose challenges for interpreting popular dye-dilution experiments objectively. It remains an unresolved challenge to avoid under- or over-interpretation of such data when phenotyping gene-targeted mouse models or patient samples. Here we develop and characterize a computational methodology to parameterize a cell population model in the context of noisy dye-dilution data. To enable objective interpretation of model fits, our method estimates fit sensitivity and redundancy by stochastically sampling the solution landscape, calculating parameter sensitivities, and clustering to determine the maximum-likelihood solution ranges. Our methodology accounts for both technical and biological variability by using a cell fluorescence model as an adaptor during population model fitting, resulting in improved fit accuracy without the need for ad hoc objective functions. We have incorporated our methodology into an integrated phenotyping tool, FlowMax, and used it to analyze B cells from two NFκB knockout mice with distinct phenotypes; we not only confirm previously published findings at a fraction of the expended effort and cost, but reveal a novel phenotype of nfkb1/p105/50 in limiting the proliferative capacity of B cells following B-cell receptor stimulation. In addition to complementing experimental work, FlowMax is suitable for high throughput analysis of dye dilution studies within clinical and pharmacological screens with objective and quantitative conclusions.

  2. Cluster Physics with Merging Galaxy Clusters

    Directory of Open Access Journals (Sweden)

    Sandor M. Molnar

    2016-02-01

    Full Text Available Collisions between galaxy clusters provide a unique opportunity to study matter in a parameter space which cannot be explored in our laboratories on Earth. In the standard LCDM model, where the total density is dominated by the cosmological constant ($Lambda$ and the matter density by cold dark matter (CDM, structure formation is hierarchical, and clusters grow mostly by merging.Mergers of two massive clusters are the most energetic events in the universe after the Big Bang,hence they provide a unique laboratory to study cluster physics.The two main mass components in clusters behave differently during collisions:the dark matter is nearly collisionless, responding only to gravity, while the gas is subject to pressure forces and dissipation, and shocks and turbulenceare developed during collisions. In the present contribution we review the different methods used to derive the physical properties of merging clusters. Different physical processes leave their signatures on different wavelengths, thusour review is based on a multifrequency analysis. In principle, the best way to analyze multifrequency observations of merging clustersis to model them using N-body/HYDRO numerical simulations. We discuss the results of such detailed analyses.New high spatial and spectral resolution ground and space based telescopeswill come online in the near future. Motivated by these new opportunities,we briefly discuss methods which will be feasible in the near future in studying merging clusters.

  3. Simple product-form bounds for queueing networks with finite clusters

    NARCIS (Netherlands)

    van Dijk, N.M.; van der Sluis, E.

    2001-01-01

    Queueing networks are studied with finite capacity constraints for clusters of stations. First, by an instructive tandem cluster example it is shown how a product-form modification method for networks with finite stations can be extended to networks with finite clusters. Next, a general result is

  4. Vedr.: Military capacity building

    DEFF Research Database (Denmark)

    Larsen, Josefine Kühnel; Struwe, Lars Bangert

    2013-01-01

    Military capacity building has increasingly become an integral part of Danish defence. Military capacity is a new way of thinking Danish defence and poses a new set of challenges and opportunities for the Danish military and the Political leadership. On the 12th of december, PhD. Candidate Josefine...... Kühnel Larsen and researcher Lars Bangert Struwe of CMS had organized a seminar in collaboration with Royal Danish Defense Colleg and the East African Security Governance Network. The seminar focused on some of the risks involved in Military capacity building and how these risks are dealt with from...

  5. Size selected metal clusters

    Indian Academy of Sciences (India)

    First page Back Continue Last page Overview Graphics. The Optical Absorption Spectra of Small Silver Clusters (5-11) ... Soft Landing and Fragmentation of Small Clusters Deposited in Noble-Gas Films. Harbich, W.; Fedrigo, S.; Buttet, J. Phys. Rev. B 1998, 58, 7428. CO combustion on supported gold clusters. Arenz M ...

  6. The Durban Auto Cluster

    DEFF Research Database (Denmark)

    Lorentzen, Jochen; Robbins, Glen; Barnes, Justin

    2004-01-01

    The paper describes the formation of the Durban Auto Cluster in the context of trade liberalization. It argues that the improvement of operational competitiveness of firms in the cluster is prominently due to joint action. It tests this proposition by comparing the gains from cluster activities...

  7. Marketing research cluster analysis

    Directory of Open Access Journals (Sweden)

    Marić Nebojša

    2002-01-01

    Full Text Available One area of applications of cluster analysis in marketing is identification of groups of cities and towns with similar demographic profiles. This paper considers main aspects of cluster analysis by an example of clustering 12 cities with the use of Minitab software.

  8. Marketing research cluster analysis

    OpenAIRE

    Marić Nebojša

    2002-01-01

    One area of applications of cluster analysis in marketing is identification of groups of cities and towns with similar demographic profiles. This paper considers main aspects of cluster analysis by an example of clustering 12 cities with the use of Minitab software.

  9. Minimalist's linux cluster

    International Nuclear Information System (INIS)

    Choi, Chang-Yeong; Kim, Jeong-Hyun; Kim, Seyong

    2004-01-01

    Using barebone PC components and NIC's, we construct a linux cluster which has 2-dimensional mesh structure. This cluster has smaller footprint, is less expensive, and use less power compared to conventional linux cluster. Here, we report our experience in building such a machine and discuss our current lattice project on the machine

  10. Range-clustering queries

    NARCIS (Netherlands)

    Abrahamsen, M.; de Berg, M.T.; Buchin, K.A.; Mehr, M.; Mehrabi, A.D.

    2017-01-01

    In a geometric k -clustering problem the goal is to partition a set of points in R d into k subsets such that a certain cost function of the clustering is minimized. We present data structures for orthogonal range-clustering queries on a point set S : given a query box Q and an integer k>2 , compute

  11. Cosmology with cluster surveys

    Indian Academy of Sciences (India)

    Abstract. Surveys of clusters of galaxies provide us with a powerful probe of the den- sity and nature of the dark energy. The red-shift distribution of detected clusters is highly sensitive to the dark energy equation of state parameter w. Upcoming Sunyaev–. Zel'dovich (SZ) surveys would provide us large yields of clusters to ...

  12. Peculiar features of heat capacity for Cu and Ni nanoclusters

    International Nuclear Information System (INIS)

    Gafner, S. L.; Redel, L. V.; Gafner, Yu. Ya.; Samsonov, V. M.

    2011-01-01

    The heat capacity of copper and nickel clusters (from 2 to 6 nm in diameter) was investigated in the temperature range 200–800 K using molecular dynamics method and a modified tight-binding potential. The simulation results demonstrate a very good agreement with the available experimental data at T = 200 K and a fairy good agreement at higher temperatures. A number of regular trends are revealed in computer experiments which agree with the corresponding theoretical predictions. A conclusion is made that in the case of single free clusters the heat capacity may exceed the capacity of the corresponding bulk material. It is found that at 200 K, the copper nanocluster (D = 6 nm) heat capacity is higher by 10% and for nickel cluster by 13%. The difference diminishes with increasing the nanoparticles size proportionally to the relative number of surface atoms. A conclusion is made that very high values of the nanostructure heat capacity observed in laboratory experiments should not be attributed to free clusters, i.e., the effect in question is caused by other reasons.

  13. Significance of flow clustering and sequencing on sediment transport: 1D sediment transport modelling

    Science.gov (United States)

    Hassan, Kazi; Allen, Deonie; Haynes, Heather

    2016-04-01

    This paper considers 1D hydraulic model data on the effect of high flow clusters and sequencing on sediment transport. Using observed flow gauge data from the River Caldew, England, a novel stochastic modelling approach was developed in order to create alternative 50 year flow sequences. Whilst the observed probability density of gauge data was preserved in all sequences, the order in which those flows occurred was varied using the output from a Hidden Markov Model (HMM) with generalised Pareto distribution (GP). In total, one hundred 50 year synthetic flow series were generated and used as the inflow boundary conditions for individual flow series model runs using the 1D sediment transport model HEC-RAS. The model routed graded sediment through the case study river reach to define the long-term morphological changes. Comparison of individual simulations provided a detailed understanding of the sensitivity of channel capacity to flow sequence. Specifically, each 50 year synthetic flow sequence was analysed using a 3-month, 6-month or 12-month rolling window approach and classified for clusters in peak discharge. As a cluster is described as a temporal grouping of flow events above a specified threshold, the threshold condition used herein is considered as a morphologically active channel forming discharge event. Thus, clusters were identified for peak discharges in excess of 10%, 20%, 50%, 100% and 150% of the 1 year Return Period (RP) event. The window of above-peak flows also required cluster definition and was tested for timeframes 1, 2, 10 and 30 days. Subsequently, clusters could be described in terms of the number of events, maximum peak flow discharge, cumulative flow discharge and skewness (i.e. a description of the flow sequence). The model output for each cluster was analysed for the cumulative flow volume and cumulative sediment transport (mass). This was then compared to the total sediment transport of a single flow event of equivalent flow volume

  14. Palladium clusters deposited on the heterogeneous substrates

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Kun, E-mail: cqdxwk@126.com [College of Power Engineering, Chongqing University, Chongqing 400044 (China); Key Laboratory of Low-grade Energy Utilization Technologies and Systems, Ministry of Education of PRC, Chongqing 400044 (China); Liu, Juanfang, E-mail: juanfang@cqu.edu.cn [College of Power Engineering, Chongqing University, Chongqing 400044 (China); Key Laboratory of Low-grade Energy Utilization Technologies and Systems, Ministry of Education of PRC, Chongqing 400044 (China); Chen, Qinghua, E-mail: qhchen@cqu.edu.cn [College of Power Engineering, Chongqing University, Chongqing 400044 (China); Key Laboratory of Low-grade Energy Utilization Technologies and Systems, Ministry of Education of PRC, Chongqing 400044 (China)

    2016-07-15

    Graphical abstract: The site-exchange between the substrate and cluster atoms can result in the formation of the surface alloys and the reconstruction of the cluster structure before the collision system approaching the thermal equilibrium. The deposited cluster adjusted the atom arrangement as possibly as to match the substrate lattice arrangement from bottom to up. The structural reconstruction is accompanied by the system potential energy minimization. - Highlights: • The deposition process can divide explicitly into three stages: adsorption, collision, relaxation. • The local melt does not emerge inside the substrate during the deposition process. • Surface alloys are formed by the site-exchange between the cluster and substrate atoms. • The cluster reconstructs the atom arrangement following as the substrate lattice arrangement from bottom to up. • The structural reconstruction ability and scope depend on the cluster size and incident energy. - Abstract: To improve the performance of the Pd composite membrane prepared by the cold spraying technology, it is extremely essential to give insights into the deposition process of the cluster and the heterogeneous deposition of the big Pd cluster at the different incident velocities on the atomic level. The deposition behavior, morphologies, energetic and interfacial configuration were examined by the molecular dynamic simulation and characterized by the cluster flattening ratio, the substrate maximum local temperature, the atom-embedded layer number and the surface-alloy formation. According to the morphology evolution, three deposition stages and the corresponding structural and energy evolution were clearly identified. The cluster deformation and penetrating depth increased with the enhancement of the incident velocity, but the increase degree also depended on the substrate hardness. The interfacial interaction between the cluster and the substrate can be improved by the higher substrate local temperature

  15. STAR FORMATION IN DENSE CLUSTERS

    International Nuclear Information System (INIS)

    Myers, Philip C.

    2011-01-01

    A model of core-clump accretion with equally likely stopping describes star formation in the dense parts of clusters, where models of isolated collapsing cores may not apply. Each core accretes at a constant rate onto its protostar, while the surrounding clump gas accretes as a power of protostar mass. Short accretion flows resemble Shu accretion and make low-mass stars. Long flows resemble reduced Bondi accretion and make massive stars. Accretion stops due to environmental processes of dynamical ejection, gravitational competition, and gas dispersal by stellar feedback, independent of initial core structure. The model matches the field star initial mass function (IMF) from 0.01 to more than 10 solar masses. The core accretion rate and the mean accretion duration set the peak of the IMF, independent of the local Jeans mass. Massive protostars require the longest accretion durations, up to 0.5 Myr. The maximum protostar luminosity in a cluster indicates the mass and age of its oldest protostar. The distribution of protostar luminosities matches those in active star-forming regions if protostars have a constant birthrate but not if their births are coeval. For constant birthrate, the ratio of young stellar objects to protostars indicates the star-forming age of a cluster, typically ∼1 Myr. The protostar accretion luminosity is typically less than its steady spherical value by a factor of ∼2, consistent with models of episodic disk accretion.

  16. Clustering recommendations to compute agent reputation

    Science.gov (United States)

    Bedi, Punam; Kaur, Harmeet

    2005-03-01

    Traditional centralized approaches to security are difficult to apply to multi-agent systems which are used nowadays in e-commerce applications. Developing a notion of trust that is based on the reputation of an agent can provide a softer notion of security that is sufficient for many multi-agent applications. Our paper proposes a mechanism for computing reputation of the trustee agent for use by the trustier agent. The trustier agent computes the reputation based on its own experience as well as the experience the peer agents have with the trustee agents. The trustier agents intentionally interact with the peer agents to get their experience information in the form of recommendations. We have also considered the case of unintentional encounters between the referee agents and the trustee agent, which can be directly between them or indirectly through a set of interacting agents. The clustering is done to filter off the noise in the recommendations in the form of outliers. The trustier agent clusters the recommendations received from referee agents on the basis of the distances between recommendations using the hierarchical agglomerative method. The dendogram hence obtained is cut at the required similarity level which restricts the maximum distance between any two recommendations within a cluster. The cluster with maximum number of elements denotes the views of the majority of recommenders. The center of this cluster represents the reputation of the trustee agent which can be computed using c-means algorithm.

  17. Evaluating Capacity Development

    International Development Research Centre (IDRC) Digital Library (Canada)

    She also had the dubious pleasure of checking and correcting the text numerous ... Has your organization received training or other types of support for capacity ...... processors, and consumer groups in its research and development work.

  18. Revisiting Absorptive Capacity

    DEFF Research Database (Denmark)

    de Araújo, Ana Luiza Lara; Ulhøi, John Parm; Lettl, Christopher

    Absorptive capacity has mostly been perceived as a 'passive' outcome of R&D investments. Recently, however, a growing interest into its 'proactive' potentials has emerged. This paper taps into this development and proposes a dynamic model for conceptualizing the determinants of the complementary...... learning processes of absorptive capacity, which comprise combinative and adaptive capabilities. Drawing on survey data (n=169), the study concludes that combinative capabilities primarily enhance transformative and exploratory learning processes, while adaptive capabilities strengthen all three learning...

  19. Identifying Clusters with Mixture Models that Include Radial Velocity Observations

    Science.gov (United States)

    Czarnatowicz, Alexis; Ybarra, Jason E.

    2018-01-01

    The study of stellar clusters plays an integral role in the study of star formation. We present a cluster mixture model that considers radial velocity data in addition to spatial data. Maximum likelihood estimation through the Expectation-Maximization (EM) algorithm is used for parameter estimation. Our mixture model analysis can be used to distinguish adjacent or overlapping clusters, and estimate properties for each cluster.Work supported by awards from the Virginia Foundation for Independent Colleges (VFIC) Undergraduate Science Research Fellowship and The Research Experience @Bridgewater (TREB).

  20. Cluster analysis for applications

    CERN Document Server

    Anderberg, Michael R

    1973-01-01

    Cluster Analysis for Applications deals with methods and various applications of cluster analysis. Topics covered range from variables and scales to measures of association among variables and among data units. Conceptual problems in cluster analysis are discussed, along with hierarchical and non-hierarchical clustering methods. The necessary elements of data analysis, statistics, cluster analysis, and computer implementation are integrated vertically to cover the complete path from raw data to a finished analysis.Comprised of 10 chapters, this book begins with an introduction to the subject o

  1. MEASUREMENT OF SPECIFIC HEAT CAPACITY OF SALTSTONE

    International Nuclear Information System (INIS)

    Harbour, J.; Williams, V.

    2008-01-01

    One of the goals of the Saltstone variability study is to identify (and quantify the impact of) the operational and compositional variables that control or influence the important processing and performance properties of Saltstone grout mixtures. The heat capacity of the Saltstone waste form is one of the important properties of Saltstone mixes that was last measured at SRNL in 1997. It is therefore important to develop a core competency for rapid and accurate analysis of the specific heat capacity of the Saltstone mixes in order to quantify the impact of compositional and operational variations on this property as part of the variability study. The heat capacity, coupled with the heat of hydration data obtained from isothermal calorimetry for a given Saltstone mix, can be used to predict the maximum temperature increase in the cells within the vaults of the Saltstone Disposal Facility (SDF). The temperature increase controls the processing rate and the pour schedule. The maximum temperature is also important to the performance properties of the Saltstone. For example, in mass pours of concrete or grout of which Saltstone is an example, the maximum temperature increase and the maximum temperature difference (between the surface and the hottest location) are controlled to ensure durability of the product and prevent or limit the cracking caused by the thermal gradients produced during curing. This report details the development and implementation of a method for the measurement of the heat capacities of Saltstone mixes as well as the heat capacities of the cementitious materials of the premix and the simulated salt solutions used to batch the mixes. The developed method utilizes the TAM Air isothermal calorimeter and takes advantage of the sophisticated heat flow measurement capabilities of the instrument. Standards and reference materials were identified and used to validate the procedure and ensure accuracy of testing. Heat capacities of Saltstone mixes were

  2. MEASUREMENT OF SPECIFIC HEAT CAPACITY OF SALTSTONE

    Energy Technology Data Exchange (ETDEWEB)

    Harbour, J; Vickie Williams, V

    2008-09-29

    One of the goals of the Saltstone variability study is to identify (and quantify the impact of) the operational and compositional variables that control or influence the important processing and performance properties of Saltstone grout mixtures. The heat capacity of the Saltstone waste form is one of the important properties of Saltstone mixes that was last measured at SRNL in 1997. It is therefore important to develop a core competency for rapid and accurate analysis of the specific heat capacity of the Saltstone mixes in order to quantify the impact of compositional and operational variations on this property as part of the variability study. The heat capacity, coupled with the heat of hydration data obtained from isothermal calorimetry for a given Saltstone mix, can be used to predict the maximum temperature increase in the cells within the vaults of the Saltstone Disposal Facility (SDF). The temperature increase controls the processing rate and the pour schedule. The maximum temperature is also important to the performance properties of the Saltstone. For example, in mass pours of concrete or grout of which Saltstone is an example, the maximum temperature increase and the maximum temperature difference (between the surface and the hottest location) are controlled to ensure durability of the product and prevent or limit the cracking caused by the thermal gradients produced during curing. This report details the development and implementation of a method for the measurement of the heat capacities of Saltstone mixes as well as the heat capacities of the cementitious materials of the premix and the simulated salt solutions used to batch the mixes. The developed method utilizes the TAM Air isothermal calorimeter and takes advantage of the sophisticated heat flow measurement capabilities of the instrument. Standards and reference materials were identified and used to validate the procedure and ensure accuracy of testing. Heat capacities of Saltstone mixes were

  3. Maximum gravitational redshift of white dwarfs

    International Nuclear Information System (INIS)

    Shapiro, S.L.; Teukolsky, S.A.

    1976-01-01

    The stability of uniformly rotating, cold white dwarfs is examined in the framework of the Parametrized Post-Newtonian (PPN) formalism of Will and Nordtvedt. The maximum central density and gravitational redshift of a white dwarf are determined as functions of five of the nine PPN parameters (γ, β, zeta 2 , zeta 3 , and zeta 4 ), the total angular momentum J, and the composition of the star. General relativity predicts that the maximum redshifts is 571 km s -1 for nonrotating carbon and helium dwarfs, but is lower for stars composed of heavier nuclei. Uniform rotation can increase the maximum redshift to 647 km s -1 for carbon stars (the neutronization limit) and to 893 km s -1 for helium stars (the uniform rotation limit). The redshift distribution of a larger sample of white dwarfs may help determine the composition of their cores

  4. Clusters in nuclei

    CERN Document Server

    Following the pioneering discovery of alpha clustering and of molecular resonances, the field of nuclear clustering is today one of those domains of heavy-ion nuclear physics that faces the greatest challenges, yet also contains the greatest opportunities. After many summer schools and workshops, in particular over the last decade, the community of nuclear molecular physicists has decided to collaborate in producing a comprehensive collection of lectures and tutorial reviews covering the field. This third volume follows the successful Lect. Notes Phys. 818 (Vol. 1) and 848 (Vol. 2), and comprises six extensive lectures covering the following topics:  - Gamma Rays and Molecular Structure - Faddeev Equation Approach for Three Cluster Nuclear Reactions - Tomography of the Cluster Structure of Light Nuclei Via Relativistic Dissociation - Clustering Effects Within the Dinuclear Model : From Light to Hyper-heavy Molecules in Dynamical Mean-field Approach - Clusterization in Ternary Fission - Clusters in Light N...

  5. Spatial cluster modelling

    CERN Document Server

    Lawson, Andrew B

    2002-01-01

    Research has generated a number of advances in methods for spatial cluster modelling in recent years, particularly in the area of Bayesian cluster modelling. Along with these advances has come an explosion of interest in the potential applications of this work, especially in epidemiology and genome research. In one integrated volume, this book reviews the state-of-the-art in spatial clustering and spatial cluster modelling, bringing together research and applications previously scattered throughout the literature. It begins with an overview of the field, then presents a series of chapters that illuminate the nature and purpose of cluster modelling within different application areas, including astrophysics, epidemiology, ecology, and imaging. The focus then shifts to methods, with discussions on point and object process modelling, perfect sampling of cluster processes, partitioning in space and space-time, spatial and spatio-temporal process modelling, nonparametric methods for clustering, and spatio-temporal ...

  6. BioCluster: Tool for Identification and Clustering of Enterobacteriaceae Based on Biochemical Data

    Directory of Open Access Journals (Sweden)

    Ahmed Abdullah

    2015-06-01

    Full Text Available Presumptive identification of different Enterobacteriaceae species is routinely achieved based on biochemical properties. Traditional practice includes manual comparison of each biochemical property of the unknown sample with known reference samples and inference of its identity based on the maximum similarity pattern with the known samples. This process is labor-intensive, time-consuming, error-prone, and subjective. Therefore, automation of sorting and similarity in calculation would be advantageous. Here we present a MATLAB-based graphical user interface (GUI tool named BioCluster. This tool was designed for automated clustering and identification of Enterobacteriaceae based on biochemical test results. In this tool, we used two types of algorithms, i.e., traditional hierarchical clustering (HC and the Improved Hierarchical Clustering (IHC, a modified algorithm that was developed specifically for the clustering and identification of Enterobacteriaceae species. IHC takes into account the variability in result of 1–47 biochemical tests within this Enterobacteriaceae family. This tool also provides different options to optimize the clustering in a user-friendly way. Using computer-generated synthetic data and some real data, we have demonstrated that BioCluster has high accuracy in clustering and identifying enterobacterial species based on biochemical test data. This tool can be freely downloaded at http://microbialgen.du.ac.bd/biocluster/.

  7. Clusters and how to make it work : Cluster Strategy Toolkit

    NARCIS (Netherlands)

    Manickam, Anu; van Berkel, Karel

    2014-01-01

    Clusters are the magic answer to regional economic development. Firms in clusters are more innovative; cluster policy dominates EU policy; ‘top-sectors’ and excellence are the choice of national policy makers; clusters are ‘in’. But, clusters are complex, clusters are ‘messy’; there is no clear

  8. Maximum entropy analysis of EGRET data

    DEFF Research Database (Denmark)

    Pohl, M.; Strong, A.W.

    1997-01-01

    EGRET data are usually analysed on the basis of the Maximum-Likelihood method \\cite{ma96} in a search for point sources in excess to a model for the background radiation (e.g. \\cite{hu97}). This method depends strongly on the quality of the background model, and thus may have high systematic unce...... uncertainties in region of strong and uncertain background like the Galactic Center region. Here we show images of such regions obtained by the quantified Maximum-Entropy method. We also discuss a possible further use of MEM in the analysis of problematic regions of the sky....

  9. The Maximum Resource Bin Packing Problem

    DEFF Research Database (Denmark)

    Boyar, J.; Epstein, L.; Favrholdt, L.M.

    2006-01-01

    Usually, for bin packing problems, we try to minimize the number of bins used or in the case of the dual bin packing problem, maximize the number or total size of accepted items. This paper presents results for the opposite problems, where we would like to maximize the number of bins used...... algorithms, First-Fit-Increasing and First-Fit-Decreasing for the maximum resource variant of classical bin packing. For the on-line variant, we define maximum resource variants of classical and dual bin packing. For dual bin packing, no on-line algorithm is competitive. For classical bin packing, we find...

  10. Shower maximum detector for SDC calorimetry

    International Nuclear Information System (INIS)

    Ernwein, J.

    1994-01-01

    A prototype for the SDC end-cap (EM) calorimeter complete with a pre-shower and a shower maximum detector was tested in beams of electrons and Π's at CERN by an SDC subsystem group. The prototype was manufactured from scintillator tiles and strips read out with 1 mm diameter wave-length shifting fibers. The design and construction of the shower maximum detector is described, and results of laboratory tests on light yield and performance of the scintillator-fiber system are given. Preliminary results on energy and position measurements with the shower max detector in the test beam are shown. (authors). 4 refs., 5 figs

  11. Topics in Bayesian statistics and maximum entropy

    International Nuclear Information System (INIS)

    Mutihac, R.; Cicuttin, A.; Cerdeira, A.; Stanciulescu, C.

    1998-12-01

    Notions of Bayesian decision theory and maximum entropy methods are reviewed with particular emphasis on probabilistic inference and Bayesian modeling. The axiomatic approach is considered as the best justification of Bayesian analysis and maximum entropy principle applied in natural sciences. Particular emphasis is put on solving the inverse problem in digital image restoration and Bayesian modeling of neural networks. Further topics addressed briefly include language modeling, neutron scattering, multiuser detection and channel equalization in digital communications, genetic information, and Bayesian court decision-making. (author)

  12. Density estimation by maximum quantum entropy

    International Nuclear Information System (INIS)

    Silver, R.N.; Wallstrom, T.; Martz, H.F.

    1993-01-01

    A new Bayesian method for non-parametric density estimation is proposed, based on a mathematical analogy to quantum statistical physics. The mathematical procedure is related to maximum entropy methods for inverse problems and image reconstruction. The information divergence enforces global smoothing toward default models, convexity, positivity, extensivity and normalization. The novel feature is the replacement of classical entropy by quantum entropy, so that local smoothing is enforced by constraints on differential operators. The linear response of the estimate is proportional to the covariance. The hyperparameters are estimated by type-II maximum likelihood (evidence). The method is demonstrated on textbook data sets

  13. 30 CFR 75.601-3 - Short circuit protection; dual element fuses; current ratings; maximum values.

    Science.gov (United States)

    2010-07-01

    ... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Short circuit protection; dual element fuses... Trailing Cables § 75.601-3 Short circuit protection; dual element fuses; current ratings; maximum values. Dual element fuses having adequate current-interrupting capacity shall meet the requirements for short...

  14. Nonsymmetric entropy and maximum nonsymmetric entropy principle

    International Nuclear Information System (INIS)

    Liu Chengshi

    2009-01-01

    Under the frame of a statistical model, the concept of nonsymmetric entropy which generalizes the concepts of Boltzmann's entropy and Shannon's entropy, is defined. Maximum nonsymmetric entropy principle is proved. Some important distribution laws such as power law, can be derived from this principle naturally. Especially, nonsymmetric entropy is more convenient than other entropy such as Tsallis's entropy in deriving power laws.

  15. Maximum speed of dewetting on a fiber

    NARCIS (Netherlands)

    Chan, Tak Shing; Gueudre, Thomas; Snoeijer, Jacobus Hendrikus

    2011-01-01

    A solid object can be coated by a nonwetting liquid since a receding contact line cannot exceed a critical speed. We theoretically investigate this forced wetting transition for axisymmetric menisci on fibers of varying radii. First, we use a matched asymptotic expansion and derive the maximum speed

  16. Maximum potential preventive effect of hip protectors

    NARCIS (Netherlands)

    van Schoor, N.M.; Smit, J.H.; Bouter, L.M.; Veenings, B.; Asma, G.B.; Lips, P.T.A.M.

    2007-01-01

    OBJECTIVES: To estimate the maximum potential preventive effect of hip protectors in older persons living in the community or homes for the elderly. DESIGN: Observational cohort study. SETTING: Emergency departments in the Netherlands. PARTICIPANTS: Hip fracture patients aged 70 and older who

  17. Maximum gain of Yagi-Uda arrays

    DEFF Research Database (Denmark)

    Bojsen, J.H.; Schjær-Jacobsen, Hans; Nilsson, E.

    1971-01-01

    Numerical optimisation techniques have been used to find the maximum gain of some specific parasitic arrays. The gain of an array of infinitely thin, equispaced dipoles loaded with arbitrary reactances has been optimised. The results show that standard travelling-wave design methods are not optimum....... Yagi–Uda arrays with equal and unequal spacing have also been optimised with experimental verification....

  18. correlation between maximum dry density and cohesion

    African Journals Online (AJOL)

    HOD

    represents maximum dry density, signifies plastic limit and is liquid limit. Researchers [6, 7] estimate compaction parameters. Aside from the correlation existing between compaction parameters and other physical quantities there are some other correlations that have been investigated by other researchers. The well-known.

  19. Weak scale from the maximum entropy principle

    Science.gov (United States)

    Hamada, Yuta; Kawai, Hikaru; Kawana, Kiyoharu

    2015-03-01

    The theory of the multiverse and wormholes suggests that the parameters of the Standard Model (SM) are fixed in such a way that the radiation of the S3 universe at the final stage S_rad becomes maximum, which we call the maximum entropy principle. Although it is difficult to confirm this principle generally, for a few parameters of the SM, we can check whether S_rad actually becomes maximum at the observed values. In this paper, we regard S_rad at the final stage as a function of the weak scale (the Higgs expectation value) vh, and show that it becomes maximum around vh = {{O}} (300 GeV) when the dimensionless couplings in the SM, i.e., the Higgs self-coupling, the gauge couplings, and the Yukawa couplings are fixed. Roughly speaking, we find that the weak scale is given by vh ˜ T_{BBN}2 / (M_{pl}ye5), where ye is the Yukawa coupling of electron, T_BBN is the temperature at which the Big Bang nucleosynthesis starts, and M_pl is the Planck mass.

  20. The maximum-entropy method in superspace

    Czech Academy of Sciences Publication Activity Database

    van Smaalen, S.; Palatinus, Lukáš; Schneider, M.

    2003-01-01

    Roč. 59, - (2003), s. 459-469 ISSN 0108-7673 Grant - others:DFG(DE) XX Institutional research plan: CEZ:AV0Z1010914 Keywords : maximum-entropy method, * aperiodic crystals * electron density Subject RIV: BM - Solid Matter Physics ; Magnetism Impact factor: 1.558, year: 2003

  1. Achieving maximum sustainable yield in mixed fisheries

    NARCIS (Netherlands)

    Ulrich, Clara; Vermard, Youen; Dolder, Paul J.; Brunel, Thomas; Jardim, Ernesto; Holmes, Steven J.; Kempf, Alexander; Mortensen, Lars O.; Poos, Jan Jaap; Rindorf, Anna

    2017-01-01

    Achieving single species maximum sustainable yield (MSY) in complex and dynamic fisheries targeting multiple species (mixed fisheries) is challenging because achieving the objective for one species may mean missing the objective for another. The North Sea mixed fisheries are a representative example

  2. 5 CFR 534.203 - Maximum stipends.

    Science.gov (United States)

    2010-01-01

    ... maximum stipend established under this section. (e) A trainee at a non-Federal hospital, clinic, or medical or dental laboratory who is assigned to a Federal hospital, clinic, or medical or dental... Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PAY UNDER OTHER SYSTEMS Student...

  3. Minimal length, Friedmann equations and maximum density

    Energy Technology Data Exchange (ETDEWEB)

    Awad, Adel [Center for Theoretical Physics, British University of Egypt,Sherouk City 11837, P.O. Box 43 (Egypt); Department of Physics, Faculty of Science, Ain Shams University,Cairo, 11566 (Egypt); Ali, Ahmed Farag [Centre for Fundamental Physics, Zewail City of Science and Technology,Sheikh Zayed, 12588, Giza (Egypt); Department of Physics, Faculty of Science, Benha University,Benha, 13518 (Egypt)

    2014-06-16

    Inspired by Jacobson’s thermodynamic approach, Cai et al. have shown the emergence of Friedmann equations from the first law of thermodynamics. We extend Akbar-Cai derivation http://dx.doi.org/10.1103/PhysRevD.75.084003 of Friedmann equations to accommodate a general entropy-area law. Studying the resulted Friedmann equations using a specific entropy-area law, which is motivated by the generalized uncertainty principle (GUP), reveals the existence of a maximum energy density closed to Planck density. Allowing for a general continuous pressure p(ρ,a) leads to bounded curvature invariants and a general nonsingular evolution. In this case, the maximum energy density is reached in a finite time and there is no cosmological evolution beyond this point which leaves the big bang singularity inaccessible from a spacetime prospective. The existence of maximum energy density and a general nonsingular evolution is independent of the equation of state and the spacial curvature k. As an example we study the evolution of the equation of state p=ωρ through its phase-space diagram to show the existence of a maximum energy which is reachable in a finite time.

  4. Agricultural Clusters in the Netherlands

    NARCIS (Netherlands)

    Schouten, M.A.; Heijman, W.J.M.

    2012-01-01

    Michael Porter was the first to use the term cluster in an economic context. He introduced the term in The Competitive Advantage of Nations (1990). The term cluster is also known as business cluster, industry cluster, competitive cluster or Porterian cluster. This article aims at determining and

  5. Deployment Strategies and Clustering Protocols Efficiency

    Directory of Open Access Journals (Sweden)

    Chérif Diallo

    2017-06-01

    Full Text Available Wireless sensor networks face significant design challenges due to limited computing and storage capacities and, most importantly, dependence on limited battery power. Energy is a critical resource and is often an important issue to the deployment of sensor applications that claim to be omnipresent in the world of future. Thus optimizing the deployment of sensors becomes a major constraint in the design and implementation of a WSN in order to ensure better network operations. In wireless networking, clustering techniques add scalability, reduce the computation complexity of routing protocols, allow data aggregation and then enhance the network performance. The well-known MaxMin clustering algorithm was previously generalized, corrected and validated. Then, in a previous work we have improved MaxMin by proposing a Single- node Cluster Reduction (SNCR mechanism which eliminates single-node clusters and then improve energy efficiency. In this paper, we show that MaxMin, because of its original pathological case, does not support the grid deployment topology, which is frequently used in WSN architectures. The unreliability feature of the wireless links could have negative impacts on Link Quality Indicator (LQI based clustering protocols. So, in the second part of this paper we show how our distributed Link Quality based d- Clustering Protocol (LQI-DCP has good performance in both stable and high unreliable link environments. Finally, performance evaluation results also show that LQI-DCP fully supports the grid deployment topology and is more energy efficient than MaxMin.

  6. Open source clustering software.

    Science.gov (United States)

    de Hoon, M J L; Imoto, S; Nolan, J; Miyano, S

    2004-06-12

    We have implemented k-means clustering, hierarchical clustering and self-organizing maps in a single multipurpose open-source library of C routines, callable from other C and C++ programs. Using this library, we have created an improved version of Michael Eisen's well-known Cluster program for Windows, Mac OS X and Linux/Unix. In addition, we generated a Python and a Perl interface to the C Clustering Library, thereby combining the flexibility of a scripting language with the speed of C. The C Clustering Library and the corresponding Python C extension module Pycluster were released under the Python License, while the Perl module Algorithm::Cluster was released under the Artistic License. The GUI code Cluster 3.0 for Windows, Macintosh and Linux/Unix, as well as the corresponding command-line program, were released under the same license as the original Cluster code. The complete source code is available at http://bonsai.ims.u-tokyo.ac.jp/mdehoon/software/cluster. Alternatively, Algorithm::Cluster can be downloaded from CPAN, while Pycluster is also available as part of the Biopython distribution.

  7. Capacity Maximizing Constellations

    Science.gov (United States)

    Barsoum, Maged; Jones, Christopher

    2010-01-01

    Some non-traditional signal constellations have been proposed for transmission of data over the Additive White Gaussian Noise (AWGN) channel using such channel-capacity-approaching codes as low-density parity-check (LDPC) or turbo codes. Computational simulations have shown performance gains of more than 1 dB over traditional constellations. These gains could be translated to bandwidth- efficient communications, variously, over longer distances, using less power, or using smaller antennas. The proposed constellations have been used in a bit-interleaved coded modulation system employing state-ofthe-art LDPC codes. In computational simulations, these constellations were shown to afford performance gains over traditional constellations as predicted by the gap between the parallel decoding capacity of the constellations and the Gaussian capacity

  8. Visual attention capacity

    DEFF Research Database (Denmark)

    Habekost, Thomas; Starrfelt, Randi

    2009-01-01

    Psychophysical studies have identified two distinct limitations of visual attention capacity: processing speed and apprehension span. Using a simple test, these cognitive factors can be analyzed by Bundesen's Theory of Visual Attention (TVA). The method has strong specificity and sensitivity......, and measurements are highly reliable. As the method is theoretically founded, it also has high validity. TVA-based assessment has recently been used to investigate a broad range of neuropsychological and neurological conditions. We present the method, including the experimental paradigm and practical guidelines...... to patient testing, and review existing TVA-based patient studies organized by lesion anatomy. Lesions in three anatomical regions affect visual capacity: The parietal lobes, frontal cortex and basal ganglia, and extrastriate cortex. Visual capacity thus depends on large, bilaterally distributed anatomical...

  9. Maximum Likelihood Blind Channel Estimation for Space-Time Coding Systems

    Directory of Open Access Journals (Sweden)

    Hakan A. Çırpan

    2002-05-01

    Full Text Available Sophisticated signal processing techniques have to be developed for capacity enhancement of future wireless communication systems. In recent years, space-time coding is proposed to provide significant capacity gains over the traditional communication systems in fading wireless channels. Space-time codes are obtained by combining channel coding, modulation, transmit diversity, and optional receive diversity in order to provide diversity at the receiver and coding gain without sacrificing the bandwidth. In this paper, we consider the problem of blind estimation of space-time coded signals along with the channel parameters. Both conditional and unconditional maximum likelihood approaches are developed and iterative solutions are proposed. The conditional maximum likelihood algorithm is based on iterative least squares with projection whereas the unconditional maximum likelihood approach is developed by means of finite state Markov process modelling. The performance analysis issues of the proposed methods are studied. Finally, some simulation results are presented.

  10. Value for railway capacity

    DEFF Research Database (Denmark)

    Sameni, Melody Khadem; Preston, John M.

    2012-01-01

    to analyze the efficiency of operators in transforming inputs of allocated capacity of infrastructure and franchise payments into valuable passenger service outputs while avoiding delays. By addressing operational and economic aspects of capacity utilization simultaneously, the paper deviates from existing...... DEA work on the economic efficiency of railways by considering a new combination of input-output that also incorporates quality of service. The constant and variable returns to scale models are applied to the case study of franchised passenger operators in Great Britain. The follow-up Tobit regression...

  11. Electron: Cluster interactions

    International Nuclear Information System (INIS)

    Scheidemann, A.A.; Knight, W.D.

    1994-02-01

    Beam depletion spectroscopy has been used to measure absolute total inelastic electron-sodium cluster collision cross sections in the energy range from E ∼ 0.1 to E ∼ 6 eV. The investigation focused on the closed shell clusters Na 8 , Na 20 , Na 40 . The measured cross sections show an increase for the lowest collision energies where electron attachment is the primary scattering channel. The electron attachment cross section can be understood in terms of Langevin scattering, connecting this measurement with the polarizability of the cluster. For energies above the dissociation energy the measured electron-cluster cross section is energy independent, thus defining an electron-cluster interaction range. This interaction range increases with the cluster size

  12. Clustering high dimensional data

    DEFF Research Database (Denmark)

    Assent, Ira

    2012-01-01

    High-dimensional data, i.e., data described by a large number of attributes, pose specific challenges to clustering. The so-called ‘curse of dimensionality’, coined originally to describe the general increase in complexity of various computational problems as dimensionality increases, is known...... to render traditional clustering algorithms ineffective. The curse of dimensionality, among other effects, means that with increasing number of dimensions, a loss of meaningful differentiation between similar and dissimilar objects is observed. As high-dimensional objects appear almost alike, new approaches...... for clustering are required. Consequently, recent research has focused on developing techniques and clustering algorithms specifically for high-dimensional data. Still, open research issues remain. Clustering is a data mining task devoted to the automatic grouping of data based on mutual similarity. Each cluster...

  13. Hydrogen storage capacity of titanium met-cars

    International Nuclear Information System (INIS)

    Akman, N; Durgun, E; Yildirim, T; Ciraci, S

    2006-01-01

    The adsorption of hydrogen molecules on the titanium metallocarbohedryne (met-car) cluster has been investigated by using the first-principles plane wave method. We have found that, while a single Ti atom at the corner can bind up to three hydrogen molecules, a single Ti atom on the surface of the cluster can bind only one hydrogen molecule. Accordingly, a Ti 8 C 12 met-car can bind up to 16 H 2 molecules and hence can be considered as a high-capacity hydrogen storage medium. Strong interaction between two met-car clusters leading to the dimer formation can affect H 2 storage capacity slightly. Increasing the storage capacity by directly inserting H 2 into the met-car or by functionalizing it with an Na atom have been explored. It is found that the insertion of neither an H 2 molecule nor an Na atom could further promote the H 2 storage capacity of a Ti 8 C 12 cluster. We have also tested the stability of the H 2 -adsorbed Ti 8 C 12 met-car with ab initio molecular dynamics calculations which have been carried out at room temperature

  14. Substructure in clusters of galaxies

    International Nuclear Information System (INIS)

    Fitchett, M.J.

    1988-01-01

    Optical observations suggesting the existence of substructure in clusters of galaxies are examined. Models of cluster formation and methods used to detect substructure in clusters are reviewed. Consideration is given to classification schemes based on a departure of bright cluster galaxies from a spherically symmetric distribution, evidence for statistically significant substructure, and various types of substructure, including velocity, spatial, and spatial-velocity substructure. The substructure observed in the galaxy distribution in clusters is discussed, focusing on observations from general cluster samples, the Virgo cluster, the Hydra cluster, Centaurus, the Coma cluster, and the Cancer cluster. 88 refs

  15. Maximum concentrations at work and maximum biologically tolerable concentration for working materials 1991

    International Nuclear Information System (INIS)

    1991-01-01

    The meaning of the term 'maximum concentration at work' in regard of various pollutants is discussed. Specifically, a number of dusts and smokes are dealt with. The valuation criteria for maximum biologically tolerable concentrations for working materials are indicated. The working materials in question are corcinogeneous substances or substances liable to cause allergies or mutate the genome. (VT) [de

  16. 75 FR 43840 - Inflation Adjustment of the Ordinary Maximum and Aggravated Maximum Civil Monetary Penalties for...

    Science.gov (United States)

    2010-07-27

    ...-17530; Notice No. 2] RIN 2130-ZA03 Inflation Adjustment of the Ordinary Maximum and Aggravated Maximum... remains at $250. These adjustments are required by the Federal Civil Penalties Inflation Adjustment Act [email protected] . SUPPLEMENTARY INFORMATION: The Federal Civil Penalties Inflation Adjustment Act of 1990...

  17. Nuclear cluster states

    International Nuclear Information System (INIS)

    Rae, W.D.M.; Merchant, A.C.

    1993-01-01

    We review clustering in light nuclei including molecular resonances in heavy ion reactions. In particular we study the systematics, paying special attention to the relationships between cluster states and superdeformed configurations. We emphasise the selection rules which govern the formation and decay of cluster states. We review some recent experimental results from Daresbury and elsewhere. In particular we report on the evidence for a 7-α chain state in 28 Si in experiments recently performed at the NSF, Daresbury. Finally we begin to address theoretically the important question of the lifetimes of cluster states as deduced from the experimental energy widths of the resonances. (Author)

  18. 15th Cluster workshop

    CERN Document Server

    Laakso, Harri; Escoubet, C. Philippe; The Cluster Active Archive : Studying the Earth’s Space Plasma Environment

    2010-01-01

    Since the year 2000 the ESA Cluster mission has been investigating the small-scale structures and processes of the Earth's plasma environment, such as those involved in the interaction between the solar wind and the magnetospheric plasma, in global magnetotail dynamics, in cross-tail currents, and in the formation and dynamics of the neutral line and of plasmoids. This book contains presentations made at the 15th Cluster workshop held in March 2008. It also presents several articles about the Cluster Active Archive and its datasets, a few overview papers on the Cluster mission, and articles reporting on scientific findings on the solar wind, the magnetosheath, the magnetopause and the magnetotail.

  19. Clusters in simple fluids

    International Nuclear Information System (INIS)

    Sator, N.

    2003-01-01

    This article concerns the correspondence between thermodynamics and the morphology of simple fluids in terms of clusters. Definitions of clusters providing a geometric interpretation of the liquid-gas phase transition are reviewed with an eye to establishing their physical relevance. The author emphasizes their main features and basic hypotheses, and shows how these definitions lead to a recent approach based on self-bound clusters. Although theoretical, this tutorial review is also addressed to readers interested in experimental aspects of clustering in simple fluids

  20. Zipf's law, power laws and maximum entropy

    International Nuclear Information System (INIS)

    Visser, Matt

    2013-01-01

    Zipf's law, and power laws in general, have attracted and continue to attract considerable attention in a wide variety of disciplines—from astronomy to demographics to software structure to economics to linguistics to zoology, and even warfare. A recent model of random group formation (RGF) attempts a general explanation of such phenomena based on Jaynes' notion of maximum entropy applied to a particular choice of cost function. In the present paper I argue that the specific cost function used in the RGF model is in fact unnecessarily complicated, and that power laws can be obtained in a much simpler way by applying maximum entropy ideas directly to the Shannon entropy subject only to a single constraint: that the average of the logarithm of the observable quantity is specified. (paper)

  1. Maximum-entropy description of animal movement.

    Science.gov (United States)

    Fleming, Chris H; Subaşı, Yiğit; Calabrese, Justin M

    2015-03-01

    We introduce a class of maximum-entropy states that naturally includes within it all of the major continuous-time stochastic processes that have been applied to animal movement, including Brownian motion, Ornstein-Uhlenbeck motion, integrated Ornstein-Uhlenbeck motion, a recently discovered hybrid of the previous models, and a new model that describes central-place foraging. We are also able to predict a further hierarchy of new models that will emerge as data quality improves to better resolve the underlying continuity of animal movement. Finally, we also show that Langevin equations must obey a fluctuation-dissipation theorem to generate processes that fall from this class of maximum-entropy distributions when the constraints are purely kinematic.

  2. Pareto versus lognormal: a maximum entropy test.

    Science.gov (United States)

    Bee, Marco; Riccaboni, Massimo; Schiavo, Stefano

    2011-08-01

    It is commonly found that distributions that seem to be lognormal over a broad range change to a power-law (Pareto) distribution for the last few percentiles. The distributions of many physical, natural, and social events (earthquake size, species abundance, income and wealth, as well as file, city, and firm sizes) display this structure. We present a test for the occurrence of power-law tails in statistical distributions based on maximum entropy. This methodology allows one to identify the true data-generating processes even in the case when it is neither lognormal nor Pareto. The maximum entropy approach is then compared with other widely used methods and applied to different levels of aggregation of complex systems. Our results provide support for the theory that distributions with lognormal body and Pareto tail can be generated as mixtures of lognormally distributed units.

  3. Maximum likelihood estimation for integrated diffusion processes

    DEFF Research Database (Denmark)

    Baltazar-Larios, Fernando; Sørensen, Michael

    We propose a method for obtaining maximum likelihood estimates of parameters in diffusion models when the data is a discrete time sample of the integral of the process, while no direct observations of the process itself are available. The data are, moreover, assumed to be contaminated...... EM-algorithm to obtain maximum likelihood estimates of the parameters in the diffusion model. As part of the algorithm, we use a recent simple method for approximate simulation of diffusion bridges. In simulation studies for the Ornstein-Uhlenbeck process and the CIR process the proposed method works...... by measurement errors. Integrated volatility is an example of this type of observations. Another example is ice-core data on oxygen isotopes used to investigate paleo-temperatures. The data can be viewed as incomplete observations of a model with a tractable likelihood function. Therefore we propose a simulated...

  4. A Maximum Radius for Habitable Planets.

    Science.gov (United States)

    Alibert, Yann

    2015-09-01

    We compute the maximum radius a planet can have in order to fulfill two constraints that are likely necessary conditions for habitability: 1- surface temperature and pressure compatible with the existence of liquid water, and 2- no ice layer at the bottom of a putative global ocean, that would prevent the operation of the geologic carbon cycle to operate. We demonstrate that, above a given radius, these two constraints cannot be met: in the Super-Earth mass range (1-12 Mearth), the overall maximum that a planet can have varies between 1.8 and 2.3 Rearth. This radius is reduced when considering planets with higher Fe/Si ratios, and taking into account irradiation effects on the structure of the gas envelope.

  5. Maximum parsimony on subsets of taxa.

    Science.gov (United States)

    Fischer, Mareike; Thatte, Bhalchandra D

    2009-09-21

    In this paper we investigate mathematical questions concerning the reliability (reconstruction accuracy) of Fitch's maximum parsimony algorithm for reconstructing the ancestral state given a phylogenetic tree and a character. In particular, we consider the question whether the maximum parsimony method applied to a subset of taxa can reconstruct the ancestral state of the root more accurately than when applied to all taxa, and we give an example showing that this indeed is possible. A surprising feature of our example is that ignoring a taxon closer to the root improves the reliability of the method. On the other hand, in the case of the two-state symmetric substitution model, we answer affirmatively a conjecture of Li, Steel and Zhang which states that under a molecular clock the probability that the state at a single taxon is a correct guess of the ancestral state is a lower bound on the reconstruction accuracy of Fitch's method applied to all taxa.

  6. Dirichlet Process Parsimonious Mixtures for clustering

    OpenAIRE

    Chamroukhi, Faicel; Bartcus, Marius; Glotin, Hervé

    2015-01-01

    The parsimonious Gaussian mixture models, which exploit an eigenvalue decomposition of the group covariance matrices of the Gaussian mixture, have shown their success in particular in cluster analysis. Their estimation is in general performed by maximum likelihood estimation and has also been considered from a parametric Bayesian prospective. We propose new Dirichlet Process Parsimonious mixtures (DPPM) which represent a Bayesian nonparametric formulation of these parsimonious Gaussian mixtur...

  7. The composite sequential clustering technique for analysis of multispectral scanner data

    Science.gov (United States)

    Su, M. Y.

    1972-01-01

    The clustering technique consists of two parts: (1) a sequential statistical clustering which is essentially a sequential variance analysis, and (2) a generalized K-means clustering. In this composite clustering technique, the output of (1) is a set of initial clusters which are input to (2) for further improvement by an iterative scheme. This unsupervised composite technique was employed for automatic classification of two sets of remote multispectral earth resource observations. The classification accuracy by the unsupervised technique is found to be comparable to that by traditional supervised maximum likelihood classification techniques. The mathematical algorithms for the composite sequential clustering program and a detailed computer program description with job setup are given.

  8. Constraints on H(0) from globular clusters

    International Nuclear Information System (INIS)

    Vandenberg, D.A.

    1988-01-01

    On the basis of canonical stellar evolutionary computations, the maximum age of the globular clusters is suggested to be near either 14 Gyr or 18 Gyr, depending on how (O/Fe) varies with (Fe/H) in the cluster stars. The lower estimate requires that H(0) = 65 km/s/Mpc or less, for all Omega(0) = O or greater, if the standard Big-Bang cosmological theory is correct - while the higher age value similarly constrains the Hubble constant to be smaller than 46 km/s/Mpc. Some reduction in the upper limit to cluster ages and a consequent increase in H(0) may be expected if helium diffusion is important in Population II stars; nevertheless, values of H(0) greater than 75 km/s/Mpc still appear to be precluded unless the cosmological constant is nonzero. 51 refs

  9. Infrared dust emission from globular clusters

    International Nuclear Information System (INIS)

    Angeletti, L.; Capuzzo-Dolcetta, R.; Giannone, P.; Blanco, A.; Bussoletti, E.

    1982-01-01

    The implications of the presence of a central cloud in the cores of globular clusters were investigated recently. A possible mechanism of confinement of dust in the central region of our cluster models was also explored. The grain temperature and infrared emission have now been computed for rather realistic grain compositions. The grain components were assumed to be graphite and/or silicates. The central clouds turned out to be roughly isothermal. The wavelengths of maximum emission came out to be larger than 20 μm in all studied cases. An application of the theoretical results to five globular clusters showed that the predictable infrared emission for 47 Tuc, M4 and M22 should be detectable by means of present instrumentation aboard flying platforms. (author)

  10. Infrared dust emission from globular clusters

    Energy Technology Data Exchange (ETDEWEB)

    Angeletti, L; Capuzzo-Dolcetta, R; Giannone, P. (Rome Univ. (Italy). Osservatorio Astronomico); Blanco, A; Bussoletti, E [Lecce Univ. (Italy). Ist. di Fisica

    1982-05-01

    The implications of the presence of a central cloud in the cores of globular clusters were investigated recently. A possible mechanism of confinement of dust in the central region of our cluster models was also explored. The grain temperature and infrared emission have now been computed for rather realistic grain compositions. The grain components were assumed to be graphite and/or silicates. The central clouds turned out to be roughly isothermal. The wavelengths of maximum emission came out to be larger than 20 ..mu..m in all studied cases. An application of the theoretical results to five globular clusters showed that the predictable infrared emission for 47 Tuc, M4 and M22 should be detectable by means of present instrumentation aboard flying platforms.

  11. ON THE MAXIMUM MASS OF STELLAR BLACK HOLES

    International Nuclear Information System (INIS)

    Belczynski, Krzysztof; Fryer, Chris L.; Bulik, Tomasz; Ruiter, Ashley; Valsecchi, Francesca; Vink, Jorick S.; Hurley, Jarrod R.

    2010-01-01

    We present the spectrum of compact object masses: neutron stars and black holes (BHs) that originate from single stars in different environments. In particular, we calculate the dependence of maximum BH mass on metallicity and on some specific wind mass loss rates (e.g., Hurley et al. and Vink et al.). Our calculations show that the highest mass BHs observed in the Galaxy M bh ∼ 15 M sun in the high metallicity environment (Z = Z sun = 0.02) can be explained with stellar models and the wind mass loss rates adopted here. To reach this result we had to set luminous blue variable mass loss rates at the level of ∼10 -4 M sun yr -1 and to employ metallicity-dependent Wolf-Rayet winds. With such winds, calibrated on Galactic BH mass measurements, the maximum BH mass obtained for moderate metallicity (Z = 0.3 Z sun = 0.006) is M bh,max = 30 M sun . This is a rather striking finding as the mass of the most massive known stellar BH is M bh = 23-34 M sun and, in fact, it is located in a small star-forming galaxy with moderate metallicity. We find that in the very low (globular cluster-like) metallicity environment the maximum BH mass can be as high as M bh,max = 80 M sun (Z = 0.01 Z sun = 0.0002). It is interesting to note that X-ray luminosity from Eddington-limited accretion onto an 80 M sun BH is of the order of ∼10 40 erg s -1 and is comparable to luminosities of some known ultra-luminous X-ray sources. We emphasize that our results were obtained for single stars only and that binary interactions may alter these maximum BH masses (e.g., accretion from a close companion). This is strictly a proof-of-principle study which demonstrates that stellar models can naturally explain even the most massive known stellar BHs.

  12. Maximum entropy analysis of liquid diffraction data

    International Nuclear Information System (INIS)

    Root, J.H.; Egelstaff, P.A.; Nickel, B.G.

    1986-01-01

    A maximum entropy method for reducing truncation effects in the inverse Fourier transform of structure factor, S(q), to pair correlation function, g(r), is described. The advantages and limitations of the method are explored with the PY hard sphere structure factor as model input data. An example using real data on liquid chlorine, is then presented. It is seen that spurious structure is greatly reduced in comparison to traditional Fourier transform methods. (author)

  13. Clustering aspects in nuclear structure functions

    International Nuclear Information System (INIS)

    Hirai, M.; Saito, K.; Watanabe, T.; Kumano, S.

    2011-01-01

    For understanding an anomalous nuclear effect experimentally observed for the beryllium-9 nucleus at the Thomas Jefferson National Accelerator Facility, clustering aspects are studied in structure functions of deep inelastic lepton-nucleus scattering by using momentum distributions calculated in antisymmetrized (or fermionic) molecular dynamics (AMD) and also in a simple shell model for comparison. According to AMD, the 9 Be nucleus consists of two α-like clusters with a surrounding neutron. The clustering produces high-momentum components in nuclear wave functions, which affects nuclear modifications of the structure functions. We investigated whether clustering features could appear in the structure function F 2 of 9 Be along with studies for other light nuclei. We found that nuclear modifications of F 2 are similar in both AMD and shell models within our simple convolution description although there are slight differences in 9 Be. It indicates that the anomalous 9 Be result should be explained by a different mechanism from the nuclear binding and Fermi motion. If nuclear-modification slopes d(F 2 A /F 2 D )/dx are shown by the maximum local densities, the 9 Be anomaly can be explained by the AMD picture, namely by the clustering structure, whereas it certainly cannot be described in the simple shell model. This fact suggests that the large nuclear modification in 9 Be should be explained by large densities in the clusters. For example, internal nucleon structure could be modified in the high-density clusters. The clustering aspect of nuclear structure functions is an unexplored topic which is interesting for future investigations.

  14. A Maximum Resonant Set of Polyomino Graphs

    Directory of Open Access Journals (Sweden)

    Zhang Heping

    2016-05-01

    Full Text Available A polyomino graph P is a connected finite subgraph of the infinite plane grid such that each finite face is surrounded by a regular square of side length one and each edge belongs to at least one square. A dimer covering of P corresponds to a perfect matching. Different dimer coverings can interact via an alternating cycle (or square with respect to them. A set of disjoint squares of P is a resonant set if P has a perfect matching M so that each one of those squares is M-alternating. In this paper, we show that if K is a maximum resonant set of P, then P − K has a unique perfect matching. We further prove that the maximum forcing number of a polyomino graph is equal to the cardinality of a maximum resonant set. This confirms a conjecture of Xu et al. [26]. We also show that if K is a maximal alternating set of P, then P − K has a unique perfect matching.

  15. Automatic maximum entropy spectral reconstruction in NMR

    International Nuclear Information System (INIS)

    Mobli, Mehdi; Maciejewski, Mark W.; Gryk, Michael R.; Hoch, Jeffrey C.

    2007-01-01

    Developments in superconducting magnets, cryogenic probes, isotope labeling strategies, and sophisticated pulse sequences together have enabled the application, in principle, of high-resolution NMR spectroscopy to biomolecular systems approaching 1 megadalton. In practice, however, conventional approaches to NMR that utilize the fast Fourier transform, which require data collected at uniform time intervals, result in prohibitively lengthy data collection times in order to achieve the full resolution afforded by high field magnets. A variety of approaches that involve nonuniform sampling have been proposed, each utilizing a non-Fourier method of spectrum analysis. A very general non-Fourier method that is capable of utilizing data collected using any of the proposed nonuniform sampling strategies is maximum entropy reconstruction. A limiting factor in the adoption of maximum entropy reconstruction in NMR has been the need to specify non-intuitive parameters. Here we describe a fully automated system for maximum entropy reconstruction that requires no user-specified parameters. A web-accessible script generator provides the user interface to the system

  16. maximum neutron flux at thermal nuclear reactors

    International Nuclear Information System (INIS)

    Strugar, P.

    1968-10-01

    Since actual research reactors are technically complicated and expensive facilities it is important to achieve savings by appropriate reactor lattice configurations. There is a number of papers, and practical examples of reactors with central reflector, dealing with spatial distribution of fuel elements which would result in higher neutron flux. Common disadvantage of all the solutions is that the choice of best solution is done starting from the anticipated spatial distributions of fuel elements. The weakness of these approaches is lack of defined optimization criteria. Direct approach is defined as follows: determine the spatial distribution of fuel concentration starting from the condition of maximum neutron flux by fulfilling the thermal constraints. Thus the problem of determining the maximum neutron flux is solving a variational problem which is beyond the possibilities of classical variational calculation. This variational problem has been successfully solved by applying the maximum principle of Pontrjagin. Optimum distribution of fuel concentration was obtained in explicit analytical form. Thus, spatial distribution of the neutron flux and critical dimensions of quite complex reactor system are calculated in a relatively simple way. In addition to the fact that the results are innovative this approach is interesting because of the optimization procedure itself [sr

  17. Evaluation of railway capacity

    DEFF Research Database (Denmark)

    Landex, Alex; Kaas, Anders H.; Schittenhelm, Bernd

    2006-01-01

    capacity consumptions. This paper describes the UIC 406 method and how it is expounded in Denmark. The paper describes the importance of choosing the right length of the line sections examined and how line sections with multiple track sections are examined. Furthermore, the possibility of using idle...

  18. Markets and Institutional Capacity

    DEFF Research Database (Denmark)

    Ingemann, Jan Holm

    2010-01-01

    Adequate explanations concerning the introduction of production and consumption of organic food in Denmark imply the necessity to engage a certain understanding of markets. Markets should subsequently not be seen as entities nor places but as complex relations between human actors. Further......, the establishment, maintenance and development of markets are depending on the capacity of the actors to enter into continuous and enhancing interplay....

  19. Lifting to cluster-tilting objects in higher cluster categories

    OpenAIRE

    Liu, Pin

    2008-01-01

    In this note, we consider the $d$-cluster-tilted algebras, the endomorphism algebras of $d$-cluster-tilting objects in $d$-cluster categories. We show that a tilting module over such an algebra lifts to a $d$-cluster-tilting object in this $d$-cluster category.

  20. Enriching the hydrogen storage capacity of carbon nanotube doped with polylithiated molecules

    Science.gov (United States)

    Panigrahi, P.; Naqvi, S. R.; Hankel, M.; Ahuja, R.; Hussain, T.

    2018-06-01

    In a quest to find optimum materials for efficient storage of clean energy, we have performed first principles calculations to study the structural and energy storage properties of one-dimensional carbon nanotubes (CNTs) functionalized with polylithiated molecules (PLMs). Van der Waals corrected calculations disclosed that various PLMs like CLi, CLi2, CLi3, OLi, OLi2, OLi3, bind strongly to CNTs even at high doping concentrations ensuring a uniform distribution of dopants without forming clusters. Bader charge analysis reveals that each Li in all the PLMs attains a partial positive charge and transform into Li+ cations. This situation allows multiple H2 molecules adsorbed with each Li+ through the polarization of incident H2 molecules via electrostatic and van der Waals type of interaction. With a maximum doping concentration, that is 3CLi2/3CLi3 and 3OLi2/3OLi3 a maximum of 36 H2 molecules could be adsorbed that corresponds to a reasonably high H2 storage capacity with the adsorption energies in the range of -0.33 to -0.15 eV/H2. This suits the ambient condition applications.

  1. Neurostimulation in cluster headache

    DEFF Research Database (Denmark)

    Pedersen, Jeppe L; Barloese, Mads; Jensen, Rigmor H

    2013-01-01

    PURPOSE OF REVIEW: Neurostimulation has emerged as a viable treatment for intractable chronic cluster headache. Several therapeutic strategies are being investigated including stimulation of the hypothalamus, occipital nerves and sphenopalatine ganglion. The aim of this review is to provide...... effective strategy must be preferred as first-line therapy for intractable chronic cluster headache....

  2. Cauchy cluster process

    DEFF Research Database (Denmark)

    Ghorbani, Mohammad

    2013-01-01

    In this paper we introduce an instance of the well-know Neyman–Scott cluster process model with clusters having a long tail behaviour. In our model the offspring points are distributed around the parent points according to a circular Cauchy distribution. Using a modified Cramér-von Misses test...

  3. When Clusters become Networks

    NARCIS (Netherlands)

    S.M.W. Phlippen (Sandra); G.A. van der Knaap (Bert)

    2007-01-01

    textabstractPolicy makers spend large amounts of public resources on the foundation of science parks and other forms of geographically clustered business activities, in order to stimulate regional innovation. Underlying the relation between clusters and innovation is the assumption that co-located

  4. Mixed-Initiative Clustering

    Science.gov (United States)

    Huang, Yifen

    2010-01-01

    Mixed-initiative clustering is a task where a user and a machine work collaboratively to analyze a large set of documents. We hypothesize that a user and a machine can both learn better clustering models through enriched communication and interactive learning from each other. The first contribution or this thesis is providing a framework of…

  5. Coma cluster of galaxies

    Science.gov (United States)

    1999-01-01

    Atlas Image mosaic, covering 34' x 34' on the sky, of the Coma cluster, aka Abell 1656. This is a particularly rich cluster of individual galaxies (over 1000 members), most prominently the two giant ellipticals, NGC 4874 (right) and NGC 4889 (left). The remaining members are mostly smaller ellipticals, but spiral galaxies are also evident in the 2MASS image. The cluster is seen toward the constellation Coma Berenices, but is actually at a distance of about 100 Mpc (330 million light years, or a redshift of 0.023) from us. At this distance, the cluster is in what is known as the 'Hubble flow,' or the overall expansion of the Universe. As such, astronomers can measure the Hubble Constant, or the universal expansion rate, based on the distance to this cluster. Large, rich clusters, such as Coma, allow astronomers to measure the 'missing mass,' i.e., the matter in the cluster that we cannot see, since it gravitationally influences the motions of the member galaxies within the cluster. The near-infrared maps the overall luminous mass content of the member galaxies, since the light at these wavelengths is dominated by the more numerous older stellar populations. Galaxies, as seen by 2MASS, look fairly smooth and homogeneous, as can be seen from the Hubble 'tuning fork' diagram of near-infrared galaxy morphology. Image mosaic by S. Van Dyk (IPAC).

  6. Cluster growth kinetics

    International Nuclear Information System (INIS)

    Dubovik, V.M.; Gal'perin, A.G.; Rikhvitskij, V.S.; Lushnikov, A.A.

    2000-01-01

    Processes of some traffic blocking coming into existence are considered as probabilistic ones. We study analytic solutions for models for the dynamics of both cluster growth and cluster growth with fragmentation in the systems of finite number of objects. Assuming rates constancy of both coalescence and fragmentation, the models under consideration are linear on the probability functions

  7. Alpha clustering in nuclei

    International Nuclear Information System (INIS)

    Hodgson, P.E.

    1990-01-01

    The effects of nucleon clustering in nuclei are described, with reference to both nuclear structure and nuclear reactions, and the advantages of using the cluster formalism to describe a range of phenomena are discussed. It is shown that bound and scattering alpha-particle states can be described in a unified way using an energy-dependent alpha-nucleus potential. (author)

  8. Capacity factors of a mixed speed railway network

    DEFF Research Database (Denmark)

    Harrod, Steven

    2009-01-01

    Fifty-four combinations of track network and speed differential are evaluated within a linear, discrete time network model that maximizes an objective function of train volume, delays, and idle train time. The results contradict accepted dispatching practice by suggesting that when introducing...... a priority, high-speed train onto a network, maximum network now is attained when the priority train operates at maximum speed. in addition, increasing siding capacity at meeting points may offer a network capacity improvement comparable to partial double track. (C) 2009 Elsevier Ltd. All rights reserved....

  9. Negotiating Cluster Boundaries

    DEFF Research Database (Denmark)

    Giacomin, Valeria

    2017-01-01

    Palm oil was introduced to Malay(si)a as an alternative to natural rubber, inheriting its cluster organizational structure. In the late 1960s, Malaysia became the world’s largest palm oil exporter. Based on archival material from British colonial institutions and agency houses, this paper focuses...... on the governance dynamics that drove institutional change within this cluster during decolonization. The analysis presents three main findings: (i) cluster boundaries are defined by continuous tug-of-war style negotiations between public and private actors; (ii) this interaction produces institutional change...... within the cluster, in the form of cumulative ‘institutional rounds’ – the correction or disruption of existing institutions or the creation of new ones; and (iii) this process leads to a broader inclusion of local actors in the original cluster configuration. The paper challenges the prevalent argument...

  10. Mathematical classification and clustering

    CERN Document Server

    Mirkin, Boris

    1996-01-01

    I am very happy to have this opportunity to present the work of Boris Mirkin, a distinguished Russian scholar in the areas of data analysis and decision making methodologies. The monograph is devoted entirely to clustering, a discipline dispersed through many theoretical and application areas, from mathematical statistics and combina­ torial optimization to biology, sociology and organizational structures. It compiles an immense amount of research done to date, including many original Russian de­ velopments never presented to the international community before (for instance, cluster-by-cluster versions of the K-Means method in Chapter 4 or uniform par­ titioning in Chapter 5). The author's approach, approximation clustering, allows him both to systematize a great part of the discipline and to develop many in­ novative methods in the framework of optimization problems. The optimization methods considered are proved to be meaningful in the contexts of data analysis and clustering. The material presented in ...

  11. Neutrosophic Hierarchical Clustering Algoritms

    Directory of Open Access Journals (Sweden)

    Rıdvan Şahin

    2014-03-01

    Full Text Available Interval neutrosophic set (INS is a generalization of interval valued intuitionistic fuzzy set (IVIFS, whose the membership and non-membership values of elements consist of fuzzy range, while single valued neutrosophic set (SVNS is regarded as extension of intuitionistic fuzzy set (IFS. In this paper, we extend the hierarchical clustering techniques proposed for IFSs and IVIFSs to SVNSs and INSs respectively. Based on the traditional hierarchical clustering procedure, the single valued neutrosophic aggregation operator, and the basic distance measures between SVNSs, we define a single valued neutrosophic hierarchical clustering algorithm for clustering SVNSs. Then we extend the algorithm to classify an interval neutrosophic data. Finally, we present some numerical examples in order to show the effectiveness and availability of the developed clustering algorithms.

  12. Maximum entropy decomposition of quadrupole mass spectra

    International Nuclear Information System (INIS)

    Toussaint, U. von; Dose, V.; Golan, A.

    2004-01-01

    We present an information-theoretic method called generalized maximum entropy (GME) for decomposing mass spectra of gas mixtures from noisy measurements. In this GME approach to the noisy, underdetermined inverse problem, the joint entropies of concentration, cracking, and noise probabilities are maximized subject to the measured data. This provides a robust estimation for the unknown cracking patterns and the concentrations of the contributing molecules. The method is applied to mass spectroscopic data of hydrocarbons, and the estimates are compared with those received from a Bayesian approach. We show that the GME method is efficient and is computationally fast

  13. Maximum power operation of interacting molecular motors

    DEFF Research Database (Denmark)

    Golubeva, Natalia; Imparato, Alberto

    2013-01-01

    , as compared to the non-interacting system, in a wide range of biologically compatible scenarios. We furthermore consider the case where the motor-motor interaction directly affects the internal chemical cycle and investigate the effect on the system dynamics and thermodynamics.......We study the mechanical and thermodynamic properties of different traffic models for kinesin which are relevant in biological and experimental contexts. We find that motor-motor interactions play a fundamental role by enhancing the thermodynamic efficiency at maximum power of the motors...

  14. Maximum entropy method in momentum density reconstruction

    International Nuclear Information System (INIS)

    Dobrzynski, L.; Holas, A.

    1997-01-01

    The Maximum Entropy Method (MEM) is applied to the reconstruction of the 3-dimensional electron momentum density distributions observed through the set of Compton profiles measured along various crystallographic directions. It is shown that the reconstruction of electron momentum density may be reliably carried out with the aid of simple iterative algorithm suggested originally by Collins. A number of distributions has been simulated in order to check the performance of MEM. It is shown that MEM can be recommended as a model-free approach. (author). 13 refs, 1 fig

  15. On the maximum drawdown during speculative bubbles

    Science.gov (United States)

    Rotundo, Giulia; Navarra, Mauro

    2007-08-01

    A taxonomy of large financial crashes proposed in the literature locates the burst of speculative bubbles due to endogenous causes in the framework of extreme stock market crashes, defined as falls of market prices that are outlier with respect to the bulk of drawdown price movement distribution. This paper goes on deeper in the analysis providing a further characterization of the rising part of such selected bubbles through the examination of drawdown and maximum drawdown movement of indices prices. The analysis of drawdown duration is also performed and it is the core of the risk measure estimated here.

  16. Multi-Channel Maximum Likelihood Pitch Estimation

    DEFF Research Database (Denmark)

    Christensen, Mads Græsbøll

    2012-01-01

    In this paper, a method for multi-channel pitch estimation is proposed. The method is a maximum likelihood estimator and is based on a parametric model where the signals in the various channels share the same fundamental frequency but can have different amplitudes, phases, and noise characteristics....... This essentially means that the model allows for different conditions in the various channels, like different signal-to-noise ratios, microphone characteristics and reverberation. Moreover, the method does not assume that a certain array structure is used but rather relies on a more general model and is hence...

  17. Conductivity maximum in a charged colloidal suspension

    Energy Technology Data Exchange (ETDEWEB)

    Bastea, S

    2009-01-27

    Molecular dynamics simulations of a charged colloidal suspension in the salt-free regime show that the system exhibits an electrical conductivity maximum as a function of colloid charge. We attribute this behavior to two main competing effects: colloid effective charge saturation due to counterion 'condensation' and diffusion slowdown due to the relaxation effect. In agreement with previous observations, we also find that the effective transported charge is larger than the one determined by the Stern layer and suggest that it corresponds to the boundary fluid layer at the surface of the colloidal particles.

  18. Dynamical maximum entropy approach to flocking.

    Science.gov (United States)

    Cavagna, Andrea; Giardina, Irene; Ginelli, Francesco; Mora, Thierry; Piovani, Duccio; Tavarone, Raffaele; Walczak, Aleksandra M

    2014-04-01

    We derive a new method to infer from data the out-of-equilibrium alignment dynamics of collectively moving animal groups, by considering the maximum entropy model distribution consistent with temporal and spatial correlations of flight direction. When bird neighborhoods evolve rapidly, this dynamical inference correctly learns the parameters of the model, while a static one relying only on the spatial correlations fails. When neighbors change slowly and the detailed balance is satisfied, we recover the static procedure. We demonstrate the validity of the method on simulated data. The approach is applicable to other systems of active matter.

  19. Maximum Temperature Detection System for Integrated Circuits

    Science.gov (United States)

    Frankiewicz, Maciej; Kos, Andrzej

    2015-03-01

    The paper describes structure and measurement results of the system detecting present maximum temperature on the surface of an integrated circuit. The system consists of the set of proportional to absolute temperature sensors, temperature processing path and a digital part designed in VHDL. Analogue parts of the circuit where designed with full-custom technique. The system is a part of temperature-controlled oscillator circuit - a power management system based on dynamic frequency scaling method. The oscillator cooperates with microprocessor dedicated for thermal experiments. The whole system is implemented in UMC CMOS 0.18 μm (1.8 V) technology.

  20. Maximum entropy PDF projection: A review

    Science.gov (United States)

    Baggenstoss, Paul M.

    2017-06-01

    We review maximum entropy (MaxEnt) PDF projection, a method with wide potential applications in statistical inference. The method constructs a sampling distribution for a high-dimensional vector x based on knowing the sampling distribution p(z) of a lower-dimensional feature z = T (x). Under mild conditions, the distribution p(x) having highest possible entropy among all distributions consistent with p(z) may be readily found. Furthermore, the MaxEnt p(x) may be sampled, making the approach useful in Monte Carlo methods. We review the theorem and present a case study in model order selection and classification for handwritten character recognition.

  1. Multiperiod Maximum Loss is time unit invariant.

    Science.gov (United States)

    Kovacevic, Raimund M; Breuer, Thomas

    2016-01-01

    Time unit invariance is introduced as an additional requirement for multiperiod risk measures: for a constant portfolio under an i.i.d. risk factor process, the multiperiod risk should equal the one period risk of the aggregated loss, for an appropriate choice of parameters and independent of the portfolio and its distribution. Multiperiod Maximum Loss over a sequence of Kullback-Leibler balls is time unit invariant. This is also the case for the entropic risk measure. On the other hand, multiperiod Value at Risk and multiperiod Expected Shortfall are not time unit invariant.

  2. Maximum a posteriori decoder for digital communications

    Science.gov (United States)

    Altes, Richard A. (Inventor)

    1997-01-01

    A system and method for decoding by identification of the most likely phase coded signal corresponding to received data. The present invention has particular application to communication with signals that experience spurious random phase perturbations. The generalized estimator-correlator uses a maximum a posteriori (MAP) estimator to generate phase estimates for correlation with incoming data samples and for correlation with mean phases indicative of unique hypothesized signals. The result is a MAP likelihood statistic for each hypothesized transmission, wherein the highest value statistic identifies the transmitted signal.

  3. Improved Maximum Parsimony Models for Phylogenetic Networks.

    Science.gov (United States)

    Van Iersel, Leo; Jones, Mark; Scornavacca, Celine

    2018-05-01

    Phylogenetic networks are well suited to represent evolutionary histories comprising reticulate evolution. Several methods aiming at reconstructing explicit phylogenetic networks have been developed in the last two decades. In this article, we propose a new definition of maximum parsimony for phylogenetic networks that permits to model biological scenarios that cannot be modeled by the definitions currently present in the literature (namely, the "hardwired" and "softwired" parsimony). Building on this new definition, we provide several algorithmic results that lay the foundations for new parsimony-based methods for phylogenetic network reconstruction.

  4. Ancestral sequence reconstruction with Maximum Parsimony

    OpenAIRE

    Herbst, Lina; Fischer, Mareike

    2017-01-01

    One of the main aims in phylogenetics is the estimation of ancestral sequences based on present-day data like, for instance, DNA alignments. One way to estimate the data of the last common ancestor of a given set of species is to first reconstruct a phylogenetic tree with some tree inference method and then to use some method of ancestral state inference based on that tree. One of the best-known methods both for tree inference as well as for ancestral sequence inference is Maximum Parsimony (...

  5. Neuronal discrimination capacity

    International Nuclear Information System (INIS)

    Deng Yingchun; Williams, Peter; Feng Jianfeng; Liu Feng

    2003-01-01

    We explore neuronal mechanisms of discriminating between masked signals. It is found that when the correlation between input signals is zero, the output signals are separable if and only if input signals are separable. With positively (negatively) correlated signals, the output signals are separable (mixed) even when input signals are mixed (separable). Exact values of discrimination capacity are obtained for two most interesting cases: the exactly balanced inhibitory and excitatory input case and the uncorrelated input case. Interestingly, the discrimination capacity obtained in these cases is independent of model parameters, input distribution and is universal. Our results also suggest a functional role of inhibitory inputs and correlated inputs or, more generally, the large variability of efferent spike trains observed in in vivo experiments: the larger the variability of efferent spike trains, the easier it is to discriminate between masked input signals

  6. Seismic capacity of switchgear

    International Nuclear Information System (INIS)

    Bandyopadhyay, K.; Hofmayer, C.; Kassir, M.; Pepper, S.

    1989-01-01

    As part of a component fragility program sponsored by the USNRC, BNL has collected existing information on the seismic capacity of switchgear assemblies from major manufacturers. Existing seismic test data for both low and medium voltage switchgear assemblies have been evaluated and the generic results are presented in this paper. The failure modes are identified and the corresponding generic lower bound capacity levels are established. The test response spectra have been used as a measure of the test vibration input. The results indicate that relays chatter at a very low input level at the base of the switchgear cabinet. This change of state of devices including relays have been observed. Breaker tripping occurs at a higher vibration level. Although the structural failure of internal elements have been noticed, the overall switchgear cabinet structure withstands a high vibration level. 5 refs., 2 figs., 2 tabs

  7. Neuronal discrimination capacity

    Energy Technology Data Exchange (ETDEWEB)

    Deng Yingchun [Department of Mathematics, Hunan Normal University 410081, Changsha (China); COGS, University of Sussex at Brighton, BN1 9QH (United Kingdom); Williams, Peter; Feng Jianfeng [COGS, University of Sussex at Brighton, BN1 9QH (United Kingdom); Liu Feng [COGS, University of Sussex at Brighton, BN1 9QH (United Kingdom); Physics Department, Nanjing University (China)

    2003-12-19

    We explore neuronal mechanisms of discriminating between masked signals. It is found that when the correlation between input signals is zero, the output signals are separable if and only if input signals are separable. With positively (negatively) correlated signals, the output signals are separable (mixed) even when input signals are mixed (separable). Exact values of discrimination capacity are obtained for two most interesting cases: the exactly balanced inhibitory and excitatory input case and the uncorrelated input case. Interestingly, the discrimination capacity obtained in these cases is independent of model parameters, input distribution and is universal. Our results also suggest a functional role of inhibitory inputs and correlated inputs or, more generally, the large variability of efferent spike trains observed in in vivo experiments: the larger the variability of efferent spike trains, the easier it is to discriminate between masked input signals.

  8. Artificial Neural Network In Maximum Power Point Tracking Algorithm Of Photovoltaic Systems

    Directory of Open Access Journals (Sweden)

    Modestas Pikutis

    2014-05-01

    Full Text Available Scientists are looking for ways to improve the efficiency of solar cells all the time. The efficiency of solar cells which are available to the general public is up to 20%. Part of the solar energy is unused and a capacity of solar power plant is significantly reduced – if slow controller or controller which cannot stay at maximum power point of solar modules is used. Various algorithms of maximum power point tracking were created, but mostly algorithms are slow or make mistakes. In the literature more and more oftenartificial neural networks (ANN in maximum power point tracking process are mentioned, in order to improve performance of the controller. Self-learner artificial neural network and IncCond algorithm were used for maximum power point tracking in created solar power plant model. The algorithm for control was created. Solar power plant model is implemented in Matlab/Simulink environment.

  9. PAYMENT CAPACITY SENSITIVITY FACTORS

    OpenAIRE

    Daniel BRÎNDESCU – OLARIU

    2014-01-01

    The current study focuses on the sensitivity of the corporate payment capacity. Through the nature of the subject, the research is based on simulating variations of the forecasted cash-flows of the companies included in the sample. The study employs 391 forecasted yearly cash-flows statements collected from 50 companies from Timis County (Romania), as well as the detailed hypotheses of the forecasts. The results of the study facilitate the determination and classification of the main se...

  10. Options on capacity imbalance

    International Nuclear Information System (INIS)

    Roggen, M.

    2002-01-01

    Since the start of this year, the Dutch energy company Nuon has been using a computer system to formulate real-time responses to national capacity imbalances in the electricity supply market. The work earns Nuon a fixed fee from TenneT (Dutch Transmission System Operator) and ensures a more stable imbalance price for everyone. The key to success has been the decision to start the project from scratch [nl

  11. Herd Clustering: A synergistic data clustering approach using collective intelligence

    KAUST Repository

    Wong, Kachun; Peng, Chengbin; Li, Yue; Chan, Takming

    2014-01-01

    , this principle is used to develop a new clustering algorithm. Inspired by herd behavior, the clustering method is a synergistic approach using collective intelligence called Herd Clustering (HC). The novel part is laid in its first stage where data instances

  12. CSTI High Capacity Power

    International Nuclear Information System (INIS)

    Winter, J.M.

    1989-01-01

    The SP-100 program was established in 1983 by DOD, DOE, and NASA as a joint program to develop the technology necessary for space nuclear power systems for military and civil application. During FY-86 and 87, the NASA SP-100 Advanced Technology Program was devised to maintain the momentum of promising technology advancement efforts started during Phase 1 of SP-100 and to strengthen, in key areas, the chances for successful development and growth capability of space nuclear reactor power systems for future space applications. In FY-88, the Advanced Technology Program was incorporated into NASA's new Civil Space Technology Initiative (CSTI). The CSTI Program was established to provide the foundation for technology development in automation and robotics, information, propulsion, and power. The CSTI High Capacity Power Program builds on the technology efforts of the SP-100 program, incorporates the previous NASA SP-100 Advanced Technology project, and provides a bridge to NASA Project Pathfinder. The elements of CSTI High Capacity Power development include Conversion Systems, Thermal Management, Power Management, System Diagnostics, and Environmental Interactions. Technology advancement in all areas, including materials, is required to assure the high reliability and 7 to 10 year lifetime demanded for future space nuclear power systems. The overall program will develop and demonstrate the technology base required to provide a wide range of modular power systems as well as allowing mission independence from solar and orbital attitude requirements. Several recent advancements in CSTI High Capacity power development will be discussed

  13. Objective Bayesianism and the Maximum Entropy Principle

    Directory of Open Access Journals (Sweden)

    Jon Williamson

    2013-09-01

    Full Text Available Objective Bayesian epistemology invokes three norms: the strengths of our beliefs should be probabilities; they should be calibrated to our evidence of physical probabilities; and they should otherwise equivocate sufficiently between the basic propositions that we can express. The three norms are sometimes explicated by appealing to the maximum entropy principle, which says that a belief function should be a probability function, from all those that are calibrated to evidence, that has maximum entropy. However, the three norms of objective Bayesianism are usually justified in different ways. In this paper, we show that the three norms can all be subsumed under a single justification in terms of minimising worst-case expected loss. This, in turn, is equivalent to maximising a generalised notion of entropy. We suggest that requiring language invariance, in addition to minimising worst-case expected loss, motivates maximisation of standard entropy as opposed to maximisation of other instances of generalised entropy. Our argument also provides a qualified justification for updating degrees of belief by Bayesian conditionalisation. However, conditional probabilities play a less central part in the objective Bayesian account than they do under the subjective view of Bayesianism, leading to a reduced role for Bayes’ Theorem.

  14. Competence building capacity shortage

    International Nuclear Information System (INIS)

    Doorman, Gerard; Wangensteen, Ivar; Bakken, Bjoern

    2005-02-01

    The objective of the project 'Competence Building Capacity Shortage' has been 'to increase knowledge about central approaches aimed at solving the peaking capacity problem in restructured power systems'. With respect to reserve markets, a model was developed in the project to analyze the relations between reserve requirements and prices in the spot and reserve markets respectively. A mathematical model was also developed and implemented, which also includes the balance market, and has a good ability to predict the relations between these markets under various assumptions. With some further development, this model can be used fore realistic analyses of these markets in a Nordic context. It was also concluded that certain system requirements with respect to frequency and time deviation can be relaxed without adverse effects. However, the requirements to system bias, Frequency Activated Operating Reserves and Frequency Activated Contingency Reserves cannot be relaxed, the latter because they must cover the dimensioning fault in the system. On the other hand, Fast Contingency Reserves can be reduced by removing requirements to national balances. Costs can furthermore be reduced by increasingly adapting a Nordic as opposed to national approach. A model for stepwise power flow was developed in the project, which is especially useful to analyze slow power system dynamics. This is relevant when analysing the effects of reserve requirements. A model for the analysis of the capacity balance in Norway and Sweden was also developed. This model is useful for looking at the future balance under various assumptions regarding e.g. weather conditions, demand growth and the development of the generation system. With respect to the present situation, if there is some price flexibility on the demand side and system operators are able to use reserves from the demand side, the probability for load shedding during the peak load hour is close to zero under the weather conditions after

  15. The self limiting effect of hydrogen cluster in gas jet under liquid nitrogen temperature

    International Nuclear Information System (INIS)

    Han Jifeng; Yang Chaowen; Miao Jingwei; Fu Pengtao; Luo Xiaobing; Shi Miangong

    2010-01-01

    The generation of hydrogen clusters in gas jet is tested using the Rayleigh scattering method under liquid nitrogen temperature of 79 K. The self limiting effect of hydrogen cluster is studied and it is found that the cluster formation is greatly affected by the number of expanded molecules. The well designed liquid nitrogen cold trap ensured that the hydrogen cluster would keep maximum size for maximum 15 ms during one gas jet. The scattered light intensity exhibits a power scaling on the backing pressure ranging from 5 to 48 bar with the power value of 4.1.

  16. Document clustering methods, document cluster label disambiguation methods, document clustering apparatuses, and articles of manufacture

    Science.gov (United States)

    Sanfilippo, Antonio [Richland, WA; Calapristi, Augustin J [West Richland, WA; Crow, Vernon L [Richland, WA; Hetzler, Elizabeth G [Kennewick, WA; Turner, Alan E [Kennewick, WA

    2009-12-22

    Document clustering methods, document cluster label disambiguation methods, document clustering apparatuses, and articles of manufacture are described. In one aspect, a document clustering method includes providing a document set comprising a plurality of documents, providing a cluster comprising a subset of the documents of the document set, using a plurality of terms of the documents, providing a cluster label indicative of subject matter content of the documents of the cluster, wherein the cluster label comprises a plurality of word senses, and selecting one of the word senses of the cluster label.

  17. Cluster-cluster correlations and constraints on the correlation hierarchy

    Science.gov (United States)

    Hamilton, A. J. S.; Gott, J. R., III

    1988-01-01

    The hypothesis that galaxies cluster around clusters at least as strongly as they cluster around galaxies imposes constraints on the hierarchy of correlation amplitudes in hierachical clustering models. The distributions which saturate these constraints are the Rayleigh-Levy random walk fractals proposed by Mandelbrot; for these fractal distributions cluster-cluster correlations are all identically equal to galaxy-galaxy correlations. If correlation amplitudes exceed the constraints, as is observed, then cluster-cluster correlations must exceed galaxy-galaxy correlations, as is observed.

  18. Formation of stable products from cluster-cluster collisions

    International Nuclear Information System (INIS)

    Alamanova, Denitsa; Grigoryan, Valeri G; Springborg, Michael

    2007-01-01

    The formation of stable products from copper cluster-cluster collisions is investigated by using classical molecular-dynamics simulations in combination with an embedded-atom potential. The dependence of the product clusters on impact energy, relative orientation of the clusters, and size of the clusters is studied. The structures and total energies of the product clusters are analysed and compared with those of the colliding clusters before impact. These results, together with the internal temperature, are used in obtaining an increased understanding of cluster fusion processes

  19. Fitting Latent Cluster Models for Networks with latentnet

    Directory of Open Access Journals (Sweden)

    Pavel N. Krivitsky

    2007-12-01

    Full Text Available latentnet is a package to fit and evaluate statistical latent position and cluster models for networks. Hoff, Raftery, and Handcock (2002 suggested an approach to modeling networks based on positing the existence of an latent space of characteristics of the actors. Relationships form as a function of distances between these characteristics as well as functions of observed dyadic level covariates. In latentnet social distances are represented in a Euclidean space. It also includes a variant of the extension of the latent position model to allow for clustering of the positions developed in Handcock, Raftery, and Tantrum (2007.The package implements Bayesian inference for the models based on an Markov chain Monte Carlo algorithm. It can also compute maximum likelihood estimates for the latent position model and a two-stage maximum likelihood method for the latent position cluster model. For latent position cluster models, the package provides a Bayesian way of assessing how many groups there are, and thus whether or not there is any clustering (since if the preferred number of groups is 1, there is little evidence for clustering. It also estimates which cluster each actor belongs to. These estimates are probabilistic, and provide the probability of each actor belonging to each cluster. It computes four types of point estimates for the coefficients and positions: maximum likelihood estimate, posterior mean, posterior mode and the estimator which minimizes Kullback-Leibler divergence from the posterior. You can assess the goodness-of-fit of the model via posterior predictive checks. It has a function to simulate networks from a latent position or latent position cluster model.

  20. Tune Your Brown Clustering, Please

    DEFF Research Database (Denmark)

    Derczynski, Leon; Chester, Sean; Bøgh, Kenneth Sejdenfaden

    2015-01-01

    Brown clustering, an unsupervised hierarchical clustering technique based on ngram mutual information, has proven useful in many NLP applications. However, most uses of Brown clustering employ the same default configuration; the appropriateness of this configuration has gone predominantly...

  1. PERHITUNGAN IDLE CAPACITY DENGAN MENGGUNAKAN CAM-I CAPACITY MODEL DALAM RANGKA EFISIENSI BIAYA PADA PT X

    Directory of Open Access Journals (Sweden)

    Muammar Aditya

    2015-09-01

    Full Text Available Aim for this research are to analyze capacity cost which incure from company production machines and human resources whose operate the production machine using CAM-I capacity model. CAM-I capacity model is an approach which focus  upon how to manage company resources. This research initiated at PT X which focus to production activity that used small mixer machine, extruder machine, oven drying machine, enrober machine, pan coting machine which consist of hot and cold pan coating machine, and packing machine which consist of vertical packing machine and horizontal packing machine as well as human resources that operates those machine. This research focus on rate capacity, productive capacity, idle capacity, and nonproductive capacity to measure capacity cost. Result of this research shows most of the capacity owned by either by production machine or human resources are not utilized to its maximum potential. There are need to reduce capacity cost owned by production machine and human resoures to increase the product sales but if its unachieveable there will be need to increase efficiency from production machine and human resources by reducing their quantityDOI: 10.15408/ess.v4i1.1961

  2. Building a cluster computer for the computing grid of tomorrow

    International Nuclear Information System (INIS)

    Wezel, J. van; Marten, H.

    2004-01-01

    The Grid Computing Centre Karlsruhe takes part in the development, test and deployment of hardware and cluster infrastructure, grid computing middleware, and applications for particle physics. The construction of a large cluster computer with thousands of nodes and several PB data storage capacity is a major task and focus of research. CERN based accelerator experiments will use GridKa, one of only 8 world wide Tier-1 computing centers, for its huge computer demands. Computing and storage is provided already for several other running physics experiments on the exponentially expanding cluster. (orig.)

  3. Hydraulic Limits on Maximum Plant Transpiration

    Science.gov (United States)

    Manzoni, S.; Vico, G.; Katul, G. G.; Palmroth, S.; Jackson, R. B.; Porporato, A. M.

    2011-12-01

    Photosynthesis occurs at the expense of water losses through transpiration. As a consequence of this basic carbon-water interaction at the leaf level, plant growth and ecosystem carbon exchanges are tightly coupled to transpiration. In this contribution, the hydraulic constraints that limit transpiration rates under well-watered conditions are examined across plant functional types and climates. The potential water flow through plants is proportional to both xylem hydraulic conductivity (which depends on plant carbon economy) and the difference in water potential between the soil and the atmosphere (the driving force that pulls water from the soil). Differently from previous works, we study how this potential flux changes with the amplitude of the driving force (i.e., we focus on xylem properties and not on stomatal regulation). Xylem hydraulic conductivity decreases as the driving force increases due to cavitation of the tissues. As a result of this negative feedback, more negative leaf (and xylem) water potentials would provide a stronger driving force for water transport, while at the same time limiting xylem hydraulic conductivity due to cavitation. Here, the leaf water potential value that allows an optimum balance between driving force and xylem conductivity is quantified, thus defining the maximum transpiration rate that can be sustained by the soil-to-leaf hydraulic system. To apply the proposed framework at the global scale, a novel database of xylem conductivity and cavitation vulnerability across plant types and biomes is developed. Conductivity and water potential at 50% cavitation are shown to be complementary (in particular between angiosperms and conifers), suggesting a tradeoff between transport efficiency and hydraulic safety. Plants from warmer and drier biomes tend to achieve larger maximum transpiration than plants growing in environments with lower atmospheric water demand. The predicted maximum transpiration and the corresponding leaf water

  4. Cluster Management Institutionalization

    DEFF Research Database (Denmark)

    Normann, Leo; Agger Nielsen, Jeppe

    2015-01-01

    of how it was legitimized as a “ready-to-use” management model. Further, our account reveals how cluster management translated into considerably different local variants as it travelled into specific organizations. However, these processes have not occurred sequentially with cluster management first...... legitimized at the field level, then spread, and finally translated into action in the adopting organizations. Instead, we observed entangled field and organizational-level processes. Accordingly, we argue that cluster management institutionalization is most readily understood by simultaneously investigating...

  5. The concept of cluster

    DEFF Research Database (Denmark)

    Laursen, Lea Louise Holst; Møller, Jørgen

    2013-01-01

    villages in order to secure their future. This paper will address the concept of cluster-villages as a possible approach to strengthen the conditions of contemporary Danish villages. Cluster-villages is a concept that gather a number of villages in a network-structure where the villages both work together...... to forskellige positioner ser vi en ny mulighed for landsbyudvikling, som vi kalder Clustervillages. In order to investigate the potentials and possibilities of the cluster-village concept the paper will seek to unfold the concept strategically; looking into the benefits of such concept. Further, the paper seeks...

  6. Raspberry Pi super cluster

    CERN Document Server

    Dennis, Andrew K

    2013-01-01

    This book follows a step-by-step, tutorial-based approach which will teach you how to develop your own super cluster using Raspberry Pi computers quickly and efficiently.Raspberry Pi Super Cluster is an introductory guide for those interested in experimenting with parallel computing at home. Aimed at Raspberry Pi enthusiasts, this book is a primer for getting your first cluster up and running.Basic knowledge of C or Java would be helpful but no prior knowledge of parallel computing is necessary.

  7. Introduction to cluster dynamics

    CERN Document Server

    Reinhard, Paul-Gerhard

    2008-01-01

    Clusters as mesoscopic particles represent an intermediate state of matter between single atoms and solid material. The tendency to miniaturise technical objects requires knowledge about systems which contain a ""small"" number of atoms or molecules only. This is all the more true for dynamical aspects, particularly in relation to the qick development of laser technology and femtosecond spectroscopy. Here, for the first time is a highly qualitative introduction to cluster physics. With its emphasis on cluster dynamics, this will be vital to everyone involved in this interdisciplinary subje

  8. Contextualizing the Cluster

    DEFF Research Database (Denmark)

    Giacomin, Valeria

    This dissertation examines the case of the palm oil cluster in Malaysia and Indonesia, today one of the largest agricultural clusters in the world. My analysis focuses on the evolution of the cluster from the 1880s to the 1970s in order to understand how it helped these two countries to integrate...... into the global economy in both colonial and post-colonial times. The study is based on empirical material drawn from five UK archives and background research using secondary sources, interviews, and archive visits to Malaysia and Singapore. The dissertation comprises three articles, each discussing a major under...

  9. Atomic cluster collisions

    Science.gov (United States)

    Korol, Andrey V.; Solov'yov, Andrey

    2013-01-01

    Atomic cluster collisions are a field of rapidly emerging research interest by both experimentalists and theorists. The international symposium on atomic cluster collisions (ISSAC) is the premier forum to present cutting-edge research in this field. It was established in 2003 and the most recent conference was held in Berlin, Germany in July of 2011. This Topical Issue presents original research results from some of the participants, who attended this conference. This issues specifically focuses on two research areas, namely Clusters and Fullerenes in External Fields and Nanoscale Insights in Radiation Biodamage.

  10. Analogue of Pontryagin's maximum principle for multiple integrals minimization problems

    OpenAIRE

    Mikhail, Zelikin

    2016-01-01

    The theorem like Pontryagin's maximum principle for multiple integrals is proved. Unlike the usual maximum principle, the maximum should be taken not over all matrices, but only on matrices of rank one. Examples are given.

  11. Lake Basin Fetch and Maximum Length/Width

    Data.gov (United States)

    Minnesota Department of Natural Resources — Linear features representing the Fetch, Maximum Length and Maximum Width of a lake basin. Fetch, maximum length and average width are calcuated from the lake polygon...

  12. Combining cluster number counts and galaxy clustering

    Energy Technology Data Exchange (ETDEWEB)

    Lacasa, Fabien; Rosenfeld, Rogerio, E-mail: fabien@ift.unesp.br, E-mail: rosenfel@ift.unesp.br [ICTP South American Institute for Fundamental Research, Instituto de Física Teórica, Universidade Estadual Paulista, São Paulo (Brazil)

    2016-08-01

    The abundance of clusters and the clustering of galaxies are two of the important cosmological probes for current and future large scale surveys of galaxies, such as the Dark Energy Survey. In order to combine them one has to account for the fact that they are not independent quantities, since they probe the same density field. It is important to develop a good understanding of their correlation in order to extract parameter constraints. We present a detailed modelling of the joint covariance matrix between cluster number counts and the galaxy angular power spectrum. We employ the framework of the halo model complemented by a Halo Occupation Distribution model (HOD). We demonstrate the importance of accounting for non-Gaussianity to produce accurate covariance predictions. Indeed, we show that the non-Gaussian covariance becomes dominant at small scales, low redshifts or high cluster masses. We discuss in particular the case of the super-sample covariance (SSC), including the effects of galaxy shot-noise, halo second order bias and non-local bias. We demonstrate that the SSC obeys mathematical inequalities and positivity. Using the joint covariance matrix and a Fisher matrix methodology, we examine the prospects of combining these two probes to constrain cosmological and HOD parameters. We find that the combination indeed results in noticeably better constraints, with improvements of order 20% on cosmological parameters compared to the best single probe, and even greater improvement on HOD parameters, with reduction of error bars by a factor 1.4-4.8. This happens in particular because the cross-covariance introduces a synergy between the probes on small scales. We conclude that accounting for non-Gaussian effects is required for the joint analysis of these observables in galaxy surveys.

  13. Inhibition delay increases neural network capacity through Stirling transform

    Science.gov (United States)

    Nogaret, Alain; King, Alastair

    2018-03-01

    Inhibitory neural networks are found to encode high volumes of information through delayed inhibition. We show that inhibition delay increases storage capacity through a Stirling transform of the minimum capacity which stabilizes locally coherent oscillations. We obtain both the exact and asymptotic formulas for the total number of dynamic attractors. Our results predict a (ln2) -N-fold increase in capacity for an N -neuron network and demonstrate high-density associative memories which host a maximum number of oscillations in analog neural devices.

  14. Near infrared photometry of globular clusters

    International Nuclear Information System (INIS)

    Evans, T.L.; Menzies, J.W.

    1977-01-01

    Photographic photometry on the V, Isub(K) system has been obtained for giant stars in the metal-rich globular clusters NGC 5927, 6171, 6352, 6356, 6388, 6522, 6528, 6712 and 6723. Colour-magnitude diagrams are presented. These data, with earlier observations of NGC 104 (47 Tuc), yield new parameters to describe the giant branch. These are the colour of the red variables, represented by their mean colour (V - Isub(K)) 0 or by the colour (V - Isub(K))sub(BO) of the bluest red variable on the giant branch of a cluster, and ΔV' which is the magnitude difference between the horizontal branch and the highest point on the giant branch. The latter is independent of reddening, since the giant branch of the most metal-rich clusters passes through a maximum in the V, V - Isub(K) plane. These parameters are correlated with the metal content, deduced from integrated photometry: the red variables are redder and the giant branch fainter the higher the metal content. Comparison with theoretical evolutionary tracks suggests that the range in metal content of these clusters is at most a factor of 10, the most metal-rich clusters possibly approaching the solar value. The cluster giant branches and those of open clusters, groups and field stars of the old disk population are compared. The assumption that all the globular clusters have an absolute magnitude on the horizontal branch of Msub(v) = + 0.9, as found recently for 47 Tuc, gives good agreement between the magnitudes of giant stars in the most metal rich of the globular clusters and those of field stars deduced from statistical parallaxes and moving group parallaxes. The values of the parameters ΔV' and (V - Isub(k))sub(BO) also approach those in the moving groups. The globular clusters have a longer horizontal branch, however, and the subgiants are bluer even when the values of ) 7Fe/H{ appear to be the same. (author). )

  15. Maximum Likelihood Reconstruction for Magnetic Resonance Fingerprinting.

    Science.gov (United States)

    Zhao, Bo; Setsompop, Kawin; Ye, Huihui; Cauley, Stephen F; Wald, Lawrence L

    2016-08-01

    This paper introduces a statistical estimation framework for magnetic resonance (MR) fingerprinting, a recently proposed quantitative imaging paradigm. Within this framework, we present a maximum likelihood (ML) formalism to estimate multiple MR tissue parameter maps directly from highly undersampled, noisy k-space data. A novel algorithm, based on variable splitting, the alternating direction method of multipliers, and the variable projection method, is developed to solve the resulting optimization problem. Representative results from both simulations and in vivo experiments demonstrate that the proposed approach yields significantly improved accuracy in parameter estimation, compared to the conventional MR fingerprinting reconstruction. Moreover, the proposed framework provides new theoretical insights into the conventional approach. We show analytically that the conventional approach is an approximation to the ML reconstruction; more precisely, it is exactly equivalent to the first iteration of the proposed algorithm for the ML reconstruction, provided that a gridding reconstruction is used as an initialization.

  16. The worst case complexity of maximum parsimony.

    Science.gov (United States)

    Carmel, Amir; Musa-Lempel, Noa; Tsur, Dekel; Ziv-Ukelson, Michal

    2014-11-01

    One of the core classical problems in computational biology is that of constructing the most parsimonious phylogenetic tree interpreting an input set of sequences from the genomes of evolutionarily related organisms. We reexamine the classical maximum parsimony (MP) optimization problem for the general (asymmetric) scoring matrix case, where rooted phylogenies are implied, and analyze the worst case bounds of three approaches to MP: The approach of Cavalli-Sforza and Edwards, the approach of Hendy and Penny, and a new agglomerative, "bottom-up" approach we present in this article. We show that the second and third approaches are faster than the first one by a factor of Θ(√n) and Θ(n), respectively, where n is the number of species.

  17. Modification of MSDR algorithm and ITS implementation on graph clustering

    Science.gov (United States)

    Prastiwi, D.; Sugeng, K. A.; Siswantining, T.

    2017-07-01

    Maximum Standard Deviation Reduction (MSDR) is a graph clustering algorithm to minimize the distance variation within a cluster. In this paper we propose a modified MSDR by replacing one technical step in MSDR which uses polynomial regression, with a new and simpler step. This leads to our new algorithm called Modified MSDR (MMSDR). We implement the new algorithm to separate a domestic flight network of an Indonesian airline into two large clusters. Further analysis allows us to discover a weak link in the network, which should be improved by adding more flights.

  18. A data-driven approach to estimating the number of clusters in hierarchical clustering [version 1; referees: 2 approved, 1 approved with reservations

    Directory of Open Access Journals (Sweden)

    Antoine E. Zambelli

    2016-12-01

    Full Text Available DNA microarray and gene expression problems often require a researcher to perform clustering on their data in a bid to better understand its structure. In cases where the number of clusters is not known, one can resort to hierarchical clustering methods. However, there currently exist very few automated algorithms for determining the true number of clusters in the data. We propose two new methods (mode and maximum difference for estimating the number of clusters in a hierarchical clustering framework to create a fully automated process with no human intervention. These methods are compared to the established elbow and gap statistic algorithms using simulated datasets and the Biobase Gene ExpressionSet. We also explore a data mixing procedure inspired by cross validation techniques. We find that the overall performance of the maximum difference method is comparable or greater to that of the gap statistic in multi-cluster scenarios, and achieves that performance at a fraction of the computational cost. This method also responds well to our mixing procedure, which opens the door to future research. We conclude that both the mode and maximum difference methods warrant further study related to their mixing and cross-validation potential. We particularly recommend the use of the maximum difference method in multi-cluster scenarios given its accuracy and execution times, and present it as an alternative to existing algorithms.

  19. Modelling maximum likelihood estimation of availability

    International Nuclear Information System (INIS)

    Waller, R.A.; Tietjen, G.L.; Rock, G.W.

    1975-01-01

    Suppose the performance of a nuclear powered electrical generating power plant is continuously monitored to record the sequence of failure and repairs during sustained operation. The purpose of this study is to assess one method of estimating the performance of the power plant when the measure of performance is availability. That is, we determine the probability that the plant is operational at time t. To study the availability of a power plant, we first assume statistical models for the variables, X and Y, which denote the time-to-failure and the time-to-repair variables, respectively. Once those statistical models are specified, the availability, A(t), can be expressed as a function of some or all of their parameters. Usually those parameters are unknown in practice and so A(t) is unknown. This paper discusses the maximum likelihood estimator of A(t) when the time-to-failure model for X is an exponential density with parameter, lambda, and the time-to-repair model for Y is an exponential density with parameter, theta. Under the assumption of exponential models for X and Y, it follows that the instantaneous availability at time t is A(t)=lambda/(lambda+theta)+theta/(lambda+theta)exp[-[(1/lambda)+(1/theta)]t] with t>0. Also, the steady-state availability is A(infinity)=lambda/(lambda+theta). We use the observations from n failure-repair cycles of the power plant, say X 1 , X 2 , ..., Xsub(n), Y 1 , Y 2 , ..., Ysub(n) to present the maximum likelihood estimators of A(t) and A(infinity). The exact sampling distributions for those estimators and some statistical properties are discussed before a simulation model is used to determine 95% simulation intervals for A(t). The methodology is applied to two examples which approximate the operating history of two nuclear power plants. (author)

  20. 1992 Annual Capacity Report

    International Nuclear Information System (INIS)

    1993-05-01

    The Standard Contract for Disposal of Spent Nuclear Fuel and/or High-Level Radioactive Waste (10 CFR Part 961) requires the Department of Energy (DOE) to issue an Annual Capacity Report (ACR) for planning purposes. This report is the fifth in the series published by DOE. In May 1993, DOE published the 1992 Acceptance Priority Ranking (APR) that established the order in which DOE will allocate projected acceptance capacity. As required by the Standard Contract, the acceptance priority ranking is based on the date the spent nuclear fuel (SNF) was permanently discharged, with the owners of the oldest SNF, on an industry-wide basis, given the highest priority. The 1992 ACR applies the projected waste acceptance rates in Table 2.1 to the 1992 APR, resulting in individual allocations for the owners and generators of the SNF. These allocations are listed in detail in the Appendix, and summarized in Table 3.1. The projected waste acceptance rates for SNF presented in Table 2.1 are nominal and assume a site for a Monitored Retrievable Storage (MRS) facility will be obtained; the facility will initiate operations in 1998; and the statutory linkages between the MRS facility and the repository set forth in the Nuclear Waste Policy Act of 1982, as amended (NWPA), will be modified. During the first ten years following projected commencement of Civilian Radioactive Waste Management System (CRWMS) operation, the total quantity of SNF that could be accepted is projected to be 8,200 metric tons of uranium (MTU). This is consistent with the storage capacity licensing conditions imposed on an MRS facility by the NWPA. The annual acceptance rates provide an approximation of the system throughput and are subject to change as the program progresses

  1. Capacity at Railway Stations

    DEFF Research Database (Denmark)

    Landex, Alex

    2011-01-01

    zone(s) the possible conflicts with other trains (also in the opposite direction) are taken into account leading to more trustworthy results. Although the UIC 406 methodology proposes that the railway network should be divided into line sections when trains turn around and when the train order...... is changed, this paper recommends that the railway lines are not always be divided. In case trains turn around on open (single track) line, the capacity consumption may be too low if a railway line is divided. The same can be the case if only few trains are overtaken at an overtaking station. For dead end...

  2. Collision cascades and sputtering induced by larger cluster ions

    International Nuclear Information System (INIS)

    Sigmund, P.

    1988-01-01

    Recent experimental work on larger cluster impact on solid surfaces suggests large deviations from the standard case of additive sputter yields both in the nuclear and electronic stopping regime. The paper concentrates on elastic collision cascades. In addition to very pronounced spike effects, two phenomena are pointed out that are specific to cluster bombardment. Multiple hits of cluster atoms on one and the same target atom may result in recoil atoms that move faster than the maximum recoil speed for monomer bombardment at the same projectile speed. This effect is important when the atomic mass of a beam atom is less than that of a target atom, M 1 2 . In the opposite case, M 1 >> M 2 , collisions between beam particles may accelerate some beam particles and slow down others. Some consequences are mentioned. Remarks on the nuclear stopping power of larger clusters and on electronic sputtering by cluster bombardment conclude the paper. 38 refs., 2 figs

  3. Metal cluster compounds - chemistry and importance; clusters containing isolated main group element atoms, large metal cluster compounds, cluster fluxionality

    International Nuclear Information System (INIS)

    Walther, B.

    1988-01-01

    This part of the review on metal cluster compounds deals with clusters containing isolated main group element atoms, with high nuclearity clusters and metal cluster fluxionality. It will be obvious that main group element atoms strongly influence the geometry, stability and reactivity of the clusters. High nuclearity clusters are of interest in there own due to the diversity of the structures adopted, but their intermediate position between molecules and the metallic state makes them a fascinating research object too. These both sites of the metal cluster chemistry as well as the frequently observed ligand and core fluxionality are related to the cluster metal and surface analogy. (author)

  4. Disentangling Porterian Clusters

    DEFF Research Database (Denmark)

    Jagtfelt, Tue

    , contested theory become so widely disseminated and applied as a normative and prescriptive strategy for economic development? The dissertation traces the introduction of the cluster notion into the EU’s Lisbon Strategy and demonstrates how its inclusion originates from Porter’s colleagues: Professor Örjan...... to his membership on the Commission on Industrial Competitiveness, and that the cluster notion found in his influential book, Nations, represents a significant shift in his conception of cluster compared with his early conceptions. This shift, it is argued, is a deliberate attempt by Porter to create...... a paradigmatic textbook that follows Kuhn’s blueprint for scientific revolutions by instilling Nations with circular references and thus creating a local linguistic holism conceptualized through an encompassing notion of cluster. The dissertation concludes that the two research questions are philosophically...

  5. Remarks on stellar clusters

    International Nuclear Information System (INIS)

    Teller, E.

    1985-01-01

    In the following, a few simple remarks on the evolution and properties of stellar clusters will be collected. In particular, globular clusters will be considered. Though details of such clusters are often not known, a few questions can be clarified with the help of primitive arguments. These are:- why are spherical clusters spherical, why do they have high densities, why do they consist of approximately a million stars, how may a black hole of great mass form within them, may they be the origin of gamma-ray bursts, may their invisible remnants account for the missing mass of our galaxy. The available data do not warrant a detailed evaluation. However, it is remarkable that exceedingly simple models can shed some light on the questions enumerated above. (author)

  6. From collisions to clusters

    DEFF Research Database (Denmark)

    Loukonen, Ville; Bork, Nicolai; Vehkamaki, Hanna

    2014-01-01

    -principles molecular dynamics collision simulations of (sulphuric acid)1(water)0, 1 + (dimethylamine) → (sulphuric acid)1(dimethylamine)1(water)0, 1 cluster formation processes. The simulations indicate that the sticking factor in the collisions is unity: the interaction between the molecules is strong enough...... control. As a consequence, the clusters show very dynamic ion pair structure, which differs from both the static structure optimisation calculations and the equilibrium first-principles molecular dynamics simulations. In some of the simulation runs, water mediates the proton transfer by acting as a proton...... to overcome the possible initial non-optimal collision orientations. No post-collisional cluster break up is observed. The reasons for the efficient clustering are (i) the proton transfer reaction which takes place in each of the collision simulations and (ii) the subsequent competition over the proton...

  7. Clustering of Emerging Flux

    Science.gov (United States)

    Ruzmaikin, A.

    1997-01-01

    Observations show that newly emerging flux tends to appear on the Solar surface at sites where there is flux already. This results in clustering of solar activity. Standard dynamo theories do not predict this effect.

  8. How Clusters Work

    Science.gov (United States)

    Technology innovation clusters are geographic concentrations of interconnected companies, universities, and other organizations with a focus on environmental technology. They play a key role in addressing the nation’s pressing environmental problems.

  9. Evolution of clustered storage

    CERN Multimedia

    CERN. Geneva; Van de Vyvre, Pierre

    2007-01-01

    The session actually featured two presentations: * Evolution of clustered storage by Lance Hukill, Quantum Corporation * ALICE DAQ - Usage of a Cluster-File System: Quantum StorNext by Pierre Vande Vyvre, CERN-PH the second one prepared at short notice by Pierre (thanks!) to present how the Quantum technologies are being used in the ALICE experiment. The abstract to Mr Hukill's follows. Clustered Storage is a technology that is driven by business and mission applications. The evolution of Clustered Storage solutions starts first at the alignment between End-users needs and Industry trends: * Push-and-Pull between managing for today versus planning for tomorrow * Breaking down the real business problems to the core applications * Commoditization of clients, servers, and target devices * Interchangeability, Interoperability, Remote Access, Centralized control * Oh, and yes, there is a budget and the "real world" to deal with This presentation will talk through these needs and trends, and then ask the question, ...

  10. Galaxy clusters and cosmology

    CERN Document Server

    White, S

    1994-01-01

    Galaxy clusters are the largest coherent objects in Universe. It has been known since 1933 that their dynamical properties require either a modification of the theory of gravity, or the presence of a dominant component of unseen material of unknown nature. Clusters still provide the best laboratories for studying the amount and distribution of this dark matter relative to the material which can be observed directly -- the galaxies themselves and the hot,X-ray-emitting gas which lies between them.Imaging and spectroscopy of clusters by satellite-borne X -ray telescopes has greatly improved our knowledge of the structure and composition of this intergalactic medium. The results permit a number of new approaches to some fundamental cosmological questions,but current indications from the data are contradictory. The observed irregularity of real clusters seems to imply recent formation epochs which would require a universe with approximately the critical density. On the other hand, the large baryon fraction observ...

  11. Applications of Clustering

    Indian Academy of Sciences (India)

    First page Back Continue Last page Overview Graphics. Applications of Clustering. Biology – medical imaging, bioinformatics, ecology, phylogenies problems etc. Market research. Data Mining. Social Networks. Any problem measuring similarity/correlation. (dimensions represent different parameters)

  12. Clustering Game Behavior Data

    DEFF Research Database (Denmark)

    Bauckhage, C.; Drachen, Anders; Sifa, Rafet

    2015-01-01

    of the causes, the proliferation of behavioral data poses the problem of how to derive insights therefrom. Behavioral data sets can be large, time-dependent and high-dimensional. Clustering offers a way to explore such data and to discover patterns that can reduce the overall complexity of the data. Clustering...... and other techniques for player profiling and play style analysis have, therefore, become popular in the nascent field of game analytics. However, the proper use of clustering techniques requires expertise and an understanding of games is essential to evaluate results. With this paper, we address game data...... scientists and present a review and tutorial focusing on the application of clustering techniques to mine behavioral game data. Several algorithms are reviewed and examples of their application shown. Key topics such as feature normalization are discussed and open problems in the context of game analytics...

  13. Clustering on Membranes

    DEFF Research Database (Denmark)

    Johannes, Ludger; Pezeshkian, Weria; Ipsen, John H

    2018-01-01

    Clustering of extracellular ligands and proteins on the plasma membrane is required to perform specific cellular functions, such as signaling and endocytosis. Attractive forces that originate in perturbations of the membrane's physical properties contribute to this clustering, in addition to direct...... protein-protein interactions. However, these membrane-mediated forces have not all been equally considered, despite their importance. In this review, we describe how line tension, lipid depletion, and membrane curvature contribute to membrane-mediated clustering. Additional attractive forces that arise...... from protein-induced perturbation of a membrane's fluctuations are also described. This review aims to provide a survey of the current understanding of membrane-mediated clustering and how this supports precise biological functions....

  14. Air void clustering.

    Science.gov (United States)

    2015-06-01

    Air void clustering around coarse aggregate in concrete has been identified as a potential source of : low strengths in concrete mixes by several Departments of Transportation around the country. Research was : carried out to (1) develop a quantitati...

  15. Effective capacity of multiple antenna channels: Correlation and keyhole

    KAUST Repository

    Zhong, Caijun

    2012-01-01

    In this study, the authors derive the effective capacity limits for multiple antenna channels which quantify the maximum achievable rate with consideration of link-layer delay-bound violation probability. Both correlated multiple-input single-output and multiple-input multiple-output keyhole channels are studied. Based on the closed-form exact expressions for the effective capacity of both channels, the authors look into the asymptotic high and low signal-to-noise ratio regimes, and derive simple expressions to gain more insights. The impact of spatial correlation on effective capacity is also characterised with the aid of a majorisation theory result. It is revealed that antenna correlation reduces the effective capacity of the channels and a stringent quality-of-service requirement causes a severe reduction in the effective capacity but can be alleviated by increasing the number of antennas. © 2012 The Institution of Engineering and Technology.

  16. Speaker segmentation and clustering

    OpenAIRE

    Kotti, M; Moschou, V; Kotropoulos, C

    2008-01-01

    07.08.13 KB. Ok to add the accepted version to Spiral, Elsevier says ok whlile mandate not enforced. This survey focuses on two challenging speech processing topics, namely: speaker segmentation and speaker clustering. Speaker segmentation aims at finding speaker change points in an audio stream, whereas speaker clustering aims at grouping speech segments based on speaker characteristics. Model-based, metric-based, and hybrid speaker segmentation algorithms are reviewed. Concerning speaker...

  17. Fermion cluster algorithms

    International Nuclear Information System (INIS)

    Chandrasekharan, Shailesh

    2000-01-01

    Cluster algorithms have been recently used to eliminate sign problems that plague Monte-Carlo methods in a variety of systems. In particular such algorithms can also be used to solve sign problems associated with the permutation of fermion world lines. This solution leads to the possibility of designing fermion cluster algorithms in certain cases. Using the example of free non-relativistic fermions we discuss the ideas underlying the algorithm

  18. BUILDING e-CLUSTERS

    OpenAIRE

    Milan Davidovic

    2013-01-01

    E-clusters are strategic alliance in TIMES technology sector (Telecommunication, Information technology, Multimedia, Entertainment, Security) where products and processes are digitalized. They enable horizontal and vertical integration of small and medium companies and establish new added value e-chains. E-clusters also build supply chains based on cooperation relationship, innovation, organizational knowledge and compliance of intellectual properties. As an innovative approach for economic p...

  19. Clusters and exotic processes

    International Nuclear Information System (INIS)

    Schiffer, J.P.

    1975-01-01

    An attempt is made to present some data which may be construed as indicating that perhaps clusters play a role in high energy and exotic pion or kaon interactions with complex (A much greater than 16) nuclei. Also an attempt is made to summarize some very recent experimental work on pion interactions with nuclei which may or may not in the end support a picture in which clusters play an important role. (U.S.)

  20. Deficiency of employability capacity

    Directory of Open Access Journals (Sweden)

    Pelse I.

    2012-10-01

    Full Text Available Young unemployed people have comprised one of the significantly largest groups of the unemployed people in Latvia in recent years. One of the reasons why young people have difficulty integrating into the labour market is the “expectation gap” that exists in the relations between employers and the new generation of workers. Employers focus on capacity-building for employability such individual factors as strength, patience, self-discipline, self-reliance, self-motivation, etc., which having a nature of habit and are developed in a long-term work socialization process, which begins even before the formal education and will continue throughout the life cycle. However, when the socialization is lost, these habits are depreciated faster than they can be restored. Currently a new generation is entering the labour market, which is missing the succession of work socialization. Factors, such as rising unemployment and poverty in the background over the past twenty years in Latvia have created a very unfavourable employability background of “personal circumstances” and “external factors”, which seriously have impaired formation of the skills and attitudes in a real work environment. The study reveals another paradox – the paradox of poverty. Common sense would want to argue that poverty can be overcome by the job. However, the real state of affairs shows that unfavourable coincidence of the individual, personal circumstances and external factors leads to deficit of employability capacity and possibility of marked social and employment deprivation.

  1. A maximum power point tracking for photovoltaic-SPE system using a maximum current controller

    Energy Technology Data Exchange (ETDEWEB)

    Muhida, Riza [Osaka Univ., Dept. of Physical Science, Toyonaka, Osaka (Japan); Osaka Univ., Dept. of Electrical Engineering, Suita, Osaka (Japan); Park, Minwon; Dakkak, Mohammed; Matsuura, Kenji [Osaka Univ., Dept. of Electrical Engineering, Suita, Osaka (Japan); Tsuyoshi, Akira; Michira, Masakazu [Kobe City College of Technology, Nishi-ku, Kobe (Japan)

    2003-02-01

    Processes to produce hydrogen from solar photovoltaic (PV)-powered water electrolysis using solid polymer electrolysis (SPE) are reported. An alternative control of maximum power point tracking (MPPT) in the PV-SPE system based on the maximum current searching methods has been designed and implemented. Based on the characteristics of voltage-current and theoretical analysis of SPE, it can be shown that the tracking of the maximum current output of DC-DC converter in SPE side will track the MPPT of photovoltaic panel simultaneously. This method uses a proportional integrator controller to control the duty factor of DC-DC converter with pulse-width modulator (PWM). The MPPT performance and hydrogen production performance of this method have been evaluated and discussed based on the results of the experiment. (Author)

  2. Robust continuous clustering.

    Science.gov (United States)

    Shah, Sohil Atul; Koltun, Vladlen

    2017-09-12

    Clustering is a fundamental procedure in the analysis of scientific data. It is used ubiquitously across the sciences. Despite decades of research, existing clustering algorithms have limited effectiveness in high dimensions and often require tuning parameters for different domains and datasets. We present a clustering algorithm that achieves high accuracy across multiple domains and scales efficiently to high dimensions and large datasets. The presented algorithm optimizes a smooth continuous objective, which is based on robust statistics and allows heavily mixed clusters to be untangled. The continuous nature of the objective also allows clustering to be integrated as a module in end-to-end feature learning pipelines. We demonstrate this by extending the algorithm to perform joint clustering and dimensionality reduction by efficiently optimizing a continuous global objective. The presented approach is evaluated on large datasets of faces, hand-written digits, objects, newswire articles, sensor readings from the Space Shuttle, and protein expression levels. Our method achieves high accuracy across all datasets, outperforming the best prior algorithm by a factor of 3 in average rank.

  3. Cluster bomb ocular injuries.

    Science.gov (United States)

    Mansour, Ahmad M; Hamade, Haya; Ghaddar, Ayman; Mokadem, Ahmad Samih; El Hajj Ali, Mohamad; Awwad, Shady

    2012-01-01

    To present the visual outcomes and ocular sequelae of victims of cluster bombs. This retrospective, multicenter case series of ocular injury due to cluster bombs was conducted for 3 years after the war in South Lebanon (July 2006). Data were gathered from the reports to the Information Management System for Mine Action. There were 308 victims of clusters bombs; 36 individuals were killed, of which 2 received ocular lacerations and; 272 individuals were injured with 18 receiving ocular injury. These 18 surviving individuals were assessed by the authors. Ocular injury occurred in 6.5% (20/308) of cluster bomb victims. Trauma to multiple organs occurred in 12 of 18 cases (67%) with ocular injury. Ocular findings included corneal or scleral lacerations (16 eyes), corneal foreign bodies (9 eyes), corneal decompensation (2 eyes), ruptured cataract (6 eyes), and intravitreal foreign bodies (10 eyes). The corneas of one patient had extreme attenuation of the endothelium. Ocular injury occurred in 6.5% of cluster bomb victims and 67% of the patients with ocular injury sustained trauma to multiple organs. Visual morbidity in civilians is an additional reason for a global ban on the use of cluster bombs.

  4. Capacity limits introduced by data fusion on cooperative spectrum sensing under correlated environments

    DEFF Research Database (Denmark)

    Pratas, Nuno; Marchetti, Nicola; Rodrigues, Antonio

    2010-01-01

    spectrum sensing scheme, by measuring the perceived capacity limits introduced by the use of data fusion on cooperative sensing schemes. The analysis is supported by evaluation metrics which account for the perceived capacity limits. The analysis is performed along the data fusion chain, comparing several...... scenarios encompassing different degree of environment correlation between the cluster nodes, number of cluster nodes and sensed channel occupation statistics. Through this study we motivate that to maximize the perceived capacity by the cooperative spectrum sensing, the use of data fusion needs...

  5. REGIONAL DEVELOPMENT BASED ON CLUSTER IN LIVESTOCK DEVELOPMENT. CLUSTER IN LIVESTOCK SECTOR IN THE KYRGYZ REPUBLIC

    Directory of Open Access Journals (Sweden)

    Meerim SYDYKOVA

    2014-11-01

    Full Text Available In most developing countries, where agriculture is the main economical source, clusters have been found as a booster to develop their economy. The Asian countries are now starting to implement agro-food clusters into the mainstream of changes in agriculture, farming and food industry. The long-term growth of meat production in the Kyrgyz Republic during the last decade, as well as the fact that agriculture has become one of the prioritized sectors of the economy, proved the importance of livestock sector in the economy of the Kyrgyz Republic. The research question is “Does the Kyrgyz Republic has strong economic opportunities and prerequisites in agriculture in order to implement an effective agro cluster in the livestock sector?” Paper focuses on describing the prerequisites of the Kyrgyz Republic in agriculture to implement livestock cluster. The main objective of the paper is to analyse the livestock sector of the Kyrgyz Republic and observe the capacity of this sector to implement agro-cluster. The study focuses on investigating livestock sector and a complex S.W.O.T. The analysis was carried out based on local and regional database and official studies. The results of research demonstrate the importance of livestock cluster for national economy. It can be concluded that cluster implementation could provide to its all members with benefits if they could build strong collaborative relationship in order to facilitate the access to the labour market and implicitly, the access to exchange of good practices. Their ability of potential cluster members to act as a convergence pole is critical for acquiring practical skills necessary for the future development of the livestock sector.

  6. Research on configuration of railway self-equipped tanker based on minimum cost maximum flow model

    Science.gov (United States)

    Yang, Yuefang; Gan, Chunhui; Shen, Tingting

    2017-05-01

    In the study of the configuration of the tanker of chemical logistics park, the minimum cost maximum flow model is adopted. Firstly, the transport capacity of the park loading and unloading area and the transportation demand of the dangerous goods are taken as the constraint condition of the model; then the transport arc capacity, the transport arc flow and the transport arc edge weight are determined in the transportation network diagram; finally, the software calculations. The calculation results show that the configuration issue of the tankers can be effectively solved by the minimum cost maximum flow model, which has theoretical and practical application value for tanker management of railway transportation of dangerous goods in the chemical logistics park.

  7. Determination of atomic cluster structure with cluster fusion algorithm

    DEFF Research Database (Denmark)

    Obolensky, Oleg I.; Solov'yov, Ilia; Solov'yov, Andrey V.

    2005-01-01

    We report an efficient scheme of global optimization, called cluster fusion algorithm, which has proved its reliability and high efficiency in determination of the structure of various atomic clusters.......We report an efficient scheme of global optimization, called cluster fusion algorithm, which has proved its reliability and high efficiency in determination of the structure of various atomic clusters....

  8. Capacity of weakly (d,k)-constrained sequences

    NARCIS (Netherlands)

    Schouhamer Immink, K.A.; Janssen, A.J.E.M.

    2000-01-01

    In the presentation we find an analytic expression for the maximum of the normalized entropy - ¿ieT pi In pi / ¿ieT ipi where the set T is the disjoint union of sets Sn of positive integers that are assigned probabilities Pn, ¿n Pn = 1. This result is applied to the computation of the capacity of

  9. Phosphorus retention capacity of sediments in Mandovi estuary (Goa)

    Digital Repository Service at National Institute of Oceanography (India)

    Rajagopal, M.D.; Reddy, C

    Experiments carried out under controlled conditions to study P retention capacity of sediments indicate that the processes of adsorption and desorption of P are pH dependent. Adsorption of P is maximum (58-99%) at pH 4. Both the exchangeable P...

  10. Cluster dynamics at different cluster size and incident laser wavelengths

    International Nuclear Information System (INIS)

    Desai, Tara; Bernardinello, Andrea

    2002-01-01

    X-ray emission spectra from aluminum clusters of diameter -0.4 μm and gold clusters of dia. ∼1.25 μm are experimentally studied by irradiating the cluster foil targets with 1.06 μm laser, 10 ns (FWHM) at an intensity ∼10 12 W/cm 2 . Aluminum clusters show a different spectra compared to bulk material whereas gold cluster evolve towards bulk gold. Experimental data are analyzed on the basis of cluster dimension, laser wavelength and pulse duration. PIC simulations are performed to study the behavior of clusters at higher intensity I≥10 17 W/cm 2 for different size of the clusters irradiated at different laser wavelengths. Results indicate the dependence of cluster dynamics on cluster size and incident laser wavelength

  11. Cluster fusion algorithm: application to Lennard-Jones clusters

    DEFF Research Database (Denmark)

    Solov'yov, Ilia; Solov'yov, Andrey V.; Greiner, Walter

    2006-01-01

    paths up to the cluster size of 150 atoms. We demonstrate that in this way all known global minima structures of the Lennard-Jones clusters can be found. Our method provides an efficient tool for the calculation and analysis of atomic cluster structure. With its use we justify the magic number sequence......We present a new general theoretical framework for modelling the cluster structure and apply it to description of the Lennard-Jones clusters. Starting from the initial tetrahedral cluster configuration, adding new atoms to the system and absorbing its energy at each step, we find cluster growing...... for the clusters of noble gas atoms and compare it with experimental observations. We report the striking correspondence of the peaks in the dependence of the second derivative of the binding energy per atom on cluster size calculated for the chain of the Lennard-Jones clusters based on the icosahedral symmetry...

  12. Cluster fusion algorithm: application to Lennard-Jones clusters

    DEFF Research Database (Denmark)

    Solov'yov, Ilia; Solov'yov, Andrey V.; Greiner, Walter

    2008-01-01

    paths up to the cluster size of 150 atoms. We demonstrate that in this way all known global minima structures of the Lennard-Jones clusters can be found. Our method provides an efficient tool for the calculation and analysis of atomic cluster structure. With its use we justify the magic number sequence......We present a new general theoretical framework for modelling the cluster structure and apply it to description of the Lennard-Jones clusters. Starting from the initial tetrahedral cluster configuration, adding new atoms to the system and absorbing its energy at each step, we find cluster growing...... for the clusters of noble gas atoms and compare it with experimental observations. We report the striking correspondence of the peaks in the dependence of the second derivative of the binding energy per atom on cluster size calculated for the chain of the Lennard-Jones clusters based on the icosahedral symmetry...

  13. Paddle River Dam : review of probable maximum flood

    Energy Technology Data Exchange (ETDEWEB)

    Clark, D. [UMA Engineering Ltd., Edmonton, AB (Canada); Neill, C.R. [Northwest Hydraulic Consultants Ltd., Edmonton, AB (Canada)

    2008-07-01

    The Paddle River Dam was built in northern Alberta in the mid 1980s for flood control. According to the 1999 Canadian Dam Association (CDA) guidelines, this 35 metre high, zoned earthfill dam with a spillway capacity sized to accommodate a probable maximum flood (PMF) is rated as a very high hazard. At the time of design, it was estimated to have a peak flow rate of 858 centimetres. A review of the PMF in 2002 increased the peak flow rate to 1,890 centimetres. In light of a 2007 revision of the CDA safety guidelines, the PMF was reviewed and the inflow design flood (IDF) was re-evaluated. This paper discussed the levels of uncertainty inherent in PMF determinations and some difficulties encountered with the SSARR hydrologic model and the HEC-RAS hydraulic model in unsteady mode. The paper also presented and discussed the analysis used to determine incremental damages, upon which a new IDF of 840 m{sup 3}/s was recommended. The paper discussed the PMF review, modelling methodology, hydrograph inputs, and incremental damage of floods. It was concluded that the PMF review, involving hydraulic routing through the valley bottom together with reconsideration of the previous runoff modeling provides evidence that the peak reservoir inflow could reasonably be reduced by approximately 20 per cent. 8 refs., 5 tabs., 8 figs.

  14. Maximum mass of magnetic white dwarfs

    International Nuclear Information System (INIS)

    Paret, Daryel Manreza; Horvath, Jorge Ernesto; Martínez, Aurora Perez

    2015-01-01

    We revisit the problem of the maximum masses of magnetized white dwarfs (WDs). The impact of a strong magnetic field on the structure equations is addressed. The pressures become anisotropic due to the presence of the magnetic field and split into parallel and perpendicular components. We first construct stable solutions of the Tolman-Oppenheimer-Volkoff equations for parallel pressures and find that physical solutions vanish for the perpendicular pressure when B ≳ 10 13 G. This fact establishes an upper bound for a magnetic field and the stability of the configurations in the (quasi) spherical approximation. Our findings also indicate that it is not possible to obtain stable magnetized WDs with super-Chandrasekhar masses because the values of the magnetic field needed for them are higher than this bound. To proceed into the anisotropic regime, we can apply results for structure equations appropriate for a cylindrical metric with anisotropic pressures that were derived in our previous work. From the solutions of the structure equations in cylindrical symmetry we have confirmed the same bound for B ∼ 10 13 G, since beyond this value no physical solutions are possible. Our tentative conclusion is that massive WDs with masses well beyond the Chandrasekhar limit do not constitute stable solutions and should not exist. (paper)

  15. TRENDS IN ESTIMATED MIXING DEPTH DAILY MAXIMUMS

    Energy Technology Data Exchange (ETDEWEB)

    Buckley, R; Amy DuPont, A; Robert Kurzeja, R; Matt Parker, M

    2007-11-12

    Mixing depth is an important quantity in the determination of air pollution concentrations. Fireweather forecasts depend strongly on estimates of the mixing depth as a means of determining the altitude and dilution (ventilation rates) of smoke plumes. The Savannah River United States Forest Service (USFS) routinely conducts prescribed fires at the Savannah River Site (SRS), a heavily wooded Department of Energy (DOE) facility located in southwest South Carolina. For many years, the Savannah River National Laboratory (SRNL) has provided forecasts of weather conditions in support of the fire program, including an estimated mixing depth using potential temperature and turbulence change with height at a given location. This paper examines trends in the average estimated mixing depth daily maximum at the SRS over an extended period of time (4.75 years) derived from numerical atmospheric simulations using two versions of the Regional Atmospheric Modeling System (RAMS). This allows for differences to be seen between the model versions, as well as trends on a multi-year time frame. In addition, comparisons of predicted mixing depth for individual days in which special balloon soundings were released are also discussed.

  16. Mammographic image restoration using maximum entropy deconvolution

    International Nuclear Information System (INIS)

    Jannetta, A; Jackson, J C; Kotre, C J; Birch, I P; Robson, K J; Padgett, R

    2004-01-01

    An image restoration approach based on a Bayesian maximum entropy method (MEM) has been applied to a radiological image deconvolution problem, that of reduction of geometric blurring in magnification mammography. The aim of the work is to demonstrate an improvement in image spatial resolution in realistic noisy radiological images with no associated penalty in terms of reduction in the signal-to-noise ratio perceived by the observer. Images of the TORMAM mammographic image quality phantom were recorded using the standard magnification settings of 1.8 magnification/fine focus and also at 1.8 magnification/broad focus and 3.0 magnification/fine focus; the latter two arrangements would normally give rise to unacceptable geometric blurring. Measured point-spread functions were used in conjunction with the MEM image processing to de-blur these images. The results are presented as comparative images of phantom test features and as observer scores for the raw and processed images. Visualization of high resolution features and the total image scores for the test phantom were improved by the application of the MEM processing. It is argued that this successful demonstration of image de-blurring in noisy radiological images offers the possibility of weakening the link between focal spot size and geometric blurring in radiology, thus opening up new approaches to system optimization

  17. Paving the road to maximum productivity.

    Science.gov (United States)

    Holland, C

    1998-01-01

    "Job security" is an oxymoron in today's environment of downsizing, mergers, and acquisitions. Workers find themselves living by new rules in the workplace that they may not understand. How do we cope? It is the leader's charge to take advantage of this chaos and create conditions under which his or her people can understand the need for change and come together with a shared purpose to effect that change. The clinical laboratory at Arkansas Children's Hospital has taken advantage of this chaos to down-size and to redesign how the work gets done to pave the road to maximum productivity. After initial hourly cutbacks, the workers accepted the cold, hard fact that they would never get their old world back. They set goals to proactively shape their new world through reorganizing, flexing staff with workload, creating a rapid response laboratory, exploiting information technology, and outsourcing. Today the laboratory is a lean, productive machine that accepts change as a way of life. We have learned to adapt, trust, and support each other as we have journeyed together over the rough roads. We are looking forward to paving a new fork in the road to the future.

  18. Maximum power flux of auroral kilometric radiation

    International Nuclear Information System (INIS)

    Benson, R.F.; Fainberg, J.

    1991-01-01

    The maximum auroral kilometric radiation (AKR) power flux observed by distant satellites has been increased by more than a factor of 10 from previously reported values. This increase has been achieved by a new data selection criterion and a new analysis of antenna spin modulated signals received by the radio astronomy instrument on ISEE 3. The method relies on selecting AKR events containing signals in the highest-frequency channel (1980, kHz), followed by a careful analysis that effectively increased the instrumental dynamic range by more than 20 dB by making use of the spacecraft antenna gain diagram during a spacecraft rotation. This analysis has allowed the separation of real signals from those created in the receiver by overloading. Many signals having the appearance of AKR harmonic signals were shown to be of spurious origin. During one event, however, real second harmonic AKR signals were detected even though the spacecraft was at a great distance (17 R E ) from Earth. During another event, when the spacecraft was at the orbital distance of the Moon and on the morning side of Earth, the power flux of fundamental AKR was greater than 3 x 10 -13 W m -2 Hz -1 at 360 kHz normalized to a radial distance r of 25 R E assuming the power falls off as r -2 . A comparison of these intense signal levels with the most intense source region values (obtained by ISIS 1 and Viking) suggests that multiple sources were observed by ISEE 3

  19. Maximum likelihood window for time delay estimation

    International Nuclear Information System (INIS)

    Lee, Young Sup; Yoon, Dong Jin; Kim, Chi Yup

    2004-01-01

    Time delay estimation for the detection of leak location in underground pipelines is critically important. Because the exact leak location depends upon the precision of the time delay between sensor signals due to leak noise and the speed of elastic waves, the research on the estimation of time delay has been one of the key issues in leak lovating with the time arrival difference method. In this study, an optimal Maximum Likelihood window is considered to obtain a better estimation of the time delay. This method has been proved in experiments, which can provide much clearer and more precise peaks in cross-correlation functions of leak signals. The leak location error has been less than 1 % of the distance between sensors, for example the error was not greater than 3 m for 300 m long underground pipelines. Apart from the experiment, an intensive theoretical analysis in terms of signal processing has been described. The improved leak locating with the suggested method is due to the windowing effect in frequency domain, which offers a weighting in significant frequencies.

  20. Ancestral Sequence Reconstruction with Maximum Parsimony.

    Science.gov (United States)

    Herbst, Lina; Fischer, Mareike

    2017-12-01

    One of the main aims in phylogenetics is the estimation of ancestral sequences based on present-day data like, for instance, DNA alignments. One way to estimate the data of the last common ancestor of a given set of species is to first reconstruct a phylogenetic tree with some tree inference method and then to use some method of ancestral state inference based on that tree. One of the best-known methods both for tree inference and for ancestral sequence inference is Maximum Parsimony (MP). In this manuscript, we focus on this method and on ancestral state inference for fully bifurcating trees. In particular, we investigate a conjecture published by Charleston and Steel in 1995 concerning the number of species which need to have a particular state, say a, at a particular site in order for MP to unambiguously return a as an estimate for the state of the last common ancestor. We prove the conjecture for all even numbers of character states, which is the most relevant case in biology. We also show that the conjecture does not hold in general for odd numbers of character states, but also present some positive results for this case.

  1. 1990s: High Capacity Backbones

    Indian Academy of Sciences (India)

    First page Back Continue Last page Overview Graphics. 1990s: High Capacity Backbones. Backbone capacities increased from 2.5 Gb/s to 100s of Gb/s during the 1990's. Wavelength division multiplexing with 160 waves of 10 Gb/s was commercially available. Several high-capacity backbones built in the US and Europe.

  2. Competitive Capacity Investment under Uncertainty

    NARCIS (Netherlands)

    X. Li (Xishu); R.A. Zuidwijk (Rob); M.B.M. de Koster (René); R. Dekker (Rommert)

    2016-01-01

    textabstractWe consider a long-term capacity investment problem in a competitive market under demand uncertainty. Two firms move sequentially in the competition and a firm’s capacity decision interacts with the other firm’s current and future capacity. Throughout the investment race, a firm can

  3. Integrated capacity and inventory decisions

    NARCIS (Netherlands)

    Dellaert, N.P.; Flapper, S.D.P.; Tan, T.; Jeunet, J.; Klatte, D.; Lüthi, H.-J.; Schmedders, K.

    2012-01-01

    This paper deals with the simultaneous acquisition of capacity and material in a situation with uncertain demand, with non-zero lead-times for the supply of both material and capacity. Although there is a lot of literature on the time-phased acquisition of capacity and material, most of this

  4. Landfill Construction and Capacity Expansion

    NARCIS (Netherlands)

    Andre, F.J.; Cerda, E.

    2003-01-01

    We study the optimal capacity and lifetime of landfills taking into account their sequential nature.Such an optimal capacity is characterized by the so-called Optimal Capacity Condition.Particular versions of this condition are obtained for two alternative settings: first, if all the landfills are

  5. 49 CFR 230.24 - Maximum allowable stress.

    Science.gov (United States)

    2010-10-01

    ... 49 Transportation 4 2010-10-01 2010-10-01 false Maximum allowable stress. 230.24 Section 230.24... Allowable Stress § 230.24 Maximum allowable stress. (a) Maximum allowable stress value. The maximum allowable stress value on any component of a steam locomotive boiler shall not exceed 1/4 of the ultimate...

  6. 20 CFR 226.52 - Total annuity subject to maximum.

    Science.gov (United States)

    2010-04-01

    ... 20 Employees' Benefits 1 2010-04-01 2010-04-01 false Total annuity subject to maximum. 226.52... COMPUTING EMPLOYEE, SPOUSE, AND DIVORCED SPOUSE ANNUITIES Railroad Retirement Family Maximum § 226.52 Total annuity subject to maximum. The total annuity amount which is compared to the maximum monthly amount to...

  7. Spatial cluster detection using dynamic programming

    Directory of Open Access Journals (Sweden)

    Sverchkov Yuriy

    2012-03-01

    Full Text Available Abstract Background The task of spatial cluster detection involves finding spatial regions where some property deviates from the norm or the expected value. In a probabilistic setting this task can be expressed as finding a region where some event is significantly more likely than usual. Spatial cluster detection is of interest in fields such as biosurveillance, mining of astronomical data, military surveillance, and analysis of fMRI images. In almost all such applications we are interested both in the question of whether a cluster exists in the data, and if it exists, we are interested in finding the most accurate characterization of the cluster. Methods We present a general dynamic programming algorithm for grid-based spatial cluster detection. The algorithm can be used for both Bayesian maximum a-posteriori (MAP estimation of the most likely spatial distribution of clusters and Bayesian model averaging over a large space of spatial cluster distributions to compute the posterior probability of an unusual spatial clustering. The algorithm is explained and evaluated in the context of a biosurveillance application, specifically the detection and identification of Influenza outbreaks based on emergency department visits. A relatively simple underlying model is constructed for the purpose of evaluating the algorithm, and the algorithm is evaluated using the model and semi-synthetic test data. Results When compared to baseline methods, tests indicate that the new algorithm can improve MAP estimates under certain conditions: the greedy algorithm we compared our method to was found to be more sensitive to smaller outbreaks, while as the size of the outbreaks increases, in terms of area affected and proportion of individuals affected, our method overtakes the greedy algorithm in spatial precision and recall. The new algorithm performs on-par with baseline methods in the task of Bayesian model averaging. Conclusions We conclude that the dynamic

  8. Capacity Ten-Seven

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2006-05-04

    The study was initiated to assemble UK solar experts who would define the parameters for the design of solar cell process plant with a capacity of about 10MWp per annum and establish the optimum substrate size, cell configuration and junction structure and show that these characteristics are suited to high volume low cost production. The study focused on proven materials and techniques rather than on innovative materials and processes. Peak production should deliver panels at a cost of less than $1 per peak watt. The study was reported under the headings of fundamental studies, front contact deposition, semiconductor deposition, rear contact deposition, cell isolation and initial design study. Since much of the work is commercially sensitive, not all findings are divulged in the report. The work, which achieved all its targets, was carried out by ICP Solar Technologies UK Ltd under contract to the DTI.

  9. GibbsCluster: unsupervised clustering and alignment of peptide sequences

    DEFF Research Database (Denmark)

    Andreatta, Massimo; Alvarez, Bruno; Nielsen, Morten

    2017-01-01

    motif characterizing each cluster. Several parameters are available to customize cluster analysis, including adjustable penalties for small clusters and overlapping groups and a trash cluster to remove outliers. As an example application, we used the server to deconvolute multiple specificities in large......-scale peptidome data generated by mass spectrometry. The server is available at http://www.cbs.dtu.dk/services/GibbsCluster-2.0....

  10. Half-width at half-maximum, full-width at half-maximum analysis

    Indian Academy of Sciences (India)

    addition to the well-defined parameter full-width at half-maximum (FWHM). The distribution of ... optical side-lobes in the diffraction pattern resulting in steep central maxima [6], reduc- tion of effects of ... and broad central peak. The idea of.

  11. Immunomodulating capacity of kefir.

    Science.gov (United States)

    Vinderola, Celso G; Duarte, Jairo; Thangavel, Deepa; Perdigón, Gabriela; Farnworth, Edward; Matar, Chantal

    2005-05-01

    Kefir is a fermented milk produced by the action of lactic acid bacteria, yeasts and acetic acid bacteria, trapped in a complex matrix of polysaccharides and proteins. Beyond its inherent high nutritional value as a source of proteins and calcium, kefir has a long tradition of being regarded as good for health in countries where it is a staple in the diet. However, published human or animal feeding trials to substantiate this view are not numerous. The aim of this work was to determine the immunomodulating capacity of kefir on the intestinal mucosal immune response in mice and to demonstrate the importance of dose and cell viability on this response. BALB/c mice were fed with commercial kefir ad libitum (diluted 1/10, 1/50, 1/100 or 1/200) or pasteurized kefir (diluted 1/6, 1/10, 1/50, 1/100) for 2, 5 or 7 consecutive days. At the end of each feeding period, the bacterial translocation assay was performed in the liver. Small intestine structure was studied by haematoxilin-eosin staining and light microscopy. The number of IgA+ and IgG+ cells was also determined. For the functional doses chosen, cytokines (IL-2, IL-4, IL-6, IL-10, IL-12, TNF-alpha and IFN-gamma) were determined. Kefir and pasteurized kefir were able to modulate the mucosal immune system in a dose-dependent manner. Kefir was administred 10-times more diluted than pasteurized kefir, but it induced an immunomodulation of similar magnitude, indicating the importance of cell viabilty. The results suggest that a Th1 response was controlled by Th2 cytokines induced by kefir feeding. Pasteurized kefir would induce both Th2 and Th1 responses. This is the first study in vivo regarding the mechanisms involved in the immunomodulating capacity of the oral administration of kefir containing viable or heat-inactivated bacteria at different doses.

  12. Cluster Implantation and Deposition Apparatus

    DEFF Research Database (Denmark)

    Hanif, Muhammad; Popok, Vladimir

    2015-01-01

    In the current report, a design and capabilities of a cluster implantation and deposition apparatus (CIDA) involving two different cluster sources are described. The clusters produced from gas precursors (Ar, N etc.) by PuCluS-2 can be used to study cluster ion implantation in order to develop...

  13. Subspace K-means clustering

    NARCIS (Netherlands)

    Timmerman, Marieke E.; Ceulemans, Eva; De Roover, Kim; Van Leeuwen, Karla

    2013-01-01

    To achieve an insightful clustering of multivariate data, we propose subspace K-means. Its central idea is to model the centroids and cluster residuals in reduced spaces, which allows for dealing with a wide range of cluster types and yields rich interpretations of the clusters. We review the

  14. Cosmic shear measurement with maximum likelihood and maximum a posteriori inference

    Science.gov (United States)

    Hall, Alex; Taylor, Andy

    2017-06-01

    We investigate the problem of noise bias in maximum likelihood and maximum a posteriori estimators for cosmic shear. We derive the leading and next-to-leading order biases and compute them in the context of galaxy ellipticity measurements, extending previous work on maximum likelihood inference for weak lensing. We show that a large part of the bias on these point estimators can be removed using information already contained in the likelihood when a galaxy model is specified, without the need for external calibration. We test these bias-corrected estimators on simulated galaxy images similar to those expected from planned space-based weak lensing surveys, with promising results. We find that the introduction of an intrinsic shape prior can help with mitigation of noise bias, such that the maximum a posteriori estimate can be made less biased than the maximum likelihood estimate. Second-order terms offer a check on the convergence of the estimators, but are largely subdominant. We show how biases propagate to shear estimates, demonstrating in our simple set-up that shear biases can be reduced by orders of magnitude and potentially to within the requirements of planned space-based surveys at mild signal-to-noise ratio. We find that second-order terms can exhibit significant cancellations at low signal-to-noise ratio when Gaussian noise is assumed, which has implications for inferring the performance of shear-measurement algorithms from simplified simulations. We discuss the viability of our point estimators as tools for lensing inference, arguing that they allow for the robust measurement of ellipticity and shear.

  15. A maximum likelihood framework for protein design

    Directory of Open Access Journals (Sweden)

    Philippe Hervé

    2006-06-01

    Full Text Available Abstract Background The aim of protein design is to predict amino-acid sequences compatible with a given target structure. Traditionally envisioned as a purely thermodynamic question, this problem can also be understood in a wider context, where additional constraints are captured by learning the sequence patterns displayed by natural proteins of known conformation. In this latter perspective, however, we still need a theoretical formalization of the question, leading to general and efficient learning methods, and allowing for the selection of fast and accurate objective functions quantifying sequence/structure compatibility. Results We propose a formulation of the protein design problem in terms of model-based statistical inference. Our framework uses the maximum likelihood principle to optimize the unknown parameters of a statistical potential, which we call an inverse potential to contrast with classical potentials used for structure prediction. We propose an implementation based on Markov chain Monte Carlo, in which the likelihood is maximized by gradient descent and is numerically estimated by thermodynamic integration. The fit of the models is evaluated by cross-validation. We apply this to a simple pairwise contact potential, supplemented with a solvent-accessibility term, and show that the resulting models have a better predictive power than currently available pairwise potentials. Furthermore, the model comparison method presented here allows one to measure the relative contribution of each component of the potential, and to choose the optimal number of accessibility classes, which turns out to be much higher than classically considered. Conclusion Altogether, this reformulation makes it possible to test a wide diversity of models, using different forms of potentials, or accounting for other factors than just the constraint of thermodynamic stability. Ultimately, such model-based statistical analyses may help to understand the forces

  16. Effect of a High-intensity Interval Training method on maximum oxygen consumption in Chilean schoolchildren

    Directory of Open Access Journals (Sweden)

    Sergio Galdames-Maliqueo

    2017-12-01

    Full Text Available Introduction: The low levels of maximum oxygen consumption (VO2max evaluated in Chilean schoolchildren suggest the startup of trainings that improve the aerobic capacity. Objective: To analyze the effect of a High-intensity Interval Training method on maximum oxygen consumption in Chilean schoolchildren. Materials and methods: Thirty-two high school students from the eighth grade, who were divided into two groups, were part of the study (experimental group = 16 students and control group = 16 students. The main analyzed variable was the maximum oxygen consumption through the Course Navette Test. A High-intensity Interval training method was applied based on the maximum aerobic speed obtained through the Test. A mixed ANOVA was used for statistical analysis. Results: The experimental group showed a significant increase in the Maximum Oxygen Consumption between the pretest and posttest when compared with the control group (p < 0.0001. Conclusion: The results of the study showed a positive effect of the High-intensity Interval Training on the maximum consumption of oxygen. At the end of the study, it is concluded that High-intensity Interval Training is a good stimulation methodology for Chilean schoolchildren.

  17. Projected coupled cluster theory.

    Science.gov (United States)

    Qiu, Yiheng; Henderson, Thomas M; Zhao, Jinmo; Scuseria, Gustavo E

    2017-08-14

    Coupled cluster theory is the method of choice for weakly correlated systems. But in the strongly correlated regime, it faces a symmetry dilemma, where it either completely fails to describe the system or has to artificially break certain symmetries. On the other hand, projected Hartree-Fock theory captures the essential physics of many kinds of strong correlations via symmetry breaking and restoration. In this work, we combine and try to retain the merits of these two methods by applying symmetry projection to broken symmetry coupled cluster wave functions. The non-orthogonal nature of states resulting from the application of symmetry projection operators furnishes particle-hole excitations to all orders, thus creating an obstacle for the exact evaluation of overlaps. Here we provide a solution via a disentanglement framework theory that can be approximated rigorously and systematically. Results of projected coupled cluster theory are presented for molecules and the Hubbard model, showing that spin projection significantly improves unrestricted coupled cluster theory while restoring good quantum numbers. The energy of projected coupled cluster theory reduces to the unprojected one in the thermodynamic limit, albeit at a much slower rate than projected Hartree-Fock.

  18. Globular Clusters - Guides to Galaxies

    CERN Document Server

    Richtler, Tom; Joint ESO-FONDAP Workshop on Globular Clusters

    2009-01-01

    The principal question of whether and how globular clusters can contribute to a better understanding of galaxy formation and evolution is perhaps the main driving force behind the overall endeavour of studying globular cluster systems. Naturally, this splits up into many individual problems. The objective of the Joint ESO-FONDAP Workshop on Globular Clusters - Guides to Galaxies was to bring together researchers, both observational and theoretical, to present and discuss the most recent results. Topics covered in these proceedings are: internal dynamics of globular clusters and interaction with host galaxies (tidal tails, evolution of cluster masses), accretion of globular clusters, detailed descriptions of nearby cluster systems, ultracompact dwarfs, formations of massive clusters in mergers and elsewhere, the ACS Virgo survey, galaxy formation and globular clusters, dynamics and kinematics of globular cluster systems and dark matter-related problems. With its wide coverage of the topic, this book constitute...

  19. An Experimental Observation of Axial Variation of Average Size of Methane Clusters in a Gas Jet

    International Nuclear Information System (INIS)

    Ji-Feng, Han; Chao-Wen, Yang; Jing-Wei, Miao; Jian-Feng, Lu; Meng, Liu; Xiao-Bing, Luo; Mian-Gong, Shi

    2010-01-01

    Axial variation of average size of methane clusters in a gas jet produced by supersonic expansion of methane through a cylindrical nozzle of 0.8 mm in diameter is observed using a Rayleigh scattering method. The scattered light intensity exhibits a power scaling on the backing pressure ranging from 16 to 50 bar, and the power is strongly Z dependent varying from 8.4 (Z = 3 mm) to 5.4 (Z = 11 mm), which is much larger than that of the argon cluster. The scattered light intensity versus axial position shows that the position of 5 mm has the maximum signal intensity. The estimation of the average cluster size on axial position Z indicates that the cluster growth process goes forward until the maximum average cluster size is reached at Z = 9 mm, and the average cluster size will decrease gradually for Z > 9 mm

  20. A Unified Simulation Approach for the Fast Outage Capacity Evaluation over Generalized Fading Channels

    KAUST Repository

    Rached, Nadhir B.; Kammoun, Abla; Alouini, Mohamed-Slim; Tempone, Raul

    2016-01-01

    The outage capacity (OC) is among the most important performance metrics of communication systems over fading channels. The evaluation of the OC, when equal gain combining (EGC) or maximum ratio combining (MRC) diversity techniques are employed

  1. What factors influence mitigative capacity?

    International Nuclear Information System (INIS)

    Winkler, Harald; Baumert, Kevin; Blanchard, Odile; Burch, Sarah; Robinson, John

    2007-01-01

    This article builds on Yohe's seminal piece on mitigative capacity, which elaborates 'determinants' of mitigative capacity, also reflected in the IPCC's third assessment report. We propose a revised definition, where mitigative capacity is a country's ability to reduce anthropogenic greenhouse gas emissions or enhance natural sinks. By 'ability' we mean skills, competencies, fitness, and proficiencies that a country has attained which can contribute to GHG emissions mitigation. A conceptual framework is proposed, linking mitigative capacity to a country's sustainable development path, and grouping the factors influencing mitigative capacity into three main sets: economic factors, institutional ones, and technology. Both quantitative and qualitative analysis of factors is presented, showing how these factors vary across countries. We suggest that it is the interplay between the three economic factors-income, abatement cost and opportunity cost-that shape mitigative capacity. We find that income is an important economic factor influencing mitigative capacity, while abatement cost is important in turning mitigative capacity into actual mitigation. Technology is a critical mitigative capacity, including the ability to absorb existing climate-friendly technologies or to develop innovative ones. Institutional factors that promote mitigative capacity include the effectiveness of government regulation, clear market rules, a skilled work force and public awareness. We briefly investigate such as high abatement cost or lack of political willingness that prevent mitigative capacity from being translated into mitigation

  2. Radio-capacity of ecosystems

    International Nuclear Information System (INIS)

    Kultakhmedov, Yu.; Kultakhmedova-Vyshnyakova, V.

    1997-01-01

    This paper consider a universal approach to ecosystems of different types, based on representation of their radio-capacity. The concept of ecosystem includes reproduction of components (bio-productivity) and conditions such as maintaining of environment quality. Radio-capacity in the case of radionuclide pollution appears in accumulation and redistribution of radionuclides in the ecosystem. As a result the radionuclides are redistributed and buried in soil or lake bottom sediments. Estimation models for the radio-capacity of water and terrestrial ecosystems are represented. The calculations of the radio-capacity factor of water ecosystems are performed, and the high radio-capacity of a freshwater reservoir (F=0.6-0.8) and extremely high radio-capacity of a reservoir cascade (F c =0.99) is shown material from the Dnieper's cascade reservoirs. The methods of radio-capacity estimation of agroecosystems, wood and marine ecosystems are developed. (authors)

  3. Overview of Maximum Power Point Tracking Techniques for Photovoltaic Energy Production Systems

    DEFF Research Database (Denmark)

    Koutroulis, Eftichios; Blaabjerg, Frede

    2015-01-01

    A substantial growth of the installed photovoltaic systems capacity has occurred around the world during the last decade, thus enhancing the availability of electric energy in an environmentally friendly way. The maximum power point tracking technique enables maximization of the energy production...... of photovoltaic sources during stochastically varying solar irradiation and ambient temperature conditions. Thus, the overall efficiency of the photovoltaic energy production system is increased. Numerous techniques have been presented during the last decade for implementing the maximum power point tracking...... process in a photovoltaic system. This article provides an overview of the operating principles of these techniques, which are suited for either uniform or non-uniform solar irradiation conditions. The operational characteristics and implementation requirements of these maximum power point tracking...

  4. Urban development control based on transportation carrying capacity

    Science.gov (United States)

    Miharja, M.; Sjafruddin, A. H.

    2017-06-01

    Severe transportation problems in Indonesian urban areas are stimulated by one fundamental factor, namely lack of awareness on transportation carrying capacity in these areas development control. Urban land use development towards more physical coverage is typically not related with the capability of transportation system to accommodate additional trips volume. Lack of clear connection between development permit with its implication on the transportation side has led to a phenomenon of exceeding transport demand over supply capacity. This paper discusses the concept of urban land use development control which will be related with transport carrying capacity. The discussion would cover both supply and demand sides of transportation. From supply side, the analysis regarding the capacity of transport system would take both existing as well as potential road network capacity could be developed. From demand side, the analysis would be through the control of a maximum floor area and public transport provision. Allowed maximum floor area for development would be at the level of generating traffic at reasonable volume. Ultimately, the objective of this paper is to introduce model to incorporate transport carrying capacity in Indonesian urban land use development control.

  5. Binary model for the coma cluster of galaxies

    International Nuclear Information System (INIS)

    Valtonen, M.J.; Byrd, G.G.

    1979-01-01

    We study the dynamics of galaxies in the Coma cluster and find that the cluster is probably dominated by a central binary of galaxies NGC 4874--NGC4889. We estimate their total mass to be about 3 x 10 14 M/sub sun/ by two independent methods (assuming in Hubble constant of 100 km s -1 Mpc -1 ). This binary is efficient in dynamically ejecting smaller galaxies, some of of which are seen in projection against the inner 3 0 radius of the cluster and which, if erroneously considered as bound members, cause a serious overestimate of the mass of the entire cluster. Taking account of the ejected galaxies, we estimate the total cluster mass to be 4--9 x 10 14 M/sub sun/, with a corresponding mass-to-light ratio for a typical galaxy in the range of 20--120 solar units. The origin of the secondary maximum observed in the radial surface density profile is studied. We consider it to be a remnant of a shell of galaxies which formed around the central binary. This shell expanded, then collapsed into the binary, and is now reexpanding. This is supported by the coincidence of the minimum in the cluster eccentricity and radical velocity dispersion at the same radial distance as the secondary maximum. Numerical simulations of a cluster model with a massive central binary and a spherical shell of test particles are performed, and they reproduce the observed shape, galaxy density, and radial velocity distributions in the Coma cluster fairly well. Consequences of extending the model to other clusters are discussed

  6. Peak capacity and peak capacity per unit time in capillary and microchip zone electrophoresis.

    Science.gov (United States)

    Foley, Joe P; Blackney, Donna M; Ennis, Erin J

    2017-11-10

    The origins of the peak capacity concept are described and the important contributions to the development of that concept in chromatography and electrophoresis are reviewed. Whereas numerous quantitative expressions have been reported for one- and two-dimensional separations, most are focused on chromatographic separations and few, if any, quantitative unbiased expressions have been developed for capillary or microchip zone electrophoresis. Making the common assumption that longitudinal diffusion is the predominant source of zone broadening in capillary electrophoresis, analytical expressions for the peak capacity are derived, first in terms of migration time, diffusion coefficient, migration distance, and desired resolution, and then in terms of the remaining underlying fundamental parameters (electric field, electroosmotic and electrophoretic mobilities) that determine the migration time. The latter expressions clearly illustrate the direct square root dependence of peak capacity on electric field and migration distance and the inverse square root dependence on solute diffusion coefficient. Conditions that result in a high peak capacity will result in a low peak capacity per unit time and vice-versa. For a given symmetrical range of relative electrophoretic mobilities for co- and counter-electroosmotic species (cations and anions), the peak capacity increases with the square root of the electric field even as the temporal window narrows considerably, resulting in a significant reduction in analysis time. Over a broad relative electrophoretic mobility interval [-0.9, 0.9], an approximately two-fold greater amount of peak capacity can be generated for counter-electroosmotic species although it takes about five-fold longer to do so, consistent with the well-known bias in migration time and resolving power for co- and counter-electroosmotic species. The optimum lower bound of the relative electrophoretic mobility interval [μ r,Z , μ r,A ] that provides the maximum

  7. Exotic cluster structures on

    CERN Document Server

    Gekhtman, M; Vainshtein, A

    2017-01-01

    This is the second paper in the series of papers dedicated to the study of natural cluster structures in the rings of regular functions on simple complex Lie groups and Poisson-Lie structures compatible with these cluster structures. According to our main conjecture, each class in the Belavin-Drinfeld classification of Poisson-Lie structures on \\mathcal{G} corresponds to a cluster structure in \\mathcal{O}(\\mathcal{G}). The authors have shown before that this conjecture holds for any \\mathcal{G} in the case of the standard Poisson-Lie structure and for all Belavin-Drinfeld classes in SL_n, n<5. In this paper the authors establish it for the Cremmer-Gervais Poisson-Lie structure on SL_n, which is the least similar to the standard one.

  8. From superdeformation to clusters

    Energy Technology Data Exchange (ETDEWEB)

    Betts, R R [Argonne National Lab., IL (United States). Physics Div.

    1992-08-01

    Much of the discussion at the conference centred on superdeformed states and their study by precise gamma spectrometry. The author suggests that the study of superdeformation by fission fragments and by auto-scattering is of importance, and may become more important. He concludes that there exists clear evidence of shell effects at extreme deformation in light nuclei studied by fission or cluster decay. The connection between the deformed shell model and the multi-center shell model can be exploited to give give insight into the cluster structure of these extremely deformed states, and also gives hope of a spectroscopy based on selection rules for cluster decay. A clear disadvantage at this stage is inability to make this spectroscopy more quantitative through calculation of the decay widths. The introduction of a new generation of high segmentation, high resolution, particle arrays has and will have a major impact on this aspect of the study of highly deformed nuclei. 20 refs., 16 figs.

  9. Capacity building for HIA

    Directory of Open Access Journals (Sweden)

    Gabriel Gulis PhD

    2007-09-01

    Full Text Available

    Background: To integrate health impact assessment (HIA into existing decision-making processes requires not only methods and procedures but also well-trained experts, aware policy makers and appropriate institutions. Capacity building is the assistance which is provided to entities, which have a need to develop a certain skill or competence, or for general upgrading of performance ability. If a new technique is planned to be introduced there is a need for capacity building with no respect to levels (local, regional, national, international or sectors (health, environment, finance, social care, education, etc.. As such, HIA is a new technique for most of the new Member States and accession countries of the European Union.

    Methods: To equip individuals with the understanding and skills needed to launch a HIA or be aware of the availability of this methodology and to access information, knowledge and training, we focused on the organization of workshops in participating countries. The workshops served also as pilot events to test a “curriculum” for HIA; a set of basic topics and presentations had been developed to be tested during workshops. In spite of classical in-class workshops we aimed to organize e-learning events as a way to over come the “busyness” problem of decision makers.

    Results: Throughout March – October 2006 we organized and ran 7 workshops in Denmark, Turkey, Lithuania, Poland, Bulgaria, Slovak Republic and Hungary. Participants came from the public health sector (141, non-public health decision makers (113 and public health students (100. A concise curriculum was developed and tested during these workshops. Participants developed a basic understanding of HIA, skills to develop and use their own screening tools as well as scoping.Within the workshop in Denmark we tested an online, real-time Internet based training method; participants highly welcomed this

  10. Maximum entropy production rate in quantum thermodynamics

    Energy Technology Data Exchange (ETDEWEB)

    Beretta, Gian Paolo, E-mail: beretta@ing.unibs.i [Universita di Brescia, via Branze 38, 25123 Brescia (Italy)

    2010-06-01

    In the framework of the recent quest for well-behaved nonlinear extensions of the traditional Schroedinger-von Neumann unitary dynamics that could provide fundamental explanations of recent experimental evidence of loss of quantum coherence at the microscopic level, a recent paper [Gheorghiu-Svirschevski 2001 Phys. Rev. A 63 054102] reproposes the nonlinear equation of motion proposed by the present author [see Beretta G P 1987 Found. Phys. 17 365 and references therein] for quantum (thermo)dynamics of a single isolated indivisible constituent system, such as a single particle, qubit, qudit, spin or atomic system, or a Bose-Einstein or Fermi-Dirac field. As already proved, such nonlinear dynamics entails a fundamental unifying microscopic proof and extension of Onsager's reciprocity and Callen's fluctuation-dissipation relations to all nonequilibrium states, close and far from thermodynamic equilibrium. In this paper we propose a brief but self-contained review of the main results already proved, including the explicit geometrical construction of the equation of motion from the steepest-entropy-ascent ansatz and its exact mathematical and conceptual equivalence with the maximal-entropy-generation variational-principle formulation presented in Gheorghiu-Svirschevski S 2001 Phys. Rev. A 63 022105. Moreover, we show how it can be extended to the case of a composite system to obtain the general form of the equation of motion, consistent with the demanding requirements of strong separability and of compatibility with general thermodynamics principles. The irreversible term in the equation of motion describes the spontaneous attraction of the state operator in the direction of steepest entropy ascent, thus implementing the maximum entropy production principle in quantum theory. The time rate at which the path of steepest entropy ascent is followed has so far been left unspecified. As a step towards the identification of such rate, here we propose a possible

  11. The system capacity view of aging and longevity

    Institute of Scientific and Technical Information of China (English)

    Jing-Dong J.Han; Lei Hou; Na Sun; Chi Xu; Joseph McDermott; Dan Wang

    2017-01-01

    Background:Aging is a complex systems level problem that needs a systems level solution.However,system models of aging and longevity,although urgently needed,are still lacking,largely due to the paucity of conceptual frameworks for modeling such a complex process.Results:We propose that aging can be viewed as a decline in system capacity,defined as the maximum level of output that a system produces to fulfill demands.Classical aging hallmarks and anti-aging strategies can be well-aligned to system capacity.Genetic variants responsible for lifespan variation across individuals or species can also be explained by their roles in system capacity.We further propose promising directions to develop systems approaches to modulate system capacity and thus extend both healthspan and lifespan.Conclusions:The system capacity model of aging provides an opportunity to examine aging at the systems level.This model predicts that the extent to which aging can be modulated is normally limited by the upper bound of the system capacity of a species.Within such a boundary,aging can be delayed by moderately increasing an individual's system capacity.Beyond such a boundary,increasing the upper bound is required,which is not unrealistic given the unlimited potential of regenerative medicine in the future,but it requires increasing the capacity of the whole system instead of only part of it.

  12. Refractory chronic cluster headache

    DEFF Research Database (Denmark)

    Mitsikostas, Dimos D; Edvinsson, Lars; Jensen, Rigmor H

    2014-01-01

    Chronic cluster headache (CCH) often resists to prophylactic pharmaceutical treatments resulting in patients' life damage. In this rare but pragmatic situation escalation to invasive management is needed but framing criteria are lacking. We aimed to reach a consensus for refractory CCH definition...... for clinical and research use. The preparation of the final consensus followed three stages. Internal between authors, a larger between all European Headache Federation members and finally an international one among all investigators that have published clinical studies on cluster headache the last five years...

  13. I Cluster geografici

    Directory of Open Access Journals (Sweden)

    Maurizio Rosina

    2010-03-01

    Full Text Available Geographic ClustersOver the past decade, public alphanumeric database have been growing at exceptional rate. Most of data can be georeferenced, so that is possible gaining new knowledge from such databases. The contribution of this paper is two-fold. We first present a model of geographic clusters, which uses only geographic and functionally data properties. The model is useful to process huge amount of public/government data, even daily upgrading. After that, we merge the model into the framework GEOPOI (GEOcoding Points Of Interest, and show some graphic map results.

  14. I Cluster geografici

    Directory of Open Access Journals (Sweden)

    Maurizio Rosina

    2010-03-01

    Full Text Available Geographic Clusters Over the past decade, public alphanumeric database have been growing at exceptional rate. Most of data can be georeferenced, so that is possible gaining new knowledge from such databases. The contribution of this paper is two-fold. We first present a model of geographic clusters, which uses only geographic and functionally data properties. The model is useful to process huge amount of public/government data, even daily upgrading. After that, we merge the model into the framework GEOPOI (GEOcoding Points Of Interest, and show some graphic map results.

  15. Clustering via Kernel Decomposition

    DEFF Research Database (Denmark)

    Have, Anna Szynkowiak; Girolami, Mark A.; Larsen, Jan

    2006-01-01

    Methods for spectral clustering have been proposed recently which rely on the eigenvalue decomposition of an affinity matrix. In this work it is proposed that the affinity matrix is created based on the elements of a non-parametric density estimator. This matrix is then decomposed to obtain...... posterior probabilities of class membership using an appropriate form of nonnegative matrix factorization. The troublesome selection of hyperparameters such as kernel width and number of clusters can be obtained using standard cross-validation methods as is demonstrated on a number of diverse data sets....

  16. Defining clusters in APT reconstructions of ODS steels.

    Science.gov (United States)

    Williams, Ceri A; Haley, Daniel; Marquis, Emmanuelle A; Smith, George D W; Moody, Michael P

    2013-09-01

    Oxide nanoclusters in a consolidated Fe-14Cr-2W-0.3Ti-0.3Y₂O₃ ODS steel and in the alloy powder after mechanical alloying (but before consolidation) are investigated by atom probe tomography (APT). The maximum separation method is a standard method to define and characterise clusters from within APT data, but this work shows that the extent of clustering between the two materials is sufficiently different that the nanoclusters in the mechanically alloyed powder and in the consolidated material cannot be compared directly using the same cluster selection parameters. As the cluster selection parameters influence the size and composition of the clusters significantly, a procedure to optimise the input parameters for the maximum separation method is proposed by sweeping the d(max) and N(min) parameter space. By applying this method of cluster parameter selection combined with a 'matrix correction' to account for trajectory aberrations, differences in the oxide nanoclusters can then be reliably quantified. Copyright © 2012 Elsevier B.V. All rights reserved.

  17. Determination of the maximum-depth to potential field sources by a maximum structural index method

    Science.gov (United States)

    Fedi, M.; Florio, G.

    2013-01-01

    A simple and fast determination of the limiting depth to the sources may represent a significant help to the data interpretation. To this end we explore the possibility of determining those source parameters shared by all the classes of models fitting the data. One approach is to determine the maximum depth-to-source compatible with the measured data, by using for example the well-known Bott-Smith rules. These rules involve only the knowledge of the field and its horizontal gradient maxima, and are independent from the density contrast. Thanks to the direct relationship between structural index and depth to sources we work out a simple and fast strategy to obtain the maximum depth by using the semi-automated methods, such as Euler deconvolution or depth-from-extreme-points method (DEXP). The proposed method consists in estimating the maximum depth as the one obtained for the highest allowable value of the structural index (Nmax). Nmax may be easily determined, since it depends only on the dimensionality of the problem (2D/3D) and on the nature of the analyzed field (e.g., gravity field or magnetic field). We tested our approach on synthetic models against the results obtained by the classical Bott-Smith formulas and the results are in fact very similar, confirming the validity of this method. However, while Bott-Smith formulas are restricted to the gravity field only, our method is applicable also to the magnetic field and to any derivative of the gravity and magnetic field. Our method yields a useful criterion to assess the source model based on the (∂f/∂x)max/fmax ratio. The usefulness of the method in real cases is demonstrated for a salt wall in the Mississippi basin, where the estimation of the maximum depth agrees with the seismic information.

  18. Integrated flexible capacity and inventory management under flexible capacity uncertainty

    OpenAIRE

    Paç, Mehmet Fazıl

    2006-01-01

    Cataloged from PDF version of article. In a manufacturing environment with volatile demand, inventory management can be coupled with dynamic capacity adjustments for handling the fluctuations more effectively. In this study we consider the integrated management of inventory and flexible capacity management under seasonal stochastic demand and uncertain labor supply. The capacity planning problem is investigated from the workforce planning perspective. We consider a manufactu...

  19. Construction and Capacity Analysis of High-Rank LoS MIMO Channels in High Speed Railway Scenarios

    Directory of Open Access Journals (Sweden)

    Jingya Yang

    2012-01-01

    Full Text Available The validity of the maximum capacity criterion applied to realize high-rank line-of-sight (LoS multiple-input multiple-output (MIMO channels is investigated for high speed railway scenarios. Performance is evaluated by ergodic capacity. Numerical results demonstrate that by simply adjusting antenna spacing according to the maximum capacity criterion, significant capacity gains are achievable. We find relatively low sensitivity of the system to displacements from the optimal point and angle in relatively short range. Thus, we present two proposals to reconfigure antenna arrays so as to maximize LoS MIMO capacity in the high speed railway scenarios

  20. Cluster size matters: Size-driven performance of subnanometer clusters in catalysis, electrocatalysis and Li-air batteries

    Science.gov (United States)

    Vajda, Stefan

    2015-03-01

    This paper discusses the strongly size-dependent performance of subnanometer cluster based catalysts in 1) heterogeneous catalysis, 2) electrocatalysis and 3) Li-air batteries. The experimental studies are based on I. fabrication of ultrasmall clusters with atomic precision control of particle size and their deposition on oxide and carbon based supports; II. test of performance, III. in situand ex situ X-ray characterization of cluster size, shape and oxidation state; and IV.electron microscopies. Heterogeneous catalysis. The pronounced effect of cluster size and support on the performance of the catalyst (catalyst activity and the yield of Cn products) will be illustrated on the example of nickel and cobalt clusters in Fischer-Tropsch reaction. Electrocatalysis. The study of the oxygen evolution reaction (OER) on size-selected palladium clusters supported on ultrananocrystalline diamond show pronounced size effects. While Pd4 clusters show no reaction, Pd6 and Pd17 clusters are among the most active catalysts known (in in terms of turnover rate per Pd atom). The system (soft-landed Pd4, Pd6, or Pd17 clusters on an UNCD Si coated electrode) shows stable electrochemical potentials over several cycles, and the characterization of the electrodes show no evidence for evolution or dissolution of either the support Theoretical calculations suggest that this striking difference may be a demonstration that bridging Pd-Pd sites, which are only present in three-dimensional clusters, are active for the oxygen evolution reaction in Pd6O6. Li-air batteries. The studies show that sub-nm silver clusters have dramatic size-dependent effect on the lowering of the overpotential, charge capacity, morphology of the discharge products, as well as on the morphology of the nm size building blocks of the discharge products. The results suggest that by precise control of the active surface sites on the cathode, the performance of Li-air cells can be significantly improved

  1. Weighted Maximum-Clique Transversal Sets of Graphs

    OpenAIRE

    Chuan-Min Lee

    2011-01-01

    A maximum-clique transversal set of a graph G is a subset of vertices intersecting all maximum cliques of G. The maximum-clique transversal set problem is to find a maximum-clique transversal set of G of minimum cardinality. Motivated by the placement of transmitters for cellular telephones, Chang, Kloks, and Lee introduced the concept of maximum-clique transversal sets on graphs in 2001. In this paper, we study the weighted version of the maximum-clique transversal set problem for split grap...

  2. Pattern formation, logistics, and maximum path probability

    Science.gov (United States)

    Kirkaldy, J. S.

    1985-05-01

    The concept of pattern formation, which to current researchers is a synonym for self-organization, carries the connotation of deductive logic together with the process of spontaneous inference. Defining a pattern as an equivalence relation on a set of thermodynamic objects, we establish that a large class of irreversible pattern-forming systems, evolving along idealized quasisteady paths, approaches the stable steady state as a mapping upon the formal deductive imperatives of a propositional function calculus. In the preamble the classical reversible thermodynamics of composite systems is analyzed as an externally manipulated system of space partitioning and classification based on ideal enclosures and diaphragms. The diaphragms have discrete classification capabilities which are designated in relation to conserved quantities by descriptors such as impervious, diathermal, and adiabatic. Differentiability in the continuum thermodynamic calculus is invoked as equivalent to analyticity and consistency in the underlying class or sentential calculus. The seat of inference, however, rests with the thermodynamicist. In the transition to an irreversible pattern-forming system the defined nature of the composite reservoirs remains, but a given diaphragm is replaced by a pattern-forming system which by its nature is a spontaneously evolving volume partitioner and classifier of invariants. The seat of volition or inference for the classification system is thus transferred from the experimenter or theoretician to the diaphragm, and with it the full deductive facility. The equivalence relations or partitions associated with the emerging patterns may thus be associated with theorems of the natural pattern-forming calculus. The entropy function, together with its derivatives, is the vehicle which relates the logistics of reservoirs and diaphragms to the analog logistics of the continuum. Maximum path probability or second-order differentiability of the entropy in isolation are

  3. Multi-Optimisation Consensus Clustering

    Science.gov (United States)

    Li, Jian; Swift, Stephen; Liu, Xiaohui

    Ensemble Clustering has been developed to provide an alternative way of obtaining more stable and accurate clustering results. It aims to avoid the biases of individual clustering algorithms. However, it is still a challenge to develop an efficient and robust method for Ensemble Clustering. Based on an existing ensemble clustering method, Consensus Clustering (CC), this paper introduces an advanced Consensus Clustering algorithm called Multi-Optimisation Consensus Clustering (MOCC), which utilises an optimised Agreement Separation criterion and a Multi-Optimisation framework to improve the performance of CC. Fifteen different data sets are used for evaluating the performance of MOCC. The results reveal that MOCC can generate more accurate clustering results than the original CC algorithm.

  4. Photochemistry in rare gas clusters

    International Nuclear Information System (INIS)

    Moeller, T.; Haeften, K. von; Pietrowski, R. von

    1999-01-01

    In this contribution photochemical processes in pure rare gas clusters will be discussed. The relaxation dynamics of electronically excited He clusters is investigated with luminescence spectroscopy. After electronic excitation of He clusters many sharp lines are observed in the visible and infrared spectral range which can be attributed to He atoms and molecules desorbing from the cluster. It turns out that the desorption of electronically excited He atoms and molecules is an important decay channel. The findings for He clusters are compared with results for Ar clusters. While desorption of electronically excited He atoms is observed for all clusters containing up to several thousand atoms a corresponding process in Ar clusters is only observed for very small clusters (N<10). (orig.)

  5. Photochemistry in rare gas clusters

    Energy Technology Data Exchange (ETDEWEB)

    Moeller, T.; Haeften, K. von; Pietrowski, R. von [Deutsches Elektronen-Synchrotron (DESY), Hamburg (Germany). Hamburger Synchrotronstrahlungslabor; Laarman, T. [Universitaet Hamburg, II. Institut fuer Experimentalphysik, Luruper Chaussee 149, D-22761 Hamburg (Germany)

    1999-12-01

    In this contribution photochemical processes in pure rare gas clusters will be discussed. The relaxation dynamics of electronically excited He clusters is investigated with luminescence spectroscopy. After electronic excitation of He clusters many sharp lines are observed in the visible and infrared spectral range which can be attributed to He atoms and molecules desorbing from the cluster. It turns out that the desorption of electronically excited He atoms and molecules is an important decay channel. The findings for He clusters are compared with results for Ar clusters. While desorption of electronically excited He atoms is observed for all clusters containing up to several thousand atoms a corresponding process in Ar clusters is only observed for very small clusters (N<10). (orig.)

  6. Globular clusters, old and young

    International Nuclear Information System (INIS)

    Samus', N.N.

    1984-01-01

    The problem of similarity of and difference in the globular and scattered star clusters is considered. Star clusters in astronomy are related either to globular or to scattered ones according to the structure of Hertzsprung-Russell diagram constructed for star clusters, but not according to the appearance. The qlobular clusters in the Galaxy are composed of giants and subgiants, which testifies to the old age of the globular clusters. The Globular clusters in the Magellanic clouds are classified into ''red'' ones - similar to the globular clusters of the Galaxy, and ''blue'' ones - similar to them in appearance but differing extremely by the star composition and so by the age. The old star clusters are suggested to be called globular ones, while another name (''populous'', for example) is suggested to be used for other clusters similar to globular ones only in appearance

  7. Globular clusters and galaxy halos

    International Nuclear Information System (INIS)

    Van Den Bergh, S.

    1984-01-01

    Using semipartial correlation coefficients and bootstrap techniques, a study is made of the important features of globular clusters with respect to the total number of galaxy clusters and dependence of specific galaxy cluster on parent galaxy type, cluster radii, luminosity functions and cluster ellipticity. It is shown that the ellipticity of LMC clusters correlates significantly with cluster luminosity functions, but not with cluster age. The cluter luminosity value above which globulars are noticeably flattened may differ by a factor of about 100 from galaxy to galaxy. Both in the Galaxy and in M31 globulars with small core radii have a Gaussian distribution over luminosity, whereas clusters with large core radii do not. In the cluster systems surrounding the Galaxy, M31 and NGC 5128 the mean radii of globular clusters was found to increase with the distance from the nucleus. Central galaxies in rich clusters have much higher values for specific globular cluster frequency than do other cluster ellipticals, suggesting that such central galaxies must already have been different from normal ellipticals at the time they were formed

  8. Clustering of resting state networks.

    Directory of Open Access Journals (Sweden)

    Megan H Lee

    Full Text Available The goal of the study was to demonstrate a hierarchical structure of resting state activity in the healthy brain using a data-driven clustering algorithm.The fuzzy-c-means clustering algorithm was applied to resting state fMRI data in cortical and subcortical gray matter from two groups acquired separately, one of 17 healthy individuals and the second of 21 healthy individuals. Different numbers of clusters and different starting conditions were used. A cluster dispersion measure determined the optimal numbers of clusters. An inner product metric provided a measure of similarity between different clusters. The two cluster result found the task-negative and task-positive systems. The cluster dispersion measure was minimized with seven and eleven clusters. Each of the clusters in the seven and eleven cluster result was associated with either the task-negative or task-positive system. Applying the algorithm to find seven clusters recovered previously described resting state networks, including the default mode network, frontoparietal control network, ventral and dorsal attention networks, somatomotor, visual, and language networks. The language and ventral attention networks had significant subcortical involvement. This parcellation was consistently found in a large majority of algorithm runs under different conditions and was robust to different methods of initialization.The clustering of resting state activity using different optimal numbers of clusters identified resting state networks comparable to previously obtained results. This work reinforces the observation that resting state networks are hierarchically organized.

  9. Assessment of genetic divergence in tomato through agglomerative hierarchical clustering and principal component analysis

    International Nuclear Information System (INIS)

    Iqbal, Q.; Saleem, M.Y.; Hameed, A.; Asghar, M.

    2014-01-01

    For the improvement of qualitative and quantitative traits, existence of variability has prime importance in plant breeding. Data on different morphological and reproductive traits of 47 tomato genotypes were analyzed for correlation,agglomerative hierarchical clustering and principal component analysis (PCA) to select genotypes and traits for future breeding program. Correlation analysis revealed significant positive association between yield and yield components like fruit diameter, single fruit weight and number of fruits plant-1. Principal component (PC) analysis depicted first three PCs with Eigen-value higher than 1 contributing 81.72% of total variability for different traits. The PC-I showed positive factor loadings for all the traits except number of fruits plant-1. The contribution of single fruit weight and fruit diameter was highest in PC-1. Cluster analysis grouped all genotypes into five divergent clusters. The genotypes in cluster-II and cluster-V exhibited uniform maturity and higher yield. The D2 statistics confirmed highest distance between cluster- III and cluster-V while maximum similarity was observed in cluster-II and cluster-III. It is therefore suggested that crosses between genotypes of cluster-II and cluster-V with those of cluster-I and cluster-III may exhibit heterosis in F1 for hybrid breeding and for selection of superior genotypes in succeeding generations for cross breeding programme. (author)

  10. Supervised maximum-likelihood weighting of composite protein networks for complex prediction

    Directory of Open Access Journals (Sweden)

    Yong Chern Han

    2012-12-01

    Full Text Available Abstract Background Protein complexes participate in many important cellular functions, so finding the set of existent complexes is essential for understanding the organization and regulation of processes in the cell. With the availability of large amounts of high-throughput protein-protein interaction (PPI data, many algorithms have been proposed to discover protein complexes from PPI networks. However, such approaches are hindered by the high rate of noise in high-throughput PPI data, including spurious and missing interactions. Furthermore, many transient interactions are detected between proteins that are not from the same complex, while not all proteins from the same complex may actually interact. As a result, predicted complexes often do not match true complexes well, and many true complexes go undetected. Results We address these challenges by integrating PPI data with other heterogeneous data sources to construct a composite protein network, and using a supervised maximum-likelihood approach to weight each edge based on its posterior probability of belonging to a complex. We then use six different clustering algorithms, and an aggregative clustering strategy, to discover complexes in the weighted network. We test our method on Saccharomyces cerevisiae and Homo sapiens, and show that complex discovery is improved: compared to previously proposed supervised and unsupervised weighting approaches, our method recalls more known complexes, achieves higher precision at all recall levels, and generates novel complexes of greater functional similarity. Furthermore, our maximum-likelihood approach allows learned parameters to be used to visualize and evaluate the evidence of novel predictions, aiding human judgment of their credibility. Conclusions Our approach integrates multiple data sources with supervised learning to create a weighted composite protein network, and uses six clustering algorithms with an aggregative clustering strategy to

  11. Clustering in Ethiopia

    African Journals Online (AJOL)

    Background: The importance of local variations in patterns of health and disease are increasingly recognised, but, particularly in the case of tropical infections, available methods and resources for characterising disease clusters in time and space are limited. Whilst the Global Positioning System. (GPS) allows accurate and ...

  12. Hardness of Clustering

    Indian Academy of Sciences (India)

    First page Back Continue Last page Overview Graphics. Hardness of Clustering. Both k-means and k-medians intractable (when n and d are both inputs even for k =2). The best known deterministic algorithms. are based on Voronoi partitioning that. takes about time. Need for approximation – “close” to optimal.

  13. On small clusters

    International Nuclear Information System (INIS)

    Bernardes, N.

    1984-01-01

    A discussion is presented of zero-point motion effects on the binding energy of a small cluster of identical particles interacting through short range attractive-repulsive forces. The model is appropriate to a discussion of both Van der Waals as well as nuclear forces. (Author) [pt

  14. Emergence of regional clusters

    DEFF Research Database (Denmark)

    Dahl, Michael S.; Østergaard, Christian Richter; Dalum, Bent

    2010-01-01

    The literature on regional clusters has increased considerably during the last decade. The emergence and growth patterns are usually explained by such factors as unique local culture, regional capabilities, tacit knowledge or the existence of location-specific externalities (knowledge spillovers...

  15. Supersonic copper clusters

    International Nuclear Information System (INIS)

    Powers, D.E.; Hansen, S.G.; Geusic, M.E.; Michalopoulos, D.L.; Smalley, R.E.

    1983-01-01

    Copper clusters ranging in size from 1 to 29 atoms have been prepared in a supersonic beam by laser vaporization of a rotating copper target rod within the throat of a pulsed supersonic nozzle using helium for the carrier gas. The clusters were cooled extensively in the supersonic expansion [T(translational) 1 to 4 K, T(rotational) = 4 K, T(vibrational) = 20 to 70 K]. These clusters were detected in the supersonic beam by laser photoionization with time-of-flight mass analysis. Using a number of fixed frequency outputs of an exciplex laser, the threshold behavior of the photoionization cross section was monitored as a function of cluster size.nce two-photon ionization (R2PI) with mass selective detection allowed the detection of five new electronic band systems in the region between 2690 and 3200 A, for each of the three naturally occurring isotopic forms of Cu 2 . In the process of scanning the R2PI spectrum of these new electronic states, the ionization potential of the copper dimer was determined to be 7.894 +- 0.015 eV

  16. Greedy subspace clustering.

    Science.gov (United States)

    2016-09-01

    We consider the problem of subspace clustering: given points that lie on or near the union of many low-dimensional linear subspaces, recover the subspaces. To this end, one first identifies sets of points close to the same subspace and uses the sets ...

  17. Available transmission capacity assessment

    Directory of Open Access Journals (Sweden)

    Škokljev Ivan

    2012-01-01

    Full Text Available Effective power system operation requires the analysis of vast amounts of information. Power market activities expose power transmission networks to high-level power transactions that threaten normal, secure operation of the power system. When there are service requests for a specific sink/source pair in a transmission system, the transmission system operator (TSO must allocate the available transfer capacity (ATC. It is common that ATC has a single numerical value. Additionally, the ATC must be calculated for the base case configuration of the system, while generation dispatch and topology remain unchanged during the calculation. Posting ATC on the internet should benefit prospective users by aiding them in formulating their requests. However, a single numerical value of ATC offers little for prospect for analysis, planning, what-if combinations, etc. A symbolic approach to the power flow problem (DC power flow and ATC offers a numerical computation at the very end, whilst the calculation beforehand is performed by using symbols for the general topology of the electrical network. Qualitative analysis of the ATC using only qualitative values, such as increase, decrease or no change, offers some new insights into ATC evaluation, multiple transactions evaluation, value of counter-flows and their impact etc. Symbolic analysis in this paper is performed after the execution of the linear, symbolic DC power flow. As control variables, the mathematical model comprises linear security constraints, ATC, PTDFs and transactions. The aim is to perform an ATC sensitivity study on a five nodes/seven lines transmission network, used for zonal market activities tests. A relatively complicated environment with twenty possible bilateral transactions is observed.

  18. Low-SNR Capacity of MIMO Optical Intensity Channels

    KAUST Repository

    Chaaban, Anas

    2017-09-18

    The capacity of the multiple-input multiple-output (MIMO) optical intensity channel is studied, under both average and peak intensity constraints. We focus on low SNR, which can be modeled as the scenario where both constraints proportionally vanish, or where the peak constraint is held constant while the average constraint vanishes. A capacity upper bound is derived, and is shown to be tight at low SNR under both scenarios. The capacity achieving input distribution at low SNR is shown to be a maximally-correlated vector-binary input distribution. Consequently, the low-SNR capacity of the channel is characterized. As a byproduct, it is shown that for a channel with peak intensity constraints only, or with peak intensity constraints and individual (per aperture) average intensity constraints, a simple scheme composed of coded on-off keying, spatial repetition, and maximum-ratio combining is optimal at low SNR.

  19. Low-SNR Capacity of MIMO Optical Intensity Channels

    KAUST Repository

    Chaaban, Anas; Rezki, Zouheir; Alouini, Mohamed-Slim

    2017-01-01

    The capacity of the multiple-input multiple-output (MIMO) optical intensity channel is studied, under both average and peak intensity constraints. We focus on low SNR, which can be modeled as the scenario where both constraints proportionally vanish, or where the peak constraint is held constant while the average constraint vanishes. A capacity upper bound is derived, and is shown to be tight at low SNR under both scenarios. The capacity achieving input distribution at low SNR is shown to be a maximally-correlated vector-binary input distribution. Consequently, the low-SNR capacity of the channel is characterized. As a byproduct, it is shown that for a channel with peak intensity constraints only, or with peak intensity constraints and individual (per aperture) average intensity constraints, a simple scheme composed of coded on-off keying, spatial repetition, and maximum-ratio combining is optimal at low SNR.

  20. Cluster headache - clinical pattern and a new severity scale in a Swedish cohort.

    Science.gov (United States)

    Steinberg, Anna; Fourier, Carmen; Ran, Caroline; Waldenlind, Elisabet; Sjöstrand, Christina; Belin, Andrea Carmine

    2018-06-01

    Background The aim of this study was to investigate clinical features of a cluster headache cohort in Sweden and to construct and test a new scale for grading severity. Methods Subjects were identified by screening medical records for the ICD 10 code G44.0, that is, cluster headache. Five hundred participating research subjects filled in a questionnaire including personal, demographic and medical aspects. We constructed a novel scale for grading cluster headache in this cohort: The Cluster Headache Severity Scale, which included number of attacks per day, attack and period duration. The lowest total score was three and the highest 12, and we used the Cluster Headache Severity Scale to grade subjects suffering from cluster headache. We further implemented the scale by defining a cluster headache maximum severity subgroup with a high Cluster Headache Severity Scale score ≥ 9. Results A majority (66.7%) of the patients reported that attacks appear at certain time intervals. In addition, cluster headache patients who were current tobacco users or had a history of tobacco consumption had a later age of disease onset (31.7 years) compared to non-tobacco users (28.5 years). The Cluster Headache Severity Scale score was higher in the patient group reporting sporadic or no alcohol intake than in the groups reporting an alcohol consumption of three to four standard units per week or more. Maximum severity cluster headache patients were characterised by higher age at disease onset, greater use of prophylactic medication, reduced hours of sleep, and lower alcohol consumption compared to the non-cluster headache maximum severity group. Conclusion There was a wide variation of severity grade among cluster headache patients, with a very marked impact on daily living for the most profoundly affected.

  1. Data clustering algorithms and applications

    CERN Document Server

    Aggarwal, Charu C

    2013-01-01

    Research on the problem of clustering tends to be fragmented across the pattern recognition, database, data mining, and machine learning communities. Addressing this problem in a unified way, Data Clustering: Algorithms and Applications provides complete coverage of the entire area of clustering, from basic methods to more refined and complex data clustering approaches. It pays special attention to recent issues in graphs, social networks, and other domains.The book focuses on three primary aspects of data clustering: Methods, describing key techniques commonly used for clustering, such as fea

  2. Cluster structures in light nuclei

    International Nuclear Information System (INIS)

    Horiuchi, H.

    2000-01-01

    Complete text of publication follows. Clustering in neutron-rich nuclei is discussed. To understand the novel features (1,2,3) of the clustering in neutron-rich nuclei, the basic features of the clustering in stable nuclei (4) are briefly reviewed. In neutron-rich nuclei, the requirement of the stability of clusters is questioned and the threshold rule is no more obeyed. Examples of clustering in Be and B isotopes (4,5) are discussed in some detail. Possible existence of novel type of clustering near neutron dripline is suggested (1). (author)

  3. Cluster model of the nucleus

    International Nuclear Information System (INIS)

    Horiuchi, H.; Ikeda, K.

    1986-01-01

    This article reviews the development of the cluster model study. The stress is put on two points; one is how the cluster structure has come to be regarded as a fundamental structure in light nuclei together with the shell-model structure, and the other is how at present the cluster model is extended to and connected with the studies of the various subjects many of which are in the neighbouring fields. The authors the present the main theme with detailed explanations of the fundamentals of the microscopic cluster model which have promoted the development of the cluster mode. Examples of the microscopic cluster model study of light nuclear structure are given

  4. Evaluation model for safety capacity of chemical industrial park based on acceptable regional risk

    Institute of Scientific and Technical Information of China (English)

    Guohua Chen; Shukun Wang; Xiaoqun Tan

    2015-01-01

    The paper defines the Safety Capacity of Chemical Industrial Park (SCCIP) from the perspective of acceptable regional risk. For the purpose of exploring the evaluation model for the SCCIP, a method based on quantitative risk assessment was adopted for evaluating transport risk and to confirm reasonable safety transport capacity of chemical industrial park, and then by combining with the safety storage capacity, a SCCIP evaluation model was put forward. The SCCIP was decided by the smaller one between the largest safety storage capacity and the maximum safety transport capacity, or else, the regional risk of the park will exceed the acceptable level. The developed method was applied to a chemical industrial park in Guangdong province to obtain the maximum safety transport capacity and the SCCIP. The results can be realized in the regional risk control of the park effectively.

  5. Nucleation from a cluster of inclusions, leading to void coalescense

    DEFF Research Database (Denmark)

    Tvergaard, Viggo

    2017-01-01

    A cell model analysis is used to study the nucleation and subsequent growth of voids from a non-uniform distribution of inclusions in a ductile material. Nucleation is modeled as either stress controlled or strain controlled. The special clusters considered consist of a number of uniformly spaced...... inclusions located along a plane perpendicular to the maximum principal tensile stress. A plane strain approximation is used, where the inclusions are parallel cylinders perpendicular to the plane. Clusters with different numbers of inclusions are compared with the nucleation and growth from a single...... inclusion, such that the total initial volume of the inclusions is the same for the clusters and the single inclusion. After nucleation, local void coalescence inside the clusters is accounted for, since this makes it possible to compare the rate of growth of the single larger void that results from...

  6. High-Intensity Femtosecond Laser Interaction with Rare Gas Clusters

    Institute of Scientific and Technical Information of China (English)

    林亚风; 钟钦; 曾淳; 陈哲

    2001-01-01

    With a 45 fs multiterawatt 790 nm laser system and jets of argon and krypton atomic clusters, a study of the interaction of fs intense laser pulses with large size rare gas dusters was conducted. The maximum laser intensity of about 7 × 1016 W/cm2 and dusters composed of thousands of atoms which were determined through Rayleigh scattering measurements were involved inthe experiments. On the one hand, the results indicate that the interaction is strongly cluster size dependent. The stronger the interaction, the larger the clusters are. On the other hand, a saturation followed by a drop of the energy of ions ejected from the interaction will occur when the laser intensity exceeds a definite value for clusters of a certain size.

  7. Clustering and phase transitions on a neutral landscape

    Science.gov (United States)

    Scott, Adam D.; King, Dawn M.; Marić, Nevena; Bahar, Sonya

    2013-06-01

    Recent computational studies have shown that speciation can occur under neutral conditions, i.e., when the simulated organisms all have identical fitness. These works bear comparison with mathematical studies of clustering on neutral landscapes in the context of branching and coalescing random walks. Here, we show that sympatric clustering/speciation can occur on a neutral landscape whose dimensions specify only the simulated organisms’ phenotypes. We demonstrate that clustering occurs not only in the case of assortative mating, but also in the case of asexual fission; it is not observed in the control case of random mating. We find that the population size and the number of clusters undergo a second-order non-equilibrium phase transition as the maximum mutation size is varied.

  8. Optimal wavelength band clustering for multispectral iris recognition.

    Science.gov (United States)

    Gong, Yazhuo; Zhang, David; Shi, Pengfei; Yan, Jingqi

    2012-07-01

    This work explores the possibility of clustering spectral wavelengths based on the maximum dissimilarity of iris textures. The eventual goal is to determine how many bands of spectral wavelengths will be enough for iris multispectral fusion and to find these bands that will provide higher performance of iris multispectral recognition. A multispectral acquisition system was first designed for imaging the iris at narrow spectral bands in the range of 420 to 940 nm. Next, a set of 60 human iris images that correspond to the right and left eyes of 30 different subjects were acquired for an analysis. Finally, we determined that 3 clusters were enough to represent the 10 feature bands of spectral wavelengths using the agglomerative clustering based on two-dimensional principal component analysis. The experimental results suggest (1) the number, center, and composition of clusters of spectral wavelengths and (2) the higher performance of iris multispectral recognition based on a three wavelengths-bands fusion.

  9. Algorithmic complexity of quantum capacity

    Science.gov (United States)

    Oskouei, Samad Khabbazi; Mancini, Stefano

    2018-04-01

    We analyze the notion of quantum capacity from the perspective of algorithmic (descriptive) complexity. To this end, we resort to the concept of semi-computability in order to describe quantum states and quantum channel maps. We introduce algorithmic entropies (like algorithmic quantum coherent information) and derive relevant properties for them. Then we show that quantum capacity based on semi-computable concept equals the entropy rate of algorithmic coherent information, which in turn equals the standard quantum capacity. Thanks to this, we finally prove that the quantum capacity, for a given semi-computable channel, is limit computable.

  10. Dementia, Decision Making, and Capacity.

    Science.gov (United States)

    Darby, R Ryan; Dickerson, Bradford C

    After participating in this activity, learners should be better able to:• Assess the neuropsychological literature on decision making and the medical and legal assessment of capacity in patients with dementia• Identify the limitations of integrating findings from decision-making research into capacity assessments for patients with dementia ABSTRACT: Medical and legal professionals face the challenge of assessing capacity and competency to make medical, legal, and financial decisions in dementia patients with impaired decision making. While such assessments have classically focused on the capacity for complex reasoning and executive functions, research in decision making has revealed that motivational and metacognitive processes are also important. We first briefly review the neuropsychological literature on decision making and on the medical and legal assessment of capacity. Next, we discuss the limitations of integrating findings from decision-making research into capacity assessments, including the group-to-individual inference problem, the unclear role of neuroimaging in capacity assessments, and the lack of capacity measures that integrate important facets of decision making. Finally, we present several case examples where we attempt to demonstrate the potential benefits and important limitations of using decision-making research to aid in capacity determinations.

  11. Accurate modeling and maximum power point detection of ...

    African Journals Online (AJOL)

    Accurate modeling and maximum power point detection of photovoltaic ... Determination of MPP enables the PV system to deliver maximum available power. ..... adaptive artificial neural network: Proposition for a new sizing procedure.

  12. Maximum power per VA control of vector controlled interior ...

    Indian Academy of Sciences (India)

    Thakur Sumeet Singh

    2018-04-11

    Apr 11, 2018 ... Department of Electrical Engineering, Indian Institute of Technology Delhi, New ... The MPVA operation allows maximum-utilization of the drive-system. ... Permanent magnet motor; unity power factor; maximum VA utilization; ...

  13. Electron density distribution in Si and Ge using multipole, maximum ...

    Indian Academy of Sciences (India)

    Si and Ge has been studied using multipole, maximum entropy method (MEM) and ... and electron density distribution using the currently available versatile ..... data should be subjected to maximum possible utility for the characterization of.

  14. Integrated capacity and inventory management with capacity acquisition lead times

    NARCIS (Netherlands)

    Mincsovics, G.Z.; Tan, T.; Alp, O.

    2009-01-01

    We model a make-to-stock production system that utilizes permanent and contingent capacity to meet non-stationary stochastic demand, where a constant lead time is associated with the acquisition of contingent capacity. We determine the structure of the optimal solution concerning both the

  15. Capacity Measurement with the UIC 406 Capacity Method

    DEFF Research Database (Denmark)

    Landex, Alex; Schittenhelm, Bernd; Kaas, Anders H.

    2008-01-01

    This article describes the fast and effective UIC 406 method for calculating capacity consumption on railway lines. It is possible to expound the UIC 406 method in different ways which can lead to different capacity consumptions. Therefore, this article describes how the methodology is expounded...

  16. Multicell Downlink Capacity with Coordinated Processing

    Directory of Open Access Journals (Sweden)

    Roberto Padovani

    2008-07-01

    Full Text Available We study the potential benefits of base-station (BS cooperation for downlink transmission in multicell networks. Based on a modified Wyner-type model with users clustered at the cell-edges, we analyze the dirty-paper-coding (DPC precoder and several linear precoding schemes, including cophasing, zero-forcing (ZF, and MMSE precoders. For the nonfading scenario with random phases, we obtain analytical performance expressions for each scheme. In particular, we characterize the high signal-to-noise ratio (SNR performance gap between the DPC and ZF precoders in large networks, which indicates a singularity problem in certain network settings. Moreover, we demonstrate that the MMSE precoder does not completely resolve the singularity problem. However, by incorporating path gain fading, we numerically show that the singularity problem can be eased by linear precoding techniques aided with multiuser selection. By extending our network model to include cell-interior users, we determine the capacity regions of the two classes of users for various cooperative strategies. In addition to an outer bound and a baseline scheme, we also consider several locally cooperative transmission approaches. The resulting capacity regions show the tradeoff between the performance improvement and the requirement for BS cooperation, signal processing complexity, and channel state information at the transmitter (CSIT.

  17. TreeCluster: Massively scalable transmission clustering using phylogenetic trees

    OpenAIRE

    Moshiri, Alexander

    2018-01-01

    Background: The ability to infer transmission clusters from molecular data is critical to designing and evaluating viral control strategies. Viral sequencing datasets are growing rapidly, but standard methods of transmission cluster inference do not scale well beyond thousands of sequences. Results: I present TreeCluster, a cross-platform tool that performs transmission cluster inference on a given phylogenetic tree orders of magnitude faster than existing inference methods and supports multi...

  18. The rotation of galaxy clusters

    International Nuclear Information System (INIS)

    Tovmassian, H.M.

    2015-01-01

    The method for detection of the galaxy cluster rotation based on the study of distribution of member galaxies with velocities lower and higher of the cluster mean velocity over the cluster image is proposed. The search for rotation is made for flat clusters with a/b> 1.8 and BMI type clusters which are expected to be rotating. For comparison there were studied also round clusters and clusters of NBMI type, the second by brightness galaxy in which does not differ significantly from the cluster cD galaxy. Seventeen out of studied 65 clusters are found to be rotating. It was found that the detection rate is sufficiently high for flat clusters, over 60 per cent, and clusters of BMI type with dominant cD galaxy, ≈ 35 per cent. The obtained results show that clusters were formed from the huge primordial gas clouds and preserved the rotation of the primordial clouds, unless they did not have mergings with other clusters and groups of galaxies, in the result of which the rotation has been prevented

  19. 40 CFR 141.13 - Maximum contaminant levels for turbidity.

    Science.gov (United States)

    2010-07-01

    ... turbidity. 141.13 Section 141.13 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) WATER... Maximum contaminant levels for turbidity. The maximum contaminant levels for turbidity are applicable to... part. The maximum contaminant levels for turbidity in drinking water, measured at a representative...

  20. Maximum Power Training and Plyometrics for Cross-Country Running.

    Science.gov (United States)

    Ebben, William P.

    2001-01-01

    Provides a rationale for maximum power training and plyometrics as conditioning strategies for cross-country runners, examining: an evaluation of training methods (strength training and maximum power training and plyometrics); biomechanic and velocity specificity (role in preventing injury); and practical application of maximum power training and…

  1. 13 CFR 107.840 - Maximum term of Financing.

    Science.gov (United States)

    2010-01-01

    ... 13 Business Credit and Assistance 1 2010-01-01 2010-01-01 false Maximum term of Financing. 107.840... COMPANIES Financing of Small Businesses by Licensees Structuring Licensee's Financing of An Eligible Small Business: Terms and Conditions of Financing § 107.840 Maximum term of Financing. The maximum term of any...

  2. 7 CFR 3565.210 - Maximum interest rate.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 15 2010-01-01 2010-01-01 false Maximum interest rate. 3565.210 Section 3565.210... AGRICULTURE GUARANTEED RURAL RENTAL HOUSING PROGRAM Loan Requirements § 3565.210 Maximum interest rate. The interest rate for a guaranteed loan must not exceed the maximum allowable rate specified by the Agency in...

  3. Characterizing graphs of maximum matching width at most 2

    DEFF Research Database (Denmark)

    Jeong, Jisu; Ok, Seongmin; Suh, Geewon

    2017-01-01

    The maximum matching width is a width-parameter that is de ned on a branch-decomposition over the vertex set of a graph. The size of a maximum matching in the bipartite graph is used as a cut-function. In this paper, we characterize the graphs of maximum matching width at most 2 using the minor o...

  4. Single-cluster dynamics for the random-cluster model

    NARCIS (Netherlands)

    Deng, Y.; Qian, X.; Blöte, H.W.J.

    2009-01-01

    We formulate a single-cluster Monte Carlo algorithm for the simulation of the random-cluster model. This algorithm is a generalization of the Wolff single-cluster method for the q-state Potts model to noninteger values q>1. Its results for static quantities are in a satisfactory agreement with those

  5. Choosing the Number of Clusters in K-Means Clustering

    Science.gov (United States)

    Steinley, Douglas; Brusco, Michael J.

    2011-01-01

    Steinley (2007) provided a lower bound for the sum-of-squares error criterion function used in K-means clustering. In this article, on the basis of the lower bound, the authors propose a method to distinguish between 1 cluster (i.e., a single distribution) versus more than 1 cluster. Additionally, conditional on indicating there are multiple…

  6. Evolution of the spherical clusters

    International Nuclear Information System (INIS)

    Surdin, V.G.

    1978-01-01

    The possible processes of the Galaxy spherical clusters formation and evolution are described on a popular level. The orbits of spherical cluster motion and their spatial velocities are determined. Given are the distrbutions of spherical cluster stars according to their velocities and the observed distribution of spherical clusters in the area of the Galaxy slow evolution. The dissipation and dynamic friction processes destructing clusters with the mass less than 10 4 of solar mass and bringing about the reduction of clusters in the Galaxy are considered. The paradox of forming mainly X-ray sources in spherical clusters is explained. The schematic image of possible ways of forming X-ray sources in spherical clusters is given

  7. Clusters in nuclei. Vol. 1

    International Nuclear Information System (INIS)

    Beck, Christian

    2010-01-01

    Following the pioneering discovery of alpha clustering and of molecular resonances, the field of nuclear clustering is presently one of the domains of heavy-ion nuclear physics facing both the greatest challenges and opportunities. After many summer schools and workshops, in particular over the last decade, the community of nuclear molecular physics decided to team up in producing a comprehensive collection of lectures and tutorial reviews covering the field. This first volume, gathering seven extensive lectures, covers the follow topics: - Cluster Radioactivity - Cluster States and Mean Field Theories - Alpha Clustering and Alpha Condensates - Clustering in Neutron-rich Nuclei - Di-neutron Clustering - Collective Clusterization in Nuclei - Giant Nuclear Molecules By promoting new ideas and developments while retaining a pedagogical nature of presentation throughout, these lectures will both serve as a reference and as advanced teaching material for future courses and schools in the fields of nuclear physics and nuclear astrophysics. (orig.)

  8. Structure and bonding in clusters

    International Nuclear Information System (INIS)

    Kumar, V.

    1991-10-01

    We review here the recent progress made in the understanding of the electronic and atomic structure of small clusters of s-p bonded materials using the density functional molecular dynamics technique within the local density approximation. Starting with a brief description of the method, results are presented for alkali metal clusters, clusters of divalent metals such as Mg and Be which show a transition from van der Waals or weak chemical bonding to metallic behaviour as the cluster size grows and clusters of Al, Sn and Sb. In the case of semiconductors, we discuss results for Si, Ge and GaAs clusters. Clusters of other materials such as P, C, S, and Se are also briefly discussed. From these and other available results we suggest the possibility of unique structures for the magic clusters. (author). 69 refs, 7 figs, 1 tab

  9. Random matrix improved subspace clustering

    KAUST Repository

    Couillet, Romain; Kammoun, Abla

    2017-01-01

    This article introduces a spectral method for statistical subspace clustering. The method is built upon standard kernel spectral clustering techniques, however carefully tuned by theoretical understanding arising from random matrix findings. We show

  10. Eclipsing binaries in open clusters

    DEFF Research Database (Denmark)

    Southworth, John; Clausen, J.V.

    2006-01-01

    Stars: fundamental parameters - Stars : binaries : eclipsing - Stars: Binaries: spectroscopic - Open clusters and ass. : general Udgivelsesdato: 5 August......Stars: fundamental parameters - Stars : binaries : eclipsing - Stars: Binaries: spectroscopic - Open clusters and ass. : general Udgivelsesdato: 5 August...

  11. Installed capacity in New York

    International Nuclear Information System (INIS)

    Charlton, J.

    2006-01-01

    This presentation discussed capacity issues related to the New York Independent System Operator (NYISO). The NYISO's market volume was approximately $11 billion in 2005, and it was responsible for providing 32,075 MW of electricity at peak load to its users. Regulatory uncertainty is currently discouraging investment in new generating resources. All load serving entities are required to contract for sufficient capacity in order to meet their capacity obligations. Market participants currently determine capacity and energy revenues. The NYISO market allows suppliers to recover variable costs for providing ancillary services, and the economic value of the revenue source governs decisions made in the wholesale electricity market. The installed capacity market was designed as a spot auction deficiency auction. Phased-in demand curves are used to modify the installed capacity market's design. A sloped demand curve mechanism is used to value capacity above the minimum requirement for both reliability and competition. Participation in the day-ahead market enhances competition and exerts downward pressure on energy and ancillary service market prices. It was concluded that the market structures and design features of the installed capacity markets recognize the need for system reliability in addition to encouraging robust competition and recognizing energy price caps and regulatory oversights. tabs., figs

  12. Concrete Operations and Attentional Capacity.

    Science.gov (United States)

    Chapman, Michael; Lindenberger, Ulman

    1989-01-01

    To test predictions regarding the attentional capacity requirements of Piaget's stage of concrete operations, a battery of concrete operational tasks and two measures of attentional capacity were administered to 120 first-, second-, and third-graders. Findings concern class inclusion, transitivity of length and weight, and multiplication of…

  13. Fiscal Capacity Equalisation in Tanzania

    NARCIS (Netherlands)

    Allers, Maarten A.; Ishemoi, Lewis J.

    2010-01-01

    Fiscal equalisation aims at enabling decentralised governments to supply similar services at similar tax rates. In order to equalise fiscal disparities, differences in both fiscal capacities and in fiscal needs have to be measured. This paper focuses on the measurement of fiscal capacity in a

  14. Dynamical aspects of galaxy clustering

    International Nuclear Information System (INIS)

    Fall, S.M.

    1980-01-01

    Some recent work on the origin and evolution of galaxy clustering is reviewed, particularly within the context of the gravitational instability theory and the hot big-bang cosmological model. Statistical measures of clustering, including correlation functions and multiplicity functions, are explained and discussed. The close connection between galaxy formation and clustering is emphasized. Additional topics include the dependence of galaxy clustering on the spectrum of primordial density fluctuations and the mean mass density of the Universe. (author)

  15. [Cluster analysis in biomedical researches].

    Science.gov (United States)

    Akopov, A S; Moskovtsev, A A; Dolenko, S A; Savina, G D

    2013-01-01

    Cluster analysis is one of the most popular methods for the analysis of multi-parameter data. The cluster analysis reveals the internal structure of the data, group the separate observations on the degree of their similarity. The review provides a definition of the basic concepts of cluster analysis, and discusses the most popular clustering algorithms: k-means, hierarchical algorithms, Kohonen networks algorithms. Examples are the use of these algorithms in biomedical research.

  16. THE ASSEMBLY OF GALAXY CLUSTERS

    International Nuclear Information System (INIS)

    Berrier, Joel C.; Stewart, Kyle R.; Bullock, James S.; Purcell, Chris W.; Barton, Elizabeth J.; Wechsler, Risa H.

    2009-01-01

    We study the formation of 53 galaxy cluster-size dark matter halos (M = 10 14.0-14.76 M sun ) formed within a pair of cosmological Λ cold dark matter N-body simulations, and track the accretion histories of cluster subhalos with masses large enough to host ∼0.3 L * galaxies. By associating subhalos with cluster galaxies, we find the majority of galaxies in clusters experience no 'preprocessing' in the group environment prior to their accretion into the cluster. On average, 70% of cluster galaxies fall into the cluster potential directly from the field, with no luminous companions in their host halos at the time of accretion; less than 12% are accreted as members of groups with five or more galaxies. Moreover, we find that cluster galaxies are significantly less likely to have experienced a merger in the recent past (∼<6 Gyr) than a field halo of the same mass. These results suggest that local cluster processes such as ram pressure stripping, galaxy harassment, or strangulation play the dominant role in explaining the difference between cluster and field populations at a fixed stellar mass, and that pre-evolution or past merging in the group environment is of secondary importance for setting cluster galaxy properties for most clusters. The accretion times for z = 0 cluster members are quite extended, with ∼20% incorporated into the cluster halo more than 7 Gyr ago and ∼20% within the last 2 Gyr. By comparing the observed morphological fractions in cluster and field populations, we estimate an approximate timescale for late-type to early-type transformation within the cluster environment to be ∼6 Gyr.

  17. Modelling of extreme rainfall events in Peninsular Malaysia based on annual maximum and partial duration series

    Science.gov (United States)

    Zin, Wan Zawiah Wan; Shinyie, Wendy Ling; Jemain, Abdul Aziz

    2015-02-01

    In this study, two series of data for extreme rainfall events are generated based on Annual Maximum and Partial Duration Methods, derived from 102 rain-gauge stations in Peninsular from 1982-2012. To determine the optimal threshold for each station, several requirements must be satisfied and Adapted Hill estimator is employed for this purpose. A semi-parametric bootstrap is then used to estimate the mean square error (MSE) of the estimator at each threshold and the optimal threshold is selected based on the smallest MSE. The mean annual frequency is also checked to ensure that it lies in the range of one to five and the resulting data is also de-clustered to ensure independence. The two data series are then fitted to Generalized Extreme Value and Generalized Pareto distributions for annual maximum and partial duration series, respectively. The parameter estimation methods used are the Maximum Likelihood and the L-moment methods. Two goodness of fit tests are then used to evaluate the best-fitted distribution. The results showed that the Partial Duration series with Generalized Pareto distribution and Maximum Likelihood parameter estimation provides the best representation for extreme rainfall events in Peninsular Malaysia for majority of the stations studied. Based on these findings, several return values are also derived and spatial mapping are constructed to identify the distribution characteristic of extreme rainfall in Peninsular Malaysia.

  18. The Effects of Topology on Throughput Capacity of Large Scale Wireless Networks

    Directory of Open Access Journals (Sweden)

    Qiuming Liu

    2017-03-01

    Full Text Available In this paper, we jointly consider the inhomogeneity and spatial dimension in large scale wireless networks. We study the effects of topology on the throughput capacity. This problem is inherently difficult since it is complex to handle the interference caused by simultaneous transmission. To solve this problem, we, according to the inhomogeneity of topology, divide the transmission into intra-cluster transmission and inter-cluster transmission. For the intra-cluster transmission, a spheroidal percolation model is constructed. The spheroidal percolation model guarantees a constant rate when a power control strategy is adopted. We also propose a cube percolation mode for the inter-cluster transmission. Different from the spheroidal percolation model, a constant transmission rate can be achieved without power control. For both transmissions, we propose a routing scheme with five phases. By comparing the achievable rate of each phase, we get the rate bottleneck, which is the throughput capacity of the network.

  19. Chronorisk in cluster headache

    DEFF Research Database (Denmark)

    Barloese, Mads; Haddock, Bryan; Lund, Nunu T

    2018-01-01

    and a spectral analysis identifying oscillations in risk. Results The Gaussian model fit for the chronorisk distribution for all patients reporting diurnal rhythmicity (n = 286) had a goodness of fit R2 value of 0.97 and identified three times of increased risk peaking at 21:41, 02:02 and 06:23 hours....... In subgroups, three to five modes of increased risk were found and goodness of fit values ranged from 0.85-0.99. Spectral analysis revealed multiple distinct oscillation frequencies in chronorisk in subgroups including a dominant circadian oscillation in episodic patients and an ultradian in chronic....... Conclusions Chronorisk in cluster headache can be characterised as a sum of individual, timed events of increased risk, each having a Gaussian distribution. In episodic cluster headache, attacks follow a circadian rhythmicity whereas, in the chronic variant, ultradian oscillations are dominant reflecting...

  20. CSR in Industrial Clusters

    DEFF Research Database (Denmark)

    Lund-Thomsen, Peter; Pillay, Renginee G.

    2012-01-01

    Purpose – The paper seeks to review the literature on CSR in industrial clusters in developing countries, identifying the main strengths, weaknesses, and gaps in this literature, pointing to future research directions and policy implications in the area of CSR and industrial cluster development....... Design/methodology/approach – A literature review is conducted of both academic and policy-oriented writings that contain the keywords “industrial clusters” and “developing countries” in combination with one or more of the following terms: corporate social responsibility, environmental management, labor...... standards, child labor, climate change, social upgrading, and environmental upgrading. The authors examine the key themes in this literature, identify the main gaps, and point to areas where future work in this area could usefully be undertaken. Feedback has been sought from some of the leading authors...