Automated verification of pointer programs in pointer logic
Zhifang WANG; Yiyun CHEN; Zhenming WANG; Baojian HUA
2008-01-01
Reasoning about pointer programs is difficult and challenging, while their safety is critical in software engineering. Storeless semantics pioneered by Jonkers provides a method to reason about pointer programs. However, the representation of memory states in Jonkers' model is costly and redundant. This paper pre-sents a new framework under a more efficient storeless model for automatically verifying properties of pointer programs related to the correct use of dynamically allo-cated memory such as absence of null dereferences, absence of dangling dereferences, absence of memory leaks, and preservation of structural invariants. The intro-duced logic-Pointer Logic, is developed to achieve such goals. To demonstrate that Pointer Logic is a useful store-less approach to verification, the Schorr-Waite tree-tra-versal algorithm which is always considered as a key test for pointer formalizations was verified via our analysis. Moreover, an experimental tool-plcc was implemented to automatically verify a number of non-trivial pointer programs.
A structured review of reasons for ecstasy use and related behaviours: pointers for future research
Kok Gerjo
2009-07-01
Full Text Available Abstract Background While the health risks of using ecstasy warrant intervention development, a recent meta-analysis of determinants of ecstasy use identified a number of lacunae in the literature. Specifically, no studies were included that address behaviours other than 'using ecstasy' (e.g. 'trying out ecstasy' or 'ceasing ecstasy use'. However, because meta-analyses aim to integrate study results quantitatively, the resulting rigid exclusion criteria cause many studies to be discarded on the basis of their qualitative methodology. Such qualitative studies may nonetheless provide valuable insights to guide future research. To provide an overview of these insights regarding ecstasy use, the current study summarizes and combines what is known from qualitative and exploratory quantitative literature on ecstasy use. Methods The databases PsycINFO and MedLine were searched for publications reporting reasons for ecstasy use and related behaviour, and the results were structured and discussed per behaviour and compared over behaviours. Results Two main categories of reasons were found. The first category comprised reasons to start using ecstasy, use ecstasy, use ecstasy more often, and refrain from ceasing ecstasy use. The second category comprised reasons to refrain from starting to use ecstasy, use less ecstasy, and cease using ecstasy. Reasons for related behaviours within each of these two categories appear to differ, but not as substantially as between the two categories. A large number of reasons that were not yet explored in quantitative research emerged. Conclusion The current summary and combination of exploratory studies yields useful lists of reasons for each behaviour. Before these lists can inform interventions, however, they beg quantitative verification. Also, similarity of determinant configurations of different behaviours can be assessed by addressing determinants of several behaviours in one study. Another important finding is that
Computer Workstation: Pointer/Mouse
... Safety and Health Program Recommendations It's the Law Poster REGULATIONS Law and Regulations Standard Interpretations Training Requirements ... when evaluating your computer workstation. Pointer Placement Pointer Size, Shape, and Settings Pointer/Mouse Quick Tips Keep ...
Pointer Sentinel Mixture Models
Merity, Stephen; Xiong, Caiming; Bradbury, James; Socher, Richard
2016-01-01
Recent neural network sequence models with softmax classifiers have achieved their best language modeling performance only with very large hidden states and large vocabularies. Even then they struggle to predict rare or unseen words even if the context makes the prediction unambiguous. We introduce the pointer sentinel mixture architecture for neural sequence models which has the ability to either reproduce a word from the recent context or produce a word from a standard softmax classifier. O...
Galang, Jemellie; Hagley, Edward W; Clark, Charles W
2010-01-01
An inexpensive green laser pointer was found to emit 20 mW of infrared radiation during normal use. This is potentially a serious hazard that would not be noticed by most users of such pointers. We find that this infrared emission derives from the design of the pointer, and describe a simple method of testing for infrared emissions using common household items.
Optimizing pointer linked data structures
Holm, Carl Wilhelm Mattias
2013-01-01
The thesis explores different ways of optimizing pointer linked data structures, and especially restructuring them. The mechanisms are based on compiler technology, theory, computer languages and hardware architecture that are capable of optimizing the memory layout of complex pointer linked data st
Harford, Steven T.; Gutierrez, Homero; Newman, Michael; Pierce, Robert; Quakenbush, Tim; Wallace, John; Bornstein, Michael
2014-03-01
Ball Aerospace & Technologies Corp. (BATC) has developed a Risley Beam Pointer (RBP) mechanism capable of agile slewing, accurate pointing and high bandwidth. The RBP is comprised of two wedged prisms that offer a wide Field of Regard (FOR) and may be manufactured and operated with diffraction limited optical quality. The tightly packaged mechanism is capable of steering a 4 inch beam over a 60° half angle cone with better than 60 μrad precision. Absolute accuracy of the beam steering is better than 1 mrad. The conformal nature of the RBP makes it an ideal mechanism for use on low altitude aircraft and unmanned aerial vehicles. Unique aspects of the opto-mechanical design include i) thermal compliance to maintain bearing preload and optical figure over a wide temperature range; and ii) packaging of a remote infrared sensor that periodically reports the temperature of both prisms for accurate determination of the index of refraction. The pointing control system operates each prism independently and employs an inner rate loop nested within an outer position loop. Mathematics for the transformation between line-of-sight coordinates and prism rotation are hosted on a 200 MHz microcontroller with just 516 KB of RAM.
The Megalithic Lunar Season Pointer
Clausen, Claus
are concentrated in the east, east- southeast and southeast. This orientation points to specific full moons during the summer period – the lunar season pointer, where the rising points of the full moons, at the horizon concentrate around a few specific directions in the Southeastern quadrant, which specially...
Maximum Correntropy Unscented Kalman Filter for Spacecraft Relative State Estimation.
Liu, Xi; Qu, Hua; Zhao, Jihong; Yue, Pengcheng; Wang, Meng
2016-09-20
A new algorithm called maximum correntropy unscented Kalman filter (MCUKF) is proposed and applied to relative state estimation in space communication networks. As is well known, the unscented Kalman filter (UKF) provides an efficient tool to solve the non-linear state estimate problem. However, the UKF usually plays well in Gaussian noises. Its performance may deteriorate substantially in the presence of non-Gaussian noises, especially when the measurements are disturbed by some heavy-tailed impulsive noises. By making use of the maximum correntropy criterion (MCC), the proposed algorithm can enhance the robustness of UKF against impulsive noises. In the MCUKF, the unscented transformation (UT) is applied to obtain a predicted state estimation and covariance matrix, and a nonlinear regression method with the MCC cost is then used to reformulate the measurement information. Finally, the UT is adopted to the measurement equation to obtain the filter state and covariance matrix. Illustrative examples demonstrate the superior performance of the new algorithm.
Safety and Liveness in Concurrent Pointer Programs
Distefano, Dino; Katoen, Joost-Pieter; Rensink, Arend; Boer, de Frank S.; Bonsangue, Marcello M.; Graf, Susanne; Roever, de Willem-Paul
2006-01-01
The incorrect use of pointers is one of the most common source of software errors. Concurrency has a similar characteristic. Proving the correctness of concurrent pointer manipulating programs, let alone algorithmically, is a highly non-trivial task. This paper proposes an automated verification tec
Maximum Correntropy Unscented Kalman Filter for Spacecraft Relative State Estimation
Xi Liu
2016-09-01
Full Text Available A new algorithm called maximum correntropy unscented Kalman filter (MCUKF is proposed and applied to relative state estimation in space communication networks. As is well known, the unscented Kalman filter (UKF provides an efficient tool to solve the non-linear state estimate problem. However, the UKF usually plays well in Gaussian noises. Its performance may deteriorate substantially in the presence of non-Gaussian noises, especially when the measurements are disturbed by some heavy-tailed impulsive noises. By making use of the maximum correntropy criterion (MCC, the proposed algorithm can enhance the robustness of UKF against impulsive noises. In the MCUKF, the unscented transformation (UT is applied to obtain a predicted state estimation and covariance matrix, and a nonlinear regression method with the MCC cost is then used to reformulate the measurement information. Finally, the UT is adopted to the measurement equation to obtain the filter state and covariance matrix. Illustrative examples demonstrate the superior performance of the new algorithm.
Relate the earthquake parameters to the maximum tsunami runup
Sharghivand, Naeimeh; Kânoǧlu, Utku
2016-04-01
Considering the 1 September 1992 Nicaraguan tsunami manifested itself with an initial shoreline recession, there was paradigm shift from solitary wave to an N-wave (Tadepalli and Synolakis, 1994, Proc. R. Soc. A: Math. Phys. Eng. Sci., 445, 99-112) to define the initial waveform of tsunamis (Kanoglu et al., 2015, Phil. Trans. R. Soc. A, 373: 20140369). The N-wave initial waveform shows specific features, which might enhance maximum runup at a target coastline. Tadepalli & Synolakis (1994) showed that the leading depression N-wave (LEN) run up higher than its mirror image, the leading elevation N-wave (LEN). Later, Kanoglu et al. (2013, Proc. R. Soc. A: Math. Phys. Eng. Sci., 469, 20130015) considered two-dimensional propagation of a finite crest length N-wave over a flat bottom and showed that focusing effect of an N-wave in the direction of leading depression, which enhance the runup. Recently, Kanoglu (2016, EGU Abstract)'s preliminary results suggest that later waves could be higher on the leading depression side for an N-wave, i.e., sequencing defined by Okal and Synolakis (2016, Geophys. J. Int. 204, 719-735) is more pronounced on the leading depression side. Here, we consider submarine earthquakes and estimate the initial ocean surface profiles through Okada's formulation (1985, Bull. Seismol. Soc. Am. 75, 1135-1040). We parameterize earthquake source parameters, such as the length and the width of the fault, the focal depth, the rake (slip) and the dip angles, and the slip amount. Then, we relate ocean surface profiles calculated through Okada (1985) to the generalized N-wave profile defined by Tadepalli and Synolakis (1994) and identify N-wave parameters. Since, for an N-wave type initial condition, Tadepalli and Synolakis (1994) presented maximum runup for a canonical problem -wave propagating over a constant depth segment first and then over a sloping beach- and Kanoglu (2004, J. Fluid Mech., 513, 363-372) for a sloping beach their results allow us to
Relative azimuth inversion by way of damped maximum correlation estimates
Ringler, A.T.; Edwards, J.D.; Hutt, C.R.; Shelly, F.
2012-01-01
Horizontal seismic data are utilized in a large number of Earth studies. Such work depends on the published orientations of the sensitive axes of seismic sensors relative to true North. These orientations can be estimated using a number of different techniques: SensOrLoc (Sensitivity, Orientation and Location), comparison to synthetics (Ekstrom and Busby, 2008), or by way of magnetic compass. Current methods for finding relative station azimuths are unable to do so with arbitrary precision quickly because of limitations in the algorithms (e.g. grid search methods). Furthermore, in order to determine instrument orientations during station visits, it is critical that any analysis software be easily run on a large number of different computer platforms and the results be obtained quickly while on site. We developed a new technique for estimating relative sensor azimuths by inverting for the orientation with the maximum correlation to a reference instrument, using a non-linear parameter estimation routine. By making use of overlapping windows, we are able to make multiple azimuth estimates, which helps to identify the confidence of our azimuth estimate, even when the signal-to-noise ratio (SNR) is low. Finally, our algorithm has been written as a stand-alone, platform independent, Java software package with a graphical user interface for reading and selecting data segments to be analyzed.
A pointer logic and certifying compiler
CHEN Yiyun; GE Lin; HUA Baojian; LI Zhaopeng; LIU Cheng; WANG Zhifang
2007-01-01
Proof-Carrying Code brings two big challenges to the research field of programming languages.One is to seek more expressive logics or type systems to specify or reason about the properties of low-level or high-level programs.The other is to study the technology of certifying compilation in which the compiler generates proofs for programs with annotations.This paper presents our progress in the above two aspects.A pointer logic was designed for PointerC (a C-like programming language) in our research.As an extension of Hoare logic,our pointer logic expresses the change of pointer information for each statement in its inference rules to support program verification.Meanwhile,based on the ideas from CAP (Certified Assembly Programming) and SCAP (Stack-based Certified Assembly Programming),a reasoning framework was built to verify the properties of object code in a Hoare style.And a certifying compiler prototype for PointerC was implemented based on this framework.The main contribution of this paper is the design of the pointer logic and the implementation of the certifying compiler prototype.In our certifying compiler,the source language contains rich pointer types and operations and also supports dynamic storage allocation and deallocation.
Laser Pointer and the Tyndall Effect
Thomas, Eugene
1996-05-01
Laser pointers provide a convenient way to demonstrate the Tyndall effect to beginning students. Since my class is at 8:00 a.m. I like to use coffee and milky water parts as part of the demonstration.
Maximum Relative Entropy Updating and the Value of Learning
Patryk Dziurosz-Serafinowicz
2015-03-01
Full Text Available We examine the possibility of justifying the principle of maximum relative entropy (MRE considered as an updating rule by looking at the value of learning theorem established in classical decision theory. This theorem captures an intuitive requirement for learning: learning should lead to new degrees of belief that are expected to be helpful and never harmful in making decisions. We call this requirement the value of learning. We consider the extent to which learning rules by MRE could satisfy this requirement and so could be a rational means for pursuing practical goals. First, by representing MRE updating as a conditioning model, we show that MRE satisfies the value of learning in cases where learning prompts a complete redistribution of one’s degrees of belief over a partition of propositions. Second, we show that the value of learning may not be generally satisfied by MRE updates in cases of updating on a change in one’s conditional degrees of belief. We explain that this is so because, contrary to what the value of learning requires, one’s prior degrees of belief might not be equal to the expectation of one’s posterior degrees of belief. This, in turn, points towards a more general moral: that the justification of MRE updating in terms of the value of learning may be sensitive to the context of a given learning experience. Moreover, this lends support to the idea that MRE is not a universal nor mechanical updating rule, but rather a rule whose application and justification may be context-sensitive.
THE EFFECT OF THE STATIC RELATIVE STRENGTH ON THE MAXIMUM RELATIVE RECEIVING OF OXYGEN
Abdulla Elezi
2011-09-01
Full Text Available Based on research on the sample of 263 students of age- 18 years, and used batteries of 9 tests for evaluation of the static relative strength and the criterion variable- maximum relative receiving of oxygen (VO2 ml / kg / min based on the Astrand test ,and on regression analysis to determine the influence of the static relative strength on the criterion variable maximum relative oxygen receiving, can be generally concluded that from 9 predictor variables statistically significant partial effect have 2variables. In hierarchical order, they are: the variable of static relative leg strength - endurance of the fingers (the angle of the lower leg and thigh 900 (SRL2 which arithmetic mean is 25.04 seconds and variable ctatic relative strength of arms and shoulders – push-up endurance in the balance beam (angle of the forearm and upper arm 900 ( SRA2 with arithmetic mean of 17.75 seconds. From the statistically influential significant predictor variables on the criterion variable one is from the static relative leg strength (SRL2 and the other is from the static relative strength of arm and shoulder area (SRA2. With the analysis of these relations we can conclude that the isometric contractions of the four headed thigh muscle and the isometric contractions of the three headed upper arm muscle are predominantly responsible for the successful execution of doing actions on a bicycle ergometer and not on the maximum relative receiving of oxygen.
Mastering C pointers tools for programming power
Traister, Robert J
2014-01-01
If you don't fully understand C pointers and how they are used, you're not getting the most out of C programming. This book features complete coverage on using and controlling C language pointers to make C applications more powerful and expressive. This new edition is completely updated and revised to reflect the changes that have been brought about with the full adoption of ANSI C. All discussions and program examples have been updated, and reading materials necessary for any modern ANSI C programmer have also been added.Includes one 3 1/2"" disk containing all of the working programs and m
Familial anthropophobia in pointer dogs?
Dykman, R A; Murphree, O D; Reese, W G
1979-08-01
This article assesses a dog model in terms of a proposed cross-species definition of phobia, the model referring to a strain of unstable dogs that has been produced by selection and inbreeding. The unstable dogs are contrasted with a strain of stable dogs. New findings are presented on approach and activity behavior toward three stimulus objects (man, another dog, and a sheet-covered chair) in a naturalistic setting. The fear response of unstable dogs to objects other than man habituates gradually, whereas the fear response to the sight of man is far more enduring, suggesting a relatively specific fear of man.
Maximum relative height of elastic interfaces in random media.
Rambeau, Joachim; Bustingorry, Sebastian; Kolton, Alejandro B; Schehr, Grégory
2011-10-01
The distribution of the maximal relative height (MRH) of self-affine one-dimensional elastic interfaces in a random potential is studied. We analyze the ground-state configuration at zero driving force, and the critical configuration exactly at the depinning threshold, both for the random-manifold and random-periodic universality classes. These configurations are sampled by exact numerical methods, and their MRH distributions are compared with those with the same roughness exponent and boundary conditions, but produced by independent Fourier modes with normally distributed amplitudes. Using Pickands' theorem we derive an exact analytical description for the right tail of the latter. After properly rescaling the MRH distributions we find that corrections from the Gaussian independent modes approximation are, in general, small, as previously found for the average width distribution of depinning configurations. In the large size limit all corrections are finite except for the ground state in the random-periodic class whose MRH distribution becomes, for periodic boundary conditions, indistinguishable from the Airy distribution. We find that the MRH distributions are, in general, sensitive to changes of boundary conditions.
Laser Pointers Can Cause Serious Eye Damage in Kids
... page: https://medlineplus.gov/news/fullstory_160743.html Laser Pointers Can Cause Serious Eye Damage in Kids ... 1, 2016 THURSDAY, Sept. 1, 2016 (HealthDay News) -- Laser pointers may look harmless enough, but when children ...
Precise Interprocedural Analysis in the Presence of Pointers to the Stack
Sotin, Pascal; Jeannet, Bertrand
In a language with procedures calls and pointers as parameters, an instruction can modify memory locations anywhere in the call-stack. The presence of such side effects breaks most generic interprocedural analysis methods, which assume that only the top of the stack may be modified. We present a method that addresses this issue, based on the definition of an equivalent local semantics in which writing through pointers has a local effect on the stack. Our second contribution in this context is an adequate representation of summary functions that models the effect of a procedure, not only on the values of its scalar and pointer variables, but also on the values contained in pointed memory locations. Our implementation in the interprocedural analyser PInterproc results in a verification tool that infers relational properties on the value of Boolean, numerical and pointer variables.
Sparse Dataflow Analysis with Pointers and Reachability
Madsen, Magnus; Møller, Anders
2014-01-01
for a sparse analysis framework that supports pointers and reachability. We present such a framework, which uses static single assignment form for heap addresses and computes def-use information on-the-fly. We also show that essential information about dominating definitions can be maintained efficiently using...... quadtrees. The framework is presented as a systematic modification of a traditional dataflow analysis algorithm. Our experimental results demonstrate the effectiveness of the technique for a suite of JavaScript programs. By also comparing the performance with an idealized staged approach that computes...
Edi Satriyanto
2013-03-01
Full Text Available The application that builded in this research is a pointer controller using eye movement (eye pointer. This application is one of image processing applications, where the users just have to move their eye to control the computer pointer. This eye pointer is expected able to assist the usage of manual pointer during the presentation. Since the title of this research is using gaze tracking that follow the eye movement, so that is important to detect the center of the pupil. To track the gaze, it is necessary to detect the center of the pupil if the eye image is from the input camera. The gaze tracking is detected using the three-step hierarchy system. First, motion detection, object (eye detection, and then pupil detection. For motion detection, the used method is identify the movement by dynamic compare the pixel ago by current pixel at t time. The eye region is detected using the Haar-Like Feature Classifier, where the sistem must be trained first to get the cascade classifier that allow the sistem to detect the object in each frame that captured by camera. The center of pupil is detect using integral projection.The final step is mapping the position of center of pupil to the screen of monitor using comparison scale between eye resolution with screen resolution. When detecting the eye gaze on the screen, the information (the distance and angle between eyes and a screen is necessary to compute pointing coordinates on the screen. In this research, the accuracy of this application is equal to 80% at eye movement with speed 1-2 second. And the optimum mean value is between 5 and 10. The optimum distance of user and the webcam is 40 cm from webcam.
A New Approach to Pointer Analysis for Assignments
HUANG Bo; ZANG Binyu; LI Jing; ZHU Chuanqi
2001-01-01
Pointer analysis is a technique to identify at compile-time the po tential values of the pointer expressions in a program, which promises significant benefits for optimizing and parallelizing compilers. In this paper, a new approach to pointer analysis for assignments is presented. In this approach, assignments are clas sified into three categories: pointer assignments, structure (union) assignments and normal assignments which don't affect the point-to information. Pointer analyses for these three kinds of assignments respectively make up the integrated algorithm. When analyzing a pointer assignment, a new method called expression expansion is used to calculate both the left targets and the right targets. The integration of recursive data structure analysis into pointer analysis is a significant originality of this paper, which uniforms the pointer analysis for heap variables and the pointer analysis for stack variables. This algorithm is implemented in Agassiz, an analyzing tool for C programs developed by Institute of Parallel Processing, Fudan University. Its accuracy and effectiveness are illustrated by experimental data.
A New Approach to Pointer Analysis for Assignments
黄波; 臧斌宇; 等
2001-01-01
Pointer analysis is a technique to identify at copile-time the potential values of the pointer expressions in a program,which promises significant benefits for optimzing and parallelizing complilers.In this paper,a new approach to pointer analysis for assignments is presented.In this approach,assignments are classified into three categories:pointer assignments,structure(union)assignents and normal assignments which don't affect the point-to information.Pointer analyses for these three kinds of assignments respectively make up the integrated algorithm.When analyzing a pointer assigemtn.a new method called expression expansion is used to calculate both the left targets and the right targets.The integration of recursive data structure analysis into pointer analysis is a significant originality of this paper,which uniforms the pointer analysis for heap variables and the pointer analysis for stack variables.This algorithm is implemented in Agassiz,an analyzing tool for C programs developed by Institute of Parallel Processing,Fudan University,Its accuracy and effectiveness are illustrated by experimental data.
X-Ray Diffraction Simulation Using Laser Pointers and Printers.
Johnson, Neil E.
2001-01-01
Uses a laser pointer to demonstrate the analogy between optical and X-ray diffraction and a laser printer with 600 or 1200 dot resolution to create and modify arrays, print them on transparencies, and illuminate them with laser pointers. Includes 14 references. (Author/YDS)
X-Ray Diffraction Simulation Using Laser Pointers and Printers.
Johnson, Neil E.
2001-01-01
Uses a laser pointer to demonstrate the analogy between optical and X-ray diffraction and a laser printer with 600 or 1200 dot resolution to create and modify arrays, print them on transparencies, and illuminate them with laser pointers. Includes 14 references. (Author/YDS)
Transparent restructuring of pointer-linked data structures
Spek, Harmen Laurens Anne van der
2010-01-01
In contemporary computer systems, data layout has great influence on performance. Traditionally, automatic restructuring in type-unsafe languages has been hard, especially in the presence of pointers. In this thesis, the foundations are laid for successful restructuring of pointer linked data struct
Balloon borne arcsecond pointer feasibility study
Ward, Philip R.; Deweese, Keith D.
2003-08-01
A major hurdle in extending the range of experiments for which balloon vehicles are useful has been the imposition of the gondola dynamics on the accuracy with which an instrument can be kept pointed at a celestial target. In this paper, the foundation for a high fidelity controller simulation is presented and it is shown that sub-arcsecond pointing stability can be achieved for a large instrument pointing at an inertial target. The flexibility of the flight train is represented through generalized modal analysis. A multiple controller scheme is introduced with a coarse azimuth pointer and a pitch-yaw gimbal mount for fine pointing. An analysis and demonstration of the necessity in eliminating static friction are provided along with a solution to eliminate static friction from the system dynamics. A control scheme involving linear accelerometers for enhanced disturbance rejection is also presented. This paper establishes that the proposed control strategy can be made robustly stable with significant design margins. Also demonstrated is the efficacy of the proposed system in rejecting disturbances larger than those considered realistic.
Advantages of nonclassical pointer states in postselected weak measurements
Turek, Yusuf; Shikano, Yutaka; Sun, Chang-Pu; Al-Amri, M
2015-01-01
We investigate, within the weak measurement theory, the advantages of non-classical pointer states over semi-classical ones for coherent, squeezed vacuum, and Schr\\"{o}inger cat states. These states are utilized as pointer state for the system operator $\\hat{A}$ with property $\\hat{A}^{2}=\\hat{I}$, where $\\hat{I}$ represents the identity operator. We calculate the ratio between the signal-to-noise ratio (SNR) of non-postselected and postselected weak measurements. The latter is used to find the quantum Fisher information for the above pointer states. The average shifts for those pointer states with arbitrary interaction strength are investigated in detail. One key result is that we find the postselected weak measurement scheme for non-classical pointer states to be superior to semi-classical ones. This can improve the precision of measurement process.
22 CFR 201.68 - Maximum prices for commodity-related services.
2010-04-01
.... 201.68 Section 201.68 Foreign Relations AGENCY FOR INTERNATIONAL DEVELOPMENT RULES AND PROCEDURES APPLICABLE TO COMMODITY TRANSACTIONS FINANCED BY USAID Price Provisions § 201.68 Maximum prices for commodity... each such service is eligible for USAID-financing under § 201.67 or § 201.68(a) of this part....
Proving Correctness for Pointer Programs in a Verifying Compiler
Kulczycki, Gregory; Singh, Amrinder
2008-01-01
This research describes a component-based approach to proving the correctness of programs involving pointer behavior. The approach supports modular reasoning and is designed to be used within the larger context of a verifying compiler. The approach consists of two parts. When a system component requires the direct manipulation of pointer operations in its implementation, we implement it using a built-in component specifically designed to capture the functional and performance behavior of pointers. When a system component requires pointer behavior via a linked data structure, we ensure that the complexities of the pointer operations are encapsulated within the data structure and are hidden to the client component. In this way, programs that rely on pointers can be verified modularly, without requiring special rules for pointers. The ultimate objective of a verifying compiler is to prove-with as little human intervention as possible-that proposed program code is correct with respect to a full behavioral specification. Full verification for software is especially important for an agency like NASA that is routinely involved in the development of mission critical systems.
Pointer Animation Implementation at Development of Multimedia Learning of Java Programming
Rusli, Muhammad; Atmojo, Yohanes Priyo
2015-01-01
This research represents the development research using the references of previous research results related to the development of interactive multimedia learning (learner controlled), specially about the effectiveness and efficiency of multimedia learning of a content that developed by pointer animation implementation showing the content in…
Preliminary investigation on the relation between maximum wave height and wave spectra
Tao, Aifeng; Wen, Cheng; Wu, Yuqing; Wu, Haoran; Li, Shuo; Cao, Guangsui
2016-04-01
The maximum wave height is important not only for the determination of design wave parameters but also for the marine disaster defense. While it cannot be predicted straightforwardly at present, since the general numerical models for wave forecasting are all based on phase averaged spectra model. Then it becomes very useful to make clear the relationship between the maximum wave height and wave spectra parameters, such as average wave steepness, spectra width and spectra type, such as one single peak spectra or multi peaks spectra. In order to perform this research procedure, plenty of observed wave data are required. We collected ten years wave data measured from a ship in North Sea, one year wave pressure data from nine points around Korea, four years buoy data from three points along Chinese coast. The preliminary investigation results on the relations between maximum waves and spectra via the mention observed data will be present here.
In-medium dispersion relations of charmonia studied by the maximum entropy method
Ikeda, Atsuro; Asakawa, Masayuki; Kitazawa, Masakiyo
2017-01-01
We study in-medium spectral properties of charmonia in the vector and pseudoscalar channels at nonzero momenta on quenched lattices, especially focusing on their dispersion relation and the weight of the peak. We measure the lattice Euclidean correlation functions with nonzero momenta on the anisotropic quenched lattices and study the spectral functions with the maximum entropy method. The dispersion relations of charmonia and the momentum dependence of the weight of the peak are analyzed with the maximum entropy method together with the errors estimated probabilistically in this method. We find a significant increase of the masses of charmonia in medium. We also find that the functional form of the charmonium dispersion relations is not changed from that in the vacuum within the error even at T ≃1.6 Tc for all the channels we analyze.
In-medium dispersion relations of charmonia studied by maximum entropy method
Ikeda, Atsuro; Kitazawa, Masakiyo
2016-01-01
We study in-medium spectral properties of charmonia in the vector and pseudoscalar channels at nonzero momenta on quenched lattices, especially focusing on their dispersion relation and weight of the peak. We measure the lattice Euclidean correlation functions with nonzero momenta on the anisotropic quenched lattices and study the spectral functions with the maximum entropy method. The dispersion relations of charmonia and the momentum dependence of the weight of the peak are analyzed with the maximum entropy method together with the errors estimated probabilistically in this method. We find significant increase of the masses of charmonia in medium. It is also found that the functional form of the charmonium dispersion relations is not changed from that in the vacuum within the error even at $T\\simeq1.6T_c$ for all the channels we analyzed.
Weak measurements with orbital angular momentum pointer states
Puentes, G; Torres, J P
2012-01-01
Weak measurements are a unique tool for accessing information about weakly interacting quantum systems with minimal back-action. Joint weak measurements of single-particle operators with pointer states characterized by a two-dimensional Gaussian distribution can provide, in turn, key information about quantum correlations which can be of relevance for quantum information applications. In this paper, we demonstrate that by employing two-dimensional pointer states endowed with orbital angular momentum (OAM), it is possible to extract second-order weak values of single particle operators, an unaccessible quantity with Gaussian pointer states only. An important application of the results presented here is in the non-destructive measurement of single-particle operator weak variances, via two-dimensional pointer displacements.
Application of the maximum relative entropy method to the physics of ferromagnetic materials
Giffin, Adom; Cafaro, Carlo; Ali, Sean Alan
2016-08-01
It is known that the Maximum relative Entropy (MrE) method can be used to both update and approximate probability distributions functions in statistical inference problems. In this manuscript, we apply the MrE method to infer magnetic properties of ferromagnetic materials. In addition to comparing our approach to more traditional methodologies based upon the Ising model and Mean Field Theory, we also test the effectiveness of the MrE method on conventionally unexplored ferromagnetic materials with defects.
Yasso, B; Li, Y; Alexander, A; Mel'nikova, N B; Mukhina, I V
2014-01-01
A comparison of the relative bioavailability and intensity of penetration of glucosamine sulfate in oral, injection and topical administration of the dosage form Hondroxid Maximum as a cream containing micellar system for transdermal delivery of glucosamine in the experiment by Sprague-Dawley rats was carried out. On the base on the pharmacokinetic profiles data of glucosamine in rat blood plasma with daily administration in 3 times a day for 1 week by cream Hondroxid Maximum 400 mg/kg and the single injection solution of 4% Glucosamine sulfate 400 mg/kg was found that the relative bioavailability was 61.6%. Calculated penetration rate of glucosamine in the plasma through the rats skin in 4 hours, equal to 26.9 μg/cm2 x h, and the penetration of glucosamine through the skin into the plasma after a single dose of cream in 4 hours was 4.12%. Comparative analysis of literature and experimental data and calculations based on them suggest that medicine Hondroxid Maximum, cream with transdermal glucosamine complex in the treatment in accordance with the instructions can provide an average concentration of glucosamine in the synovial fluid of an inflamed joint in the range (0.7 - 1.5) μg/ml, much higher than the concentration of endogenous glucosamine human synovial joint fluid (0.02 - 0.07 μg/ml). By theoretical calculations taking into account experimental data it is shown that the medicine Hondroxid Maximum can reach the bioavailability level of the modern injection forms and exceed the bioavailability level of modern oral forms of glucosamine up to 2 times.
Maximum relative speeds of living organisms: Why do bacteria perform as fast as ostriches?
Meyer-Vernet, Nicole; Rospars, Jean-Pierre
2016-12-01
Self-locomotion is central to animal behaviour and survival. It is generally analysed by focusing on preferred speeds and gaits under particular biological and physical constraints. In the present paper we focus instead on the maximum speed and we study its order-of-magnitude scaling with body size, from bacteria to the largest terrestrial and aquatic organisms. Using data for about 460 species of various taxonomic groups, we find a maximum relative speed of the order of magnitude of ten body lengths per second over a 1020-fold mass range of running and swimming animals. This result implies a locomotor time scale of the order of one tenth of second, virtually independent on body size, anatomy and locomotion style, whose ubiquity requires an explanation building on basic properties of motile organisms. From first-principle estimates, we relate this generic time scale to other basic biological properties, using in particular the recent generalisation of the muscle specific tension to molecular motors. Finally, we go a step further by relating this time scale to still more basic quantities, as environmental conditions at Earth in addition to fundamental physical and chemical constants.
Weak measurements with orbital-angular-momentum pointer states.
Puentes, G; Hermosa, N; Torres, J P
2012-07-27
Weak measurements are a unique tool for accessing information about weakly interacting quantum systems with minimal back action. Joint weak measurements of single-particle operators with pointer states characterized by a two-dimensional Gaussian distribution can provide, in turn, key information about quantum correlations that can be relevant for quantum information applications. Here we demonstrate that by employing two-dimensional pointer states endowed with orbital angular momentum (OAM), it is possible to extract weak values of the higher order moments of single-particle operators, an inaccessible quantity with Gaussian pointer states only. We provide a specific example that illustrates the advantages of our method both in terms of signal enhancement and information retrieval.
Implementation of Binary Search Trees Via Smart Pointers
Ivaylo Donchev
2015-03-01
Full Text Available Study of binary trees has prominent place in the training course of DSA (Data Structures and Algorithms. Their implementation in C++ however is traditionally difficult for students. To a large extent these difficulties are due not so much to the complexity of algorithms as to language complexity in terms of memory management by raw pointers – the programmer must consider too many details to ensure a reliable, efficient and secure implementation. Evolution of C++ regarded to automated resource management, as well as experience in implementation of linear lists by means of C++ 11/14 lead to an attempt to implement binary search trees (BST via smart pointers as well. In the present paper, the authors share experience in this direction. Some conclusions about pedagogical aspects and effectiveness of the new classes, compared to traditional library containers and implementation with built-in pointers, are made.
Implementation of ADS Linked List Via Smart Pointers
Ivaylo Donchev
2015-02-01
Full Text Available Students traditionally have difficulties in implementing abstract data structures (ADS in C++. To a large extent, these difficulties are due to language complexity in terms of memory management with raw pointers – the programmer must take care of too many details to provide reliable, efficient and secure implementation. Since all these technical details distract students from the essence of the studied algorithms, we decided to use in the course in DSA (Data Structures and Algorithms an automated resource management, provided by the C++ standard ISO/IEC 14882:2011. In this work we share experience of using smart pointers to implement linked lists and discuss pedagogical aspects and effectiveness of the new classes, compared to the traditional library containers and implementation via built-in pointers.
Special relativity and theory of gravity via maximum symmetry and localization
2008-01-01
Like Euclid,Riemann and Lobachevski geometries are on an almost equal footing,based on the principle of relativity of maximum symmetry proposed by Professor Lu Qikeng and the postulate on invariant universal constants c and R,the de Sitter/anti-de Sitter（dS/AdS）special relativity on dS/AdS-space with radius R can be set up on an almost equal footing with Einstein’s special relativity on the Minkowski-space in the case of R→∞. Thus the dS-space is coin-like:a law of inertia in Beltrami atlas with Beltrami time simultaneity for the principle of relativity on one side,and the proper-time simultaneity and a Robertson-Walker-like dS-space with entropy and an accelerated expanding S3 fitting the cosmological principle on another side. If our universe is asymptotic to the Robertson-Walker-like dS-space of R（?）（3/Λ）1/2,it should be slightly closed in O（A）with entropy bound S（?）3πc3kB/ΛGh.Contrarily,via its asymptotic behavior, it can fix on Beltrami inertial frames without‘an argument in a circle’and acts as the origin of inertia. There is a triality of conformal extensions of three kinds of special relativity and their null physics on the projective boundary of a 5-d AdS-space,a null cone modulo projective equivalence[N]（?）p（AdS5）. Thus there should be a dS-space on the boundary of S5×AdS5 as a vacuum of supergravity. In the light of Einstein’s‘Galilean regions’,gravity should be based on the localized principle of relativity of full maximum symmetry with a gauge-like dynamics.Thus,this may lead to the theory of gravity of corresponding local symmetry.A simple model of dS-gravity characterized by a dimensionless constant g（?）（AGh/3c3）1/2～10-61shows the features on umbilical manifolds of local dS-invariance. Some gravitational effects out of general relativity may play a role as dark matter. The dark universe and its asymptotic behavior may already indicate that the dS special relativity and dS-gravity be the
Special relativity and theory of gravity via maximum symmetry and localization
GUO HanYing
2008-01-01
Like Euclid,Riemann and Lobachevski geometries are on an almost equal footing,based on the principle of relativity of maximum symmetry proposed by Professor Lu Qikeng and the postulate on invariant universal constants c and R,the de Sitter/anti-de Sitter (dS/AdS) special relativity on dS/AdS-space with radius R can be set up on an almost equal footing with Einstein's special relativity on the Minkowski-space in the case of R →∞.Thus the dS-space is coin-like: a law of inertia in Beltrami atlas with Beltrami time simultaneity for the principle of relativity on one side,and the proper-time simultaneity and a Robertson-Walker-like dS-space with entropy and an accelerated expanding S3 fitting the cosmological principle on another side.If our universe is asymptotic to the Robertson-Walker-like dS-space of R≈(3/∧)1/2,it should be slightly closed in O(A) with entropy bound S≈3πc3kB/∧Gh.Contrarily,via its asymptotic behavior,it can fix on Beltrami inertial frames without 'an argument in a circle' and acts as the origin of inertia.There is a triality of conformal extensions of three kinds of special relativity and their null physics on the projective boundary of a 5-d AdS-space,a null cone modulo projective equivalence [N]≈(e)p(AdS5).Thus there should be a dS-space on the boundary of S5 × AdS5 as a vacuum of supergravity.In the light of Einstein's 'Galilean regions',gravity should be based on the localized principle of relativity of full maximum symmetry with a gauge-like dynamics.Thus,this may lead to the theory of gravity of corresponding local symmetry.A simple model of dS-gravity characterized by a dimensionless constant g≈(∧Gh/3c3)1/2 ～ 10-61 shows the features on umbilical manifolds of local dS-invariance.Some gravitational effects out of general relativity may play a role as dark matter.The dark universe and its asymptotic behavior may already indicate that the dS special relativity and dS-gravity be the foundation of large scale physics.
Mohammad H. Radfar
2006-11-01
Full Text Available We present a new technique for separating two speech signals from a single recording. The proposed method bridges the gap between underdetermined blind source separation techniques and those techniques that model the human auditory system, that is, computational auditory scene analysis (CASA. For this purpose, we decompose the speech signal into the excitation signal and the vocal-tract-related filter and then estimate the components from the mixed speech using a hybrid model. We first express the probability density function (PDF of the mixed speech's log spectral vectors in terms of the PDFs of the underlying speech signal's vocal-tract-related filters. Then, the mean vectors of PDFs of the vocal-tract-related filters are obtained using a maximum likelihood estimator given the mixed signal. Finally, the estimated vocal-tract-related filters along with the extracted fundamental frequencies are used to reconstruct estimates of the individual speech signals. The proposed technique effectively adds vocal-tract-related filter characteristics as a new cue to CASA models using a new grouping technique based on an underdetermined blind source separation. We compare our model with both an underdetermined blind source separation and a CASA method. The experimental results show that our model outperforms both techniques in terms of SNR improvement and the percentage of crosstalk suppression.
Dansereau Richard M
2007-01-01
Full Text Available We present a new technique for separating two speech signals from a single recording. The proposed method bridges the gap between underdetermined blind source separation techniques and those techniques that model the human auditory system, that is, computational auditory scene analysis (CASA. For this purpose, we decompose the speech signal into the excitation signal and the vocal-tract-related filter and then estimate the components from the mixed speech using a hybrid model. We first express the probability density function (PDF of the mixed speech's log spectral vectors in terms of the PDFs of the underlying speech signal's vocal-tract-related filters. Then, the mean vectors of PDFs of the vocal-tract-related filters are obtained using a maximum likelihood estimator given the mixed signal. Finally, the estimated vocal-tract-related filters along with the extracted fundamental frequencies are used to reconstruct estimates of the individual speech signals. The proposed technique effectively adds vocal-tract-related filter characteristics as a new cue to CASA models using a new grouping technique based on an underdetermined blind source separation. We compare our model with both an underdetermined blind source separation and a CASA method. The experimental results show that our model outperforms both techniques in terms of SNR improvement and the percentage of crosstalk suppression.
An aggressive location restoration proposal based on pointer
Yu Yimin; Huang Zailu; Tian Chen; Zhou Zongyi
2006-01-01
A proposal for smooth aggressive location restoration by forwarding pointer is given in this paper. A mobile communication system's robustness in case of a location-database failure is improved without the need for periodic location update operations. Radio resources would be saved at the cost of wire signal and operation of location databases. Meanwhile, a chain of forwarding location pointers has been used during the period of HLR failure. So mobile stations are unconscious of the failure of HLR, and mobile subscribers can always make outgoing call.
SEBA SUSAN; NANDINI AGGARWAL; SHEFALI CHAND; AYUSH GUPTA
2016-12-01
In this paper we investigate information-theoretic image coding techniques that assign longer codes to improbable, imprecise and non-distinct intensities in the image. The variable length coding techniques when applied to cropped facial images of subjects with different facial expressions, highlight the set of low probability intensities that characterize the facial expression such as the creases in the forehead, the widening of the eyes and the opening and closing of the mouth. A new coding scheme based on maximum entropy partitioning is proposed in our work, particularly to identify the improbable intensities related to different emotions. The improbable intensities when used as a mask decode the facial expression correctly, providing an effectiveplatform for future emotion categorization experiments
Maximum jaw opening capacity in adolescents in relation to general joint mobility.
Westling, L; Helkimo, E
1992-09-01
Mandibular jaw opening was related with general joint mobility in a non-patient adolescent group. The angular rotation of the mandible at maximum jaw opening was slightly larger in females than in males and significantly larger in hypermobile individuals. No significant relationship between linear measuring of maximal mandibular opening capacity and peripheral joint mobility was found either at active (AROM) or at passive range of mandibular opening (PROM). PROM was strongly correlated to the mandibular length. Clinical signs in the great jaw closer muscles could not be associated to decreased AROM. The mean value of the difference between PROM-AROM (DPA) was 1.2 mm. Frequent clenching and/or grinding was correlated to increased DPA only in hypermobile adolescents (r = 0.49***). Those with DPA exceeding 5mm had all reciprocal clicking.
Optimal Finger Search Trees in the Pointer Machine
Brodal, Gerth Stølting; Lagogiannis, George; Makris, Christos
2003-01-01
We develop a new finger search tree with worst-case constant update time in the Pointer Machine (PM) model of computation. This was a major problem in the field of Data Structures and was tantalizingly open for over twenty years while many attempts by researchers were made to solve it. The result...
An Approach for Search Based Testing of Null Pointer Exceptions
Romano, D.; Di Penta, M.; Antoniol, G.
2011-01-01
Uncaught exceptions, and in particular null pointer exceptions (NPEs), constitute a major cause of crashes for software systems. Although tools for the static identification of potential NPEs exist, there is need for proper approaches able to identify system execution scenarios causing NPEs. This
[Laser pointers are not toys; eye injury with permanent loss of visual acuity].
Keunen, Jan E E; Delbecq, Ann-Laure M H; Cruysberg, J R M Hans; van Meurs, Jan C; Gan, Ivan M; Berendschot, Tos T J M
2014-01-01
In the nineteen-nineties, there was much hype in the European media about presumed laser pointer maculopathy. However, the recent introduction of more powerful and therefore more dangerous laser pointers and their easy availability on the internet necessitates vigilance on the issue. This is an urgent matter, as here we report three cases of proven maculopathy due to an unsafe laser pointer. Three boys aged 13, 9 and 12 years used an unsafe laser pointer as a toy and looked repeatedly into the pointer, resulting in a permanent reduction in visual acuity due to macular damage. Laser pointers are not designed to be children's toys or instruments to annoy people in a crowd. Health authorities and the ophthalmic community should be aware of the potential danger of improper use of high-output laser pointers and warn the general public before the widespread availability of unsafe laser pointers and consequently laser pointer-induced macular damage becomes a true social problem.
Program Optimization Based Pointer Analysis and Live Stack-Heap Analysis
Mohamed A El-Zawawy
2011-03-01
Full Text Available In this paper, we present type systems for flow-sensitive pointer analysis, live stack-heap (variables analysis, and program optimization. The type system for live stack-heap analysis is an enrichment of that for pointer analysis; the enrichment has the form of a second component being added to types of the latter system. Results of pointer analysis are proved useful via their use in the type system for live stack-heap analysis. The type system for program optimization is also an augmentation of that for live stack-heap analysis, but the augmentation takes the form of a transformation component being added to inference rules of the latter system. The form of program optimization being achieved is that of dead-code elimination. A form of program correction may result indirectly from eliminating faulty code (causing the program to abort that is dead. Therefore program optimization can result in program correction. Our type systems have the advantage of being compositional and relatively-simply structured. The novelty of our work comes from the fact that our type system for program optimization associates the optimized version of a program with a justification (in the form of a type derivation for the optimization. This justification is pretty much appreciated in many research areas like certified code (proof-carrying code which is the motivation of this work.
Program Optimization Based Pointer Analysis and Live Stack-Heap Analysis
El-Zawawy, Mohamed A
2011-01-01
In this paper, we present type systems for flow-sensitive pointer analysis, live stack-heap (variables) analysis, and program optimization. The type system for live stack-heap analysis is an enrichment of that for pointer analysis; the enrichment has the form of a second component being added to types of the latter system. Results of pointer analysis are proved useful via their use in the type system for live stack-heap analysis. The type system for program optimization is also an augmentation of that for live stack-heap analysis, but the augmentation takes the form of a transformation component being added to inference rules of the latter system. The form of program optimization being achieved is that of dead-code elimination. A form of program correction may result indirectly from eliminating faulty code (causing the program to abort) that is dead. Therefore program optimization can result in program correction. Our type systems have the advantage of being compositional and relatively-simply structured. The...
Reppert, Michael; Tokmakoff, Andrei
The structural characterization of intrinsically disordered peptides (IDPs) presents a challenging biophysical problem. Extreme heterogeneity and rapid conformational interconversion make traditional methods difficult to interpret. Due to its ultrafast (ps) shutter speed, Amide I vibrational spectroscopy has received considerable interest as a novel technique to probe IDP structure and dynamics. Historically, Amide I spectroscopy has been limited to delivering global secondary structural information. More recently, however, the method has been adapted to study structure at the local level through incorporation of isotope labels into the protein backbone at specific amide bonds. Thanks to the acute sensitivity of Amide I frequencies to local electrostatic interactions-particularly hydrogen bonds-spectroscopic data on isotope labeled residues directly reports on local peptide conformation. Quantitative information can be extracted using electrostatic frequency maps which translate molecular dynamics trajectories into Amide I spectra for comparison with experiment. Here we present our recent efforts in the development of a rigorous approach to incorporating Amide I spectroscopic restraints into refined molecular dynamics structural ensembles using maximum entropy and related approaches. By combining force field predictions with experimental spectroscopic data, we construct refined structural ensembles for a family of short, strongly disordered, elastin-like peptides in aqueous solution.
Rotating proto-neutron stars: spin evolution, maximum mass and I-Love-Q relations
Martinon, Grégoire; Gualtieri, Leonardo; Ferrari, Valeria
2014-01-01
Shortly after its birth in a gravitational collapse, a proto-neutron star enters in a phase of quasi-stationary evolution characterized by large gradients of the thermodynamical variables and intense neutrino emission. In few tens of seconds the gradients smooth out while the star contracts and cools down, until it becomes a neutron star. In this paper we study this phase of the proto-neutron star life including rotation, and employing finite temperature equations of state. We model the evolution of the rotation rate, and determine the relevant quantities characterizing the star. Our results show that an isolated neutron star cannot reach, at the end of the evolution, the maximum values of mass and rotation rate allowed by the zero-temperature equation of state. Moreover, a mature neutron star evolved in isolation cannot rotate too rapidly, even if it is born from a proto-neutron star rotating at the mass-shedding limit. We also show that the I-Love-Q relations are violated in the first second of life, but th...
Adom Giffin
2014-09-01
Full Text Available In this paper, we continue our efforts to show how maximum relative entropy (MrE can be used as a universal updating algorithm. Here, our purpose is to tackle a joint state and parameter estimation problem where our system is nonlinear and in a non-equilibrium state, i.e., perturbed by varying external forces. Traditional parameter estimation can be performed by using filters, such as the extended Kalman filter (EKF. However, as shown with a toy example of a system with first order non-homogeneous ordinary differential equations, assumptions made by the EKF algorithm (such as the Markov assumption may not be valid. The problem can be solved with exponential smoothing, e.g., exponentially weighted moving average (EWMA. Although this has been shown to produce acceptable filtering results in real exponential systems, it still cannot simultaneously estimate both the state and its parameters and has its own assumptions that are not always valid, for example when jump discontinuities exist. We show that by applying MrE as a filter, we can not only develop the closed form solutions, but we can also infer the parameters of the differential equation simultaneously with the means. This is useful in real, physical systems, where we want to not only filter the noise from our measurements, but we also want to simultaneously infer the parameters of the dynamics of a nonlinear and non-equilibrium system. Although there were many assumptions made throughout the paper to illustrate that EKF and exponential smoothing are special cases ofMrE, we are not “constrained”, by these assumptions. In other words, MrE is completely general and can be used in broader ways.
Elastic pointer directory organization for scalable shared memory multiprocessors
Yuhang Liu; Mingfa Zhu; Limin Xiao
2014-01-01
In the field of supercomputing, one key issue for scal-able shared-memory multiprocessors is the design of the directory which denotes the sharing state for a cache block. A good direc-tory design intends to achieve three key attributes: reasonable memory overhead, sharer position precision and implementation complexity. However, researchers often face the problem that gain-ing one attribute may result in losing another. The paper proposes an elastic pointer directory (EPD) structure based on the analysis of shared-memory applications, taking the fact that the number of sharers for each directory entry is typical y smal . Analysis re-sults show that for 4 096 nodes, the ratio of memory overhead to the ful-map directory is 2.7%. Theoretical analysis and cycle-accurate execution-driven simulations on a 16 and 64-node cache coherence non uniform memory access (CC-NUMA) multiproces-sor show that the corresponding pointer overflow probability is reduced significantly. The performance is observed to be better than that of a limited pointers directory and almost identical to the ful-map directory, except for the slight implementation complex-ity. Using the directory cache to explore directory access locality is also studied. The experimental result shows that this is a promis-ing approach to be used in the state-of-the-art high performance computing domain.
Nutrient maximums related to low oxygen concentrations in the southern Canada Basin
JIN Ming-ming; SHI Jiuxin; LU Yong; CHEN Jianfang; GAO Guoping; WU Jingfeng; ZHANG Haisheng
2005-01-01
The phenomenon of nutrient maximums at 70～200 m occurred only in the region of the Canada Basin among the world oceans. The prevailing hypothesis was that the direct injection of the low-temperature high-nutrient brines from the Chukchi Sea shelf (＜50 m) in winter provided the nutrient maximums. However, we found that there are five problems in the direct injection process. Formerly Jin et al. considered that the formation of nutrient maximums can be a process of locally long-term regeneration. Here we propose a regeneration-mixture process. Data of temperature, salinity, oxygen and nutrients were collected at three stations in the southern Canada Basin during the summer 1999 cruise. We identified the cores of the surface, near-surface, potential temperature maximum waters and Arctic Bottom Water by the diagrams and vertical profiles of salinity, potential temperature, oxygen and nutrients. The historical 129Ⅰ data indicated that the surface and near-surface waters were Pacific-origin, but the waters below the potential temperature maximum core depth was Atlantic-origin. Along with the correlation of nutrient maximums and very low oxygen contents in the near-surface water, we hypothesize that, the putative organic matter was decomposed to inorganic nutrients; and the Pacific water was mixed with the Atlantic water in the transition zone. The idea of the regeneration-mixture process agrees with the historical observations of no apparent seasonal changes, the smooth nutrient profiles, the lowest saturation of CaCO3 above 400 m, low rate of CFC-11 ventilation and 3H-3He ages of 8～18 a around the nutrient maximum depths.
You, Daekeun; Simpson, Matthew; Antani, Sameer; Demner-Fushman, Dina; Thoma, George R.
2013-01-01
Pointers (arrows and symbols) are frequently used in biomedical images to highlight specific image regions of interest (ROIs) that are mentioned in figure captions and/or text discussion. Detection of pointers is the first step toward extracting relevant visual features from ROIs and combining them with textual descriptions for a multimodal (text and image) biomedical article retrieval system. Recently we developed a pointer recognition algorithm based on an edge-based pointer segmentation method, and subsequently reported improvements made on our initial approach involving the use of Active Shape Models (ASM) for pointer recognition and region growing-based method for pointer segmentation. These methods contributed to improving the recall of pointer recognition but not much to the precision. The method discussed in this article is our recent effort to improve the precision rate. Evaluation performed on two datasets and compared with other pointer segmentation methods show significantly improved precision and the highest F1 score.
Marcelo Matida Hamata
2009-02-01
Full Text Available Fabrication of occlusal splints in centric relation for temporomandibular disorders (TMD patients is arguable, since this position has been defined for asymptomatic stomatognathic system. Thus, maximum intercuspation might be employed in patients with occlusal stability, eliminating the need for interocclusal records. This study compared occlusal splints fabricated in centric relation and maximum intercuspation in muscle pain reduction of TMD patients. Twenty patients with TMD of myogenous origin and bruxism were divided into 2 groups treated with splints in maximum intercuspation (I or centric relation (II. Clinical, electrognathographic and electromyographic examinations were performed before and 3 months after therapy. Data were analyzed by the Student's t test. Differences at 5% level of probability were considered statistically significant. There was a remarkable reduction in pain symptomatology, without statistically significant differences (p>0.05 between the groups. There was mandibular repositioning during therapy, as demonstrated by the change in occlusal contacts on the splints. Electrognathographic examination demonstrated a significant increase in maximum left lateral movement for group I and right lateral movement for group II (p0.05 in the electromyographic activities at rest after utilization of both splints. In conclusion, both occlusal splints were effective for pain control and presented similar action. The results suggest that maximum intercuspation may be used for fabrication of occlusal splints in patients with occlusal stability without large discrepancies between centric relation and maximum intercuspation. Moreover, this technique is simpler and less expensive.
Structural features of a potential gas hydrate area in the Pointer Ridge off southwest Taiwan
Wang, Hsueh-Fen; Hsu, Shu-Kun; Tsai, Ching-Hui; Chen, Song-Chuen; Liu, Char-Shine; Lin, Hsiao-Shan
2015-04-01
The offshore area of the southwest Taiwan is located in the oblique convergence zone between the northern continental margin of South China Sea and the Manila accretionary wedge. To the west of the deformation front offshore southwestern Taiwan, the Pointer Ridge is located in the passive South China Sea continental margin. The continental margin is compose of extensional horst-and-graben structures. There are numerous submarine channels and linear ridge, formed due to the submarine erosion across the continental slope region. According to geophysical research off SW Taiwan, abundant gas hydrate may exist. In this study, our purpose is to understand the relationship between the near-seafloor structures of the Pointer Ridge and the gas hydrate formation off SW Taiwan. The data we used include multi-beam echo sounder (MBES), side-scan sonar (SSS), sub-bottom profiler (SBP) and the multi-channel reflection seismic (MCS) data. Our results show the pockmark and gas seepage structures mainly appear in the place where the gradient of the BSR thickness is maximum. Those sites contain authigenic carbonate signature shown in the sub-bottom profiler. We also observe several folds and faults structures in this extensional background; however, these compressional features need further studies.
25(OHD3 Levels Relative to Muscle Strength and Maximum Oxygen Uptake in Athletes
Książek Anna
2016-04-01
Full Text Available Vitamin D is mainly known for its effects on the bone and calcium metabolism. The discovery of Vitamin D receptors in many extraskeletal cells suggests that it may also play a significant role in other organs and systems. The aim of our study was to assess the relationship between 25(OHD3 levels, lower limb isokinetic strength and maximum oxygen uptake in well-trained professional football players. We enrolled 43 Polish premier league soccer players. The mean age was 22.7±5.3 years. Our study showed decreased serum 25(OHD3 levels in 74.4% of the professional players. The results also demonstrated a lack of statistically significant correlation between 25(OHD3 levels and lower limb muscle strength with the exception of peak torque of the left knee extensors at an angular velocity of 150°/s (r=0.41. No significant correlations were found between hand grip strength and maximum oxygen uptake. Based on our study we concluded that in well-trained professional soccer players, there was no correlation between serum levels of 25(OHD3 and muscle strength or maximum oxygen uptake.
Li, Dingcheng
2011-01-01
Coreference resolution (CR) and entity relation detection (ERD) aim at finding predefined relations between pairs of entities in text. CR focuses on resolving identity relations while ERD focuses on detecting non-identity relations. Both CR and ERD are important as they can potentially improve other natural language processing (NLP) related tasks…
Inner change in the Corpus Paulinum: pointers for pastoral counselling
Y. Campbell-Lane
2007-07-01
Full Text Available The aim of this article is to establish what perspectives exist on inner change within the “Corpus Paulinum” and how it should be applied in pastoral counselling. The Scriptural guidelines of change that will be examined for the purposes of this article, are found in the following references: Ephesians 4:22-24, Colos- sians 3:8-10, and Romans 12:1-2. The work of the Holy Spirit as “Agent of change” will also be discussed and finally some pointers on inner change and the implications for pastoral counselling will be proposed.
Terhune, Claire E; Hylander, William L; Vinyard, Christopher J; Taylor, Andrea B
2015-05-01
Maximum jaw gape is a performance variable related to feeding and non-feeding oral behaviors, such as canine gape displays, and is influenced by several factors including jaw-muscle fiber architecture, muscle position on the skull, and jaw morphology. Maximum gape, jaw length, and canine height are strongly correlated across catarrhine primates, but relationships between gape and other aspects of masticatory apparatus morphology are less clear. We examine the effects of jaw-adductor fiber architecture, jaw-muscle leverage, and jaw form on gape in an intraspecific sample of sexually dimorphic crab-eating macaques (Macaca fascicularis). As M. fascicularis males have relatively larger maximum gapes than females, we predict that males will have muscle and jaw morphologies that facilitate large gape, but these morphologies may come at some expense to bite force. Male crab-eating macaques have relatively longer jaw-muscle fibers, masseters with decreased leverage, and temporomandibular joint morphologies that facilitate the production of wide gapes. Because relative canine height is correlated with maximum gape in catarrhines, and males have relatively longer canines than females, these results support the hypothesis that male M. fascicularis have experienced selection to increase maximum gape. The sexes do not differ in relative masseter physiologic cross-sectional area (PCSA), but males compensate for a potential trade-off between muscle excursion versus muscle force with increased temporalis weight and PCSA. This musculoskeletal configuration is likely functionally significant for behaviors involving aggressive canine biting and displays in male M. fascicularis and provides additional evidence supporting the multifactorial nature of the catarrhine masticatory apparatus. Our results have implications for the evolution of craniofacial morphology in catarrhine primates and reinforce the importance of evaluating additional factors other than feeding behavior and diet
2010-04-01
... 20 Employees' Benefits 2 2010-04-01 2010-04-01 false Computing the Special Minimum Primary Insurance Amount and Related Maximum Family Benefits V Appendix V to Subpart C of Part 404 Employees...- ) Computing Primary Insurance Amounts Pt. 404, Subpt. C, App. V Appendix V to Subpart C of Part 404—Computing...
Ferreira,Amanda de Freitas; Henriques,João César Guimarães; Almeida,Guilherme de Araújo; Machado,Asbel Rodrigues; Machado, Naila Aparecida de Godoi; Fernandes Neto,Alfredo Júlio
2009-01-01
This research consisted of a quantitative assessment, and aimed to measure the possible discrepancies between the maxillomandibular positions for centric relation (CR) and maximum intercuspation (MI), using computed tomography volumetric cone beam (cone beam method). The sample of the study consisted of 10 asymptomatic young adult patients divided into two types of standard occlusion: normal occlusion and Angle Class I occlusion. In order to obtain the centric relation, a JIG device and mandi...
Controlling mouse pointer position using an infrared head-operated joystick.
Evans, D G; Drew, R; Blenkhorn, P
2000-03-01
This paper describes the motivation for and the design considerations of a low-cost head-operated joystick. The paper briefly summarizes the requirements of head-operated mouse pointer control for people with disabilities before discussing a set of technological approaches that can be used to satisfy these requirements. The paper focuses on the design of a head-operated joystick that uses infrared light emitting diodes (LED's) and photodetectors to determine head position, which is subsequently converted into signals that emulate a Microsoft mouse. There are two significant findings. The first is that, while nonideal device characteristics might appear to make the joystick difficult to use, users naturally compensate for nonlinearities, in a transparent manner, because of visual feedback of mouse pointer position. The second finding, from relatively informal, independent trials, indicates that disabled users prefer a head-operated device that has the characteristics of a joystick (a relative pointing device) to those of a mouse (an absolute pointing device).
Sheng, Shiqi; Tu, Z C
2015-02-01
We present a unified perspective on nonequilibrium heat engines by generalizing nonlinear irreversible thermodynamics. For tight-coupling heat engines, a generic constitutive relation for nonlinear response accurate up to the quadratic order is derived from the stalling condition and the symmetry argument. By applying this generic nonlinear constitutive relation to finite-time thermodynamics, we obtain the necessary and sufficient condition for the universality of efficiency at maximum power, which states that a tight-coupling heat engine takes the universal efficiency at maximum power up to the quadratic order if and only if either the engine symmetrically interacts with two heat reservoirs or the elementary thermal energy flowing through the engine matches the characteristic energy of the engine. Hence we solve the following paradox: On the one hand, the quadratic term in the universal efficiency at maximum power for tight-coupling heat engines turned out to be a consequence of symmetry [Esposito, Lindenberg, and Van den Broeck, Phys. Rev. Lett. 102, 130602 (2009); Sheng and Tu, Phys. Rev. E 89, 012129 (2014)]; On the other hand, typical heat engines such as the Curzon-Ahlborn endoreversible heat engine [Curzon and Ahlborn, Am. J. Phys. 43, 22 (1975)] and the Feynman ratchet [Tu, J. Phys. A 41, 312003 (2008)] recover the universal efficiency at maximum power regardless of any symmetry.
Direct state reconstruction with coupling-deformed pointer observables
Zhu, Xuanmin; Zhang, Yu-Xiang; Wu, Shengjun
2016-06-01
Direct state tomography (DST) using weak measurements has received wide attention. Based on the concept of coupling-deformed pointer observables presented by Zhang et al. [Y.-X. Zhang, S. Wu, and Z.-B. Chen, Phys. Rev. A 93, 032128 (2016), 10.1103/PhysRevA.93.032128], a modified direct state tomography (MDST) is proposed, examined, and compared with other typical state tomography schemes. MDST has exact validity for measurements of any strength. We identify the strength needed to attain the highest efficiency level of MDST by using statistical theory. MDST is much more efficient than DST in the sense that far fewer samples are needed to reach DST's level of reconstruction accuracy. Moreover, MDST has no inherent bias when compared to DST.
Open quantum dots in graphene: Scaling relativistic pointer states
Ferry, D. K.; Huang, L.; Yang, R.; Lai, Y.-C.; Akis, R.
2010-04-01
Open quantum dots provide a window into the connection between quantum and classical physics, particularly through the decoherence theory, in which an important set of quantum states are not "washed out" through interaction with the environment-the pointer states provide connection to trapped classical orbits which remain stable in the dots. Graphene is a recently discovered material with highly unusual properties. This single layer, one atom thick, sheet of carbon has a unique bandstructure, governed by the Dirac equation, in which charge carriers imitate relativistic particles with zero rest mass. Here, an atomic orbital-based recursive Green's function method is used for studying the quantum transport. We study quantum fluctuations in graphene and bilayer graphene quantum dots with this recursive Green's function method. Finally, we examine the scaling of the domiant fluctuation frequency with dot size.
[Chronic thallium intoxication in five German pointers of one litter].
Gassner, G; Tholen, V; Ternes, W
2000-01-01
The difficulty of diagnosis and therapy of chronic thallium intoxication is described in five German Pointers with the same skin disease. The detection of thallium in cases of skin lesions like the cutaneous erythema with oedema and crusts or in chronic cases with multifocal alopecia is difficult. The first diagnostic information was gathered in this case from the high thallium level in the urine. The thallium concentration in the hair is subject to great variations, even in physiologic conditions. The trichogramme showed in this case pathognomonic changes like adhesion of the hair follicles. Differential diagnosis for this symmetric alopezia without pruritus are hormonal disturbances or, in puppies, the generalized form of demodicosis. The five affected dogs were treated with Fe III-Hexacyanoferrat. The clinical appearance of the skin improved slowly during a period of 1-2 months.
Usage of the Pointer Variable in C Language%C语言中指针变量的用法
周二强; 张妍琰
2014-01-01
C语言中使用指针变量时通常需两步：第一步为指针变量赋值，把某个存储单元的地址存入与指针变量相关的存储单元中，即让指针变量指向某个存储单元；第二步以间接引用的方式使用指针变量指向的存储单元。利用变量直接使用存储单元时，存储单元的使用范围受限于变量的作用域，而通过指针变量以间接引用的方式可以扩展存储单元的使用范围。以间接引用方式使用的存储单元不仅可以是基本数据类型的、数组等，而且可以是没有变量标识的堆空间上的存储单元，甚至还可以是与代码相关的“函数类型”的存储单元。%Uses a pointer variable in C programming language need to take two steps:the first step is to assign the pointer variable a value, stores the address of the storage unit into place that is associated with the pointer variable, makes the pointer variable pointing to a storage unit; the second step is to use the storage unit which is pointed by the pointer variable in an indirect way. When uses a storage unit directly by the variable, the range of the storage unit is limited by the scope of the variable. By contrast,it can extend the range of the storage unit by using a pointer variable in the indirect way. The storage unit used in an indirect way can not only be the basic types or arrays, but also can be the storage unit that is not signed as a variable on the heap space, even can be the function type storage unit that is related to the code.
Izawa, Kazuhiro P.; Watanabe, Satoshi; Hirano, Yasuyuki; Matsushima, Shinya; Suzuki, Tomohiro; Oka, Koichiro; Kida, Keisuke; Suzuki, Kengo; Osada, Naohiko; Omiya, Kazuto; Brubaker, Peter H.; Shimizu, Hiroyuki; Akashi, Yoshihiro J.
2015-01-01
Abstract Maximum gait speed and physical activity (PA) relate to mortality and morbidity, but little is known about gender-related differences in these factors in elderly hospitalized cardiac inpatients. This study aimed to determine differences in maximum gait speed and daily measured PA based on sex and the relationship between these measures in elderly cardiac inpatients. A consecutive 268 elderly Japanese cardiac inpatients (mean age, 73.3 years) were enrolled and divided by sex into female (n = 75, 28%) and male (n = 193, 72%) groups. Patient characteristics and maximum gait speed, average step count, and PA energy expenditure (PAEE) in kilocalorie per day for 2 days assessed by accelerometer were compared between groups. Gait speed correlated positively with in-hospital PA measured by average daily step count (r = 0.46, P < 0.001) and average daily PAEE (r = 0.47, P < 0.001) in all patients. After adjustment for left ventricular ejection fraction, step counts and PAEE were significantly lower in females than males (2651.35 ± 1889.92 vs 4037.33 ± 1866.81 steps, P < 0.001; 52.74 ± 51.98 vs 99.33 ± 51.40 kcal, P < 0.001), respectively. Maximum gait speed was slower and PA lower in elderly female versus male inpatients. Minimum gait speed and step count values in this study might be minimum target values for elderly male and female Japanese cardiac inpatients. PMID:25789953
Ferreira, Amanda de Freitas; Henriques, João César Guimarães; Almeida, Guilherme Araújo; Machado, Asbel Rodrigues; Machado, Naila Aparecida de Godoi; Fernandes Neto, Alfredo Júlio
2009-01-01
This research consisted of a quantitative assessment, and aimed to measure the possible discrepancies between the maxillomandibular positions for centric relation (CR) and maximum intercuspation (MI), using computed tomography volumetric cone beam (cone beam method). The sample of the study consisted of 10 asymptomatic young adult patients divided into two types of standard occlusion: normal occlusion and Angle Class I occlusion. In order to obtain the centric relation, a JIG device and mandible manipulation were used to deprogram the habitual conditions of the jaw. The evaluations were conducted in both frontal and lateral tomographic images, showing the condyle/articular fossa relation. The images were processed in the software included in the NewTom 3G device (QR NNT software version 2.00), and 8 tomographic images were obtained per patient, four laterally and four frontally exhibiting the TMA's (in CR and MI, on both sides, right and left). By means of tools included in another software, linear and angular measurements were performed and statistically analyzed by student t test. According to the methodology and the analysis performed in asymptomatic patients, it was not possible to detect statistically significant differences between the positions of centric relation and maximum intercuspation. However, the resources of cone beam tomography are of extreme relevance to the completion of further studies that use heterogeneous groups of samples in order to compare the results.
Amanda de Freitas Ferreira
2009-01-01
Full Text Available This research consisted of a quantitative assessment, and aimed to measure the possible discrepancies between the maxillomandibular positions for centric relation (CR and maximum intercuspation (MI, using computed tomography volumetric cone beam (cone beam method. The sample of the study consisted of 10 asymptomatic young adult patients divided into two types of standard occlusion: normal occlusion and Angle Class I occlusion. In order to obtain the centric relation, a JIG device and mandible manipulation were used to deprogram the habitual conditions of the jaw. The evaluations were conducted in both frontal and lateral tomographic images, showing the condyle/articular fossa relation. The images were processed in the software included in the NewTom 3G device (QR NNT software version 2.00, and 8 tomographic images were obtained per patient, four laterally and four frontally exhibiting the TMA's (in CR and MI, on both sides, right and left. By means of tools included in another software, linear and angular measurements were performed and statistically analyzed by student t test. According to the methodology and the analysis performed in asymptomatic patients, it was not possible to detect statistically significant differences between the positions of centric relation and maximum intercuspation. However, the resources of cone beam tomography are of extreme relevance to the completion of further studies that use heterogeneous groups of samples in order to compare the results.
Parks, A D
2010-01-01
Exact pointer states are obtained for projection operator measurements performed upon pre-selected (PS) and upon pre- and post-selected (PPS) quantum systems. These states are used to provide simple exact expressions for both the pointer spatial probability distribution profiles and the mean values of arbitrary pointer observables associated with PS and PPS projection operator measurements that are valid for any strength of the interaction which couples a measurement pointer to the quantum system. These profiles and mean values are compared in order to identify the effects of post-selection upon projector measurement pointers. As a special case, these mean value results are applied to the weak measurement regime - yielding PS and PPS mean value expressions which are valid for any operator (projector or non-projector). Measurement sensitivities which are useful for estimating weak measurement accuracies for PS and PPS systems are also obtained and discussed.
Larecki, Wieslaw; Banach, Zbigniew
2014-01-01
This paper analyzes the propagation of the waves of weak discontinuity in a phonon gas described by the four-moment maximum entropy phonon hydrodynamics involving a nonlinear isotropic phonon dispersion relation. For the considered hyperbolic equations of phonon gas hydrodynamics, the eigenvalue problem is analyzed and the condition of genuine nonlinearity is discussed. The speed of the wave front propagating into the region in thermal equilibrium is first determined in terms of the integral formula dependent on the phonon dispersion relation and subsequently explicitly calculated for the Dubey dispersion-relation model: |k|=ωc-1(1+bω2). The specification of the parameters c and b for sodium fluoride (NaF) and semimetallic bismuth (Bi) then makes it possible to compare the calculated dependence of the wave-front speed on the sample’s temperature with the empirical relations of Coleman and Newman (1988) describing for NaF and Bi the variation of the second-sound speed with temperature. It is demonstrated that the calculated temperature dependence of the wave-front speed resembles the empirical relation and that the parameters c and b obtained from fitting respectively the empirical relation and the original material parameters of Dubey (1973) are of the same order of magnitude, the difference being in the values of the numerical factors. It is also shown that the calculated temperature dependence is in good agreement with the predictions of Hardy and Jaswal’s theory (Hardy and Jaswal, 1971) on second-sound propagation. This suggests that the nonlinearity of a phonon dispersion relation should be taken into account in the theories aiming at the description of the wave-type phonon heat transport and that the Dubey nonlinear isotropic dispersion-relation model can be very useful for this purpose.
Kimmel, David G.; McGlaughon, Benjamin D.; Leonard, Jeremy; Paerl, Hans W.; Taylor, J. Christopher; Cira, Emily K.; Wetz, Michael S.
2015-05-01
Estuaries often have distinct zones of high chlorophyll a concentrations, known as chlorophyll maximum (CMAX). The persistence of these features is often attributed to physical (mixing and light availability) and chemical (nutrient availability) features, but the role of mesozooplankton grazing is rarely explored. We measured the spatial and temporal variability of the CMAX and mesozooplankton community in the eutrophic Neuse River Estuary, North Carolina. We also conducted grazing experiments to determine the relative impact of mesozooplankton grazing on the CMAX during the phytoplankton growing season (spring through late summer). The CMAX was consistently located upriver of the zone of maximum zooplankton abundance, with an average spatial separation of 18 km. Grazing experiments in the CMAX region revealed negligible effect of mesozooplankton on chlorophyll a during March, and no effect during June or August. These results suggest that the spatial separation of the peak in chlorophyll a concentration and mesozooplankton abundance results in minimal impact of mesozooplankton grazing, contributing to persistence of the CMAX for prolonged time periods. In the Neuse River Estuary, the low mesozooplankton abundance in the CMAX region is attributed to lack of a low salinity tolerant species, predation by the ctenophore Mnemiopsis leidyi, and/or physiologic impacts on mesozooplankton growth rates due to temperature (in the case of low wintertime abundances). The consequences of this lack of overlap result in exacerbation of the effects of eutrophication; namely a lack of trophic transfer to mesozooplankton in this region and the sinking of phytodetritus to the benthos that fuels hypoxia.
Grossi Márcio L
2007-04-01
Full Text Available Abstract Background Vertical facial pattern may be related to the direction of pull of the masticatory muscles, yet its effect on occlusal force and elastic deformation of the mandible still is unclear. This study tested whether the variation in vertical facial pattern is related to the variation in maximum occlusal force (MOF and medial mandibular flexure (MMF in 51 fully-dentate adults. Methods Data from cephalometric analysis according to the method of Ricketts were used to divide the subjects into three groups: Dolichofacial (n = 6, Mesofacial (n = 10 and Brachyfacial (n = 35. Bilateral MOF was measured using a cross-arch force transducer placed in the first molar region. For MMF, impressions of the mandibular occlusal surface were made in rest (R and in maximum opening (O positions. The impressions were scanned, and reference points were selected on the occlusal surface of the contralateral first molars. MMF was calculated by subtracting the intermolar distance in O from the intermolar distance in R. Data were analysed by ANCOVA (fixed factors: facial pattern, sex; covariate: body mass index (BMI; alpha = 0.05. Results No significant difference of MOF or MMF was found among the three facial patterns (P = 0.62 and P = 0.72, respectively. BMI was not a significant covariate for MOF or MMF (P > 0.05. Sex was a significant factor only for MOF (P = 0.007; males had higher MOF values than females. Conclusion These results suggest that MOF and MMF did not vary as a function of vertical facial pattern in this Brazilian sample.
Brunner, R.; Akis, R.; Ferry, D. K.; Kuchar, F.; Meisels, R.
2008-07-01
We discuss a quantum system coupled to the environment, composed of an open array of billiards (dots) in series. Beside pointer states occurring in individual dots, we observe sets of robust states which arise only in the array. We define these new states as bipartite pointer states, since they cannot be described in terms of simple linear combinations of robust single-dot states. The classical existence of bipartite pointer states is confirmed by comparing the quantum-mechanical and classical results. The ability of the robust states to create “offspring” indicates that quantum Darwinism is in action.
Ureña-López, L. Arturo; Robles, Victor H.; Matos, T.
2017-08-01
Recent analysis of the rotation curves of a large sample of galaxies with very diverse stellar properties reveals a relation between the radial acceleration purely due to the baryonic matter and the one inferred directly from the observed rotation curves. Assuming the dark matter (DM) exists, this acceleration relation is tantamount to an acceleration relation between DM and baryons. This leads us to a universal maximum acceleration for all halos. Using the latter in DM profiles that predict inner cores implies that the central surface density μDM=ρsrs must be a universal constant, as suggested by previous studies of selected galaxies, revealing a strong correlation between the density ρs and scale rs parameters in each profile. We then explore the consequences of the constancy of μDM in the context of the ultralight scalar field dark matter model (SFDM). We find that for this model μDM=648 M⊙ pc-2 and that the so-called WaveDM soliton profile should be a universal feature of the DM halos. Comparing with the data from the Milky Way and Andromeda satellites, we find that they are all consistent with a boson mass of the scalar field particle of the order of 10-21 eV /c2, which puts the SFDM model in agreement with recent cosmological constraints.
Eggers, G. L.; Lewis, K. W.; Simons, F. J.; Olhede, S.
2013-12-01
Venus does not possess a plate-tectonic system like that observed on Earth, and many surface features--such as tesserae and coronae--lack terrestrial equivalents. To understand Venus' tectonics is to understand its lithosphere, requiring a study of topography and gravity, and how they relate. Past studies of topography dealt with mapping and classification of visually observed features, and studies of gravity dealt with inverting the relation between topography and gravity anomalies to recover surface density and elastic thickness in either the space (correlation) or the spectral (admittance, coherence) domain. In the former case, geological features could be delineated but not classified quantitatively. In the latter case, rectangular or circular data windows were used, lacking geological definition. While the estimates of lithospheric strength on this basis were quantitative, they lacked robust error estimates. Here, we remapped the surface into 77 regions visually and qualitatively defined from a combination of Magellan topography, gravity, and radar images. We parameterize the spectral covariance of the observed topography, treating it as a Gaussian process assumed to be stationary over the mapped regions, using a three-parameter isotropic Matern model, and perform maximum-likelihood based inversions for the parameters. We discuss the parameter distribution across the Venusian surface and across terrain types such as coronoae, dorsae, tesserae, and their relation with mean elevation and latitudinal position. We find that the three-parameter model, while mathematically established and applicable to Venus topography, is overparameterized, and thus reduce the results to a two-parameter description of the peak spectral variance and the range-to-half-peak variance (in function of the wavenumber). With the reduction the clustering of geological region types in two-parameter space becomes promising. Finally, we perform inversions for the JOINT spectral variance of
The Measurement of the Speed of Light Using a Laser Pointer.
Mak, Se-yuen; Yip, Din-yan
2000-01-01
Presents a method for measuring the speed of light using a laser pointer with adjustable focus as the signal carrier, a signal generator to modulate the light beam, and a student oscilloscope to detect the phase shift. (Author/CCM)
The Measurement of the Speed of Light Using a Laser Pointer.
Mak, Se-yuen; Yip, Din-yan
2000-01-01
Presents a method for measuring the speed of light using a laser pointer with adjustable focus as the signal carrier, a signal generator to modulate the light beam, and a student oscilloscope to detect the phase shift. (Author/CCM)
Augmenting a Laser Pointer with a Diffraction Grating for Monoscopic 6DOF Detection
Elmar Bomberg
2008-02-01
Full Text Available This article illustrates the detection of 6 degrees of freedom (DOF for Virtual Environment interactions using a modified simple laser pointer device and a camera. The laser pointer is combined with a diffraction grating to project a unique laser grid onto the projection planes used in projection-based immersive VR setups. The distortion of the projected grid is used to calculate the translational and rotational degrees of freedom required for human-computer interaction purposes.
Kim, Kyungsoo; Lim, Sung-Ho; Lee, Jaeseok; Kang, Won-Seok; Moon, Cheil; Choi, Ji-Woong
2016-06-16
Electroencephalograms (EEGs) measure a brain signal that contains abundant information about the human brain function and health. For this reason, recent clinical brain research and brain computer interface (BCI) studies use EEG signals in many applications. Due to the significant noise in EEG traces, signal processing to enhance the signal to noise power ratio (SNR) is necessary for EEG analysis, especially for non-invasive EEG. A typical method to improve the SNR is averaging many trials of event related potential (ERP) signal that represents a brain's response to a particular stimulus or a task. The averaging, however, is very sensitive to variable delays. In this study, we propose two time delay estimation (TDE) schemes based on a joint maximum likelihood (ML) criterion to compensate the uncertain delays which may be different in each trial. We evaluate the performance for different types of signals such as random, deterministic, and real EEG signals. The results show that the proposed schemes provide better performance than other conventional schemes employing averaged signal as a reference, e.g., up to 4 dB gain at the expected delay error of 10°.
Kim, Kyungsoo; Lim, Sung-Ho; Lee, Jaeseok; Kang, Won-Seok; Moon, Cheil; Choi, Ji-Woong
2016-01-01
Electroencephalograms (EEGs) measure a brain signal that contains abundant information about the human brain function and health. For this reason, recent clinical brain research and brain computer interface (BCI) studies use EEG signals in many applications. Due to the significant noise in EEG traces, signal processing to enhance the signal to noise power ratio (SNR) is necessary for EEG analysis, especially for non-invasive EEG. A typical method to improve the SNR is averaging many trials of event related potential (ERP) signal that represents a brain’s response to a particular stimulus or a task. The averaging, however, is very sensitive to variable delays. In this study, we propose two time delay estimation (TDE) schemes based on a joint maximum likelihood (ML) criterion to compensate the uncertain delays which may be different in each trial. We evaluate the performance for different types of signals such as random, deterministic, and real EEG signals. The results show that the proposed schemes provide better performance than other conventional schemes employing averaged signal as a reference, e.g., up to 4 dB gain at the expected delay error of 10°. PMID:27322267
Kyungsoo Kim
2016-06-01
Full Text Available Electroencephalograms (EEGs measure a brain signal that contains abundant information about the human brain function and health. For this reason, recent clinical brain research and brain computer interface (BCI studies use EEG signals in many applications. Due to the significant noise in EEG traces, signal processing to enhance the signal to noise power ratio (SNR is necessary for EEG analysis, especially for non-invasive EEG. A typical method to improve the SNR is averaging many trials of event related potential (ERP signal that represents a brain’s response to a particular stimulus or a task. The averaging, however, is very sensitive to variable delays. In this study, we propose two time delay estimation (TDE schemes based on a joint maximum likelihood (ML criterion to compensate the uncertain delays which may be different in each trial. We evaluate the performance for different types of signals such as random, deterministic, and real EEG signals. The results show that the proposed schemes provide better performance than other conventional schemes employing averaged signal as a reference, e.g., up to 4 dB gain at the expected delay error of 10°.
McInerney, F. A.; Bloch, J. I.; Secord, R.; Wing, S. L.; Kraus, M. J.; Boyer, D. M.
2009-12-01
The Paleocene-Eocene Thermal Maximum (PETM) presents an opportunity to characterize continental hydrologic changes during rapid and extreme global warming. The Bighorn Basin, Wyoming, USA, has long been recognized for the PETM sequences preserved there and sits in an ideal location for recording hydrologic changes in the interior of North America. The southeast Bighorn Basin is of particular interest because it contains not only alluvial paleosols and vertebrate fossils, but also macrofloral remains from the PETM. The carbon isotope excursion associated with this event is preserved in this part of the Basin in leaf wax lipids, tooth enamel, and bulk organic matter. To characterize the hydrologic changes that occurred during the PETM we are applying a suite of isotopic, paleobotanical and paleopedological approaches to sections in the southeast Bighorn Basin. Reported here are results from the combined hydrogen and oxygen isotope analysis aimed at reconstructing relative humidity. Oxygen isotope ratios (δ18O) of biogenic apatite from mammalian tooth enamel and fish scales vary with environment, physiology and diet. Because mammals are homeothermic, they primarily track surface water values with predictable physiological offsets. Hydrogen isotope ratios (δD) of leaf-wax lipids (long-chain n-alkanes) reflect both meteoric water δD values and additional D-enrichment caused by evapotranspiration. The enrichment factor between water δD and n-alkane δD can therefore be used as a proxy for relative humidity (RH). In this study, δ18O of surface water is estimated using the δ18O of Coryphodon tooth enamel. We use these δ18O values to estimate surface water δD values using the Global Meteoric Water Line (δD = 8δ18O + 10). We then calculate relative humidity from n-alkane δD values using a Craig-Gordon type isotopic model for D-enrichment caused by transpiration from leaves. Results of the combined hydrogen-oxygen isotope paleohygrometer indicate a general rise in
Kinkhabwala, Ali
2013-01-01
The most fundamental problem in statistics is the inference of an unknown probability distribution from a finite number of samples. For a specific observed data set, answers to the following questions would be desirable: (1) Estimation: Which candidate distribution provides the best fit to the observed data?, (2) Goodness-of-fit: How concordant is this distribution with the observed data?, and (3) Uncertainty: How concordant are other candidate distributions with the observed data? A simple unified approach for univariate data that addresses these traditionally distinct statistical notions is presented called "maximum fidelity". Maximum fidelity is a strict frequentist approach that is fundamentally based on model concordance with the observed data. The fidelity statistic is a general information measure based on the coordinate-independent cumulative distribution and critical yet previously neglected symmetry considerations. An approximation for the null distribution of the fidelity allows its direct conversi...
A General Low-Cost Indirect Branch Prediction Using Target Address Pointers
谢子超; 佟冬; 黄明凯
2014-01-01
Nowadays energy-eﬃciency becomes the first design metric in chip development. To pursue higher energy eﬃciency, the processor architects should reduce or eliminate those unnecessary energy dissipations. Indirect-branch pre-diction has become a performance bottleneck, especially for the applications written in object-oriented languages. Previous hardware-based indirect-branch predictors are generally ineﬃcient, for they either require significant hardware storage or predict indirect-branch targets slowly. In this paper, we propose an energy-eﬃcient indirect-branch prediction technique called TAP (target address pointer) prediction. Its key idea includes two parts: utilizing specific hardware pointers to accelerate the indirect branch prediction flow and reusing the existing processor components to reduce additional hardware costs and power consumption. When fetching an indirect branch, TAP prediction first gets the specific pointers called target address pointers from the conditional branch predictor, and then uses such pointers to generate virtual addresses which index the indirect-branch targets. This technique spends similar time compared to the dedicated storage techniques without requiring additional large amounts of storage. Our evaluation shows that TAP prediction with some representative state-of-the-art branch predictors can improve performance significantly over the baseline processor. Compared with those hardware-based indirect-branch predictors, the TAP-Perceptron scheme achieves performance improvement equivalent to that provided by an 8 K-entry TTC predictor, and also outperforms the VPC predictor.
Pointer years in the growth of beech trees of the NP "Đerdap" area
Stajić Branko
2014-01-01
Full Text Available This paper defines pointer years and years with distinctive tree rings in the growth of beech in 3 sites (ecological units-EU of the area of Đerdap, in order to observe the reactions of trees to the effects of various factors of growth. The pointer years were determined by the Schweingruber (1983 methodology. The years with distintive tree rings were defined as the years of beech growth with very pronounced distinctive growth rings (annual ring width at least ± 2 standard deviations higher or lower than the arithmetic mean and pronounced typical growth rings (annual ring width at least ± 1.5 standard deviation higher or lower than the arithmetic mean. The common pointer years for the growth of beech in the site conditions of all three ecological units are 1977 and 1988 (negative pointer years. A particular pointer year is 1988, when in more than 90% of trees under the analyzed environmental conditions the marked reduction of tree rings width of the trees (EUB or very pronounced reduction in the width of tree rings (EUA and ESV were observed. [Projekat Ministarstva nauke Republike Srbije, br. 43007: Istraživanje klimatskih promena na životnu sredinu: praćenje uticaja, adaptacija i ublažavanje
M. Génova
2012-04-01
Full Text Available The study of pointer years of numerous tree-ring chronologies of the central Iberian Peninsula (Sierra de Guadarrama could provide complementary information about climate variability over the last 405 yr. In total, 64 pointer years have been identified: 30 negative (representing minimum growths and 34 positive (representing maximum growths, the most significant of these being 1601, 1963 and 1996 for the negative ones, and 1734 and 1737 for the positive ones. Given that summer precipitation was found to be the most limiting factor for the growth of Pinus in the Sierra de Guadarrama in the second half of the 20th century, it is also an explanatory factor in almost 50% of the extreme growths. Furthermore, these pointer years and intervals are not evenly distributed throughout time. Both in the first half of the 17th and in the second half of 20th, they were more frequent and more extreme and these periods are the most notable for the frequency of negative pointer years in Central Spain. The interval 1600–1602 is of special significance, being one of the most unfavourable for tree growth in the centre of Spain, with 1601 representing the minimum index in the regional chronology. We infer that this special minimum annual increase was the effect of the eruption of Huaynaputina, which occurred in Peru at the beginning of 1600 AD. This is the first time that the effects of this eruption in the tree-ring records of Southern Europe have been demonstrated.
Kobayashi, A; Yoneda, T; Yoshikawa, M; Ikuno, M; Takenaka, H; Fukuoka, A; Narita, N; Nezu, K
2000-01-01
To assess the factors determining maximum exercise performance in patients with chronic obstructive pulmonary disease (COPD), we examined nutritional status with special reference to body composition and pulmonary function in 50 stable COPD patients. Nutritional status was evaluated by body weight and body composition, including fat mass (FM) and fat-free mass (FFM) assessed by bioelectrical impedance analysis (BIA). Exercise performance was evaluated by maximum oxygen uptake (Vo(2max)) on a cycle ergometer. A total of 50 patients (FEV(1) = 0.98 L) was divided randomly into either a study group (group A, n = 25) or validation group (group B, n = 25). Stepwise regression analysis was performed in group A to determine the best predictors of Vo(2max) from measurements of pulmonary function and nutritional status. Stepwise regression analysis revealed that Vo(2max) was predicted best by the following equation in group A: Vo(2max) (mL/min) = 10.223 x FFM (kg) + 4.188 x MVV (L/min) + 9.952 x DL(co) (mL/min/mmHg) - 127.9 (r = 0.84, p equation was then cross-validated in group B: Measured Vo(2max) (mL/min) = 1.554 x Predicted Vo(2max) (mL/min) - 324.0 (r = 0.87, p < 0.001). We conclude that FFM is an important factor in determining maximum exercise performance, along with pulmonary function parameters, in patients with COPD.
The Trainee Teacher and His Practice Class. Fifty Pointers for the Student-Teacher.
Rees, Alun L. W.
1969-01-01
This handbook, based on the author's experience of supervising the English practice-classes of trainee teachers, was originally compiled for the specific use of students at the National University of Trujillo, Peru, and consists of a list of pointers embracing the most prevalent of trainees' shortcomings observed over a period of years at all…
Who is pointing when to whom: on model-checking pointer structures
Distefano, Dino; Rensink, Arend; Katoen, Joost-Pieter
2003-01-01
This paper introduces a new model to reason about systems composed by entities that can refer to each other via pointers, such as objects in an object-based system. The model, based on History-Dependent Automata, treats particular cases of unboundedness by a special layered mechanism of abstraction.
Wearable Laser Pointer Versus Head-mounted Display for Tele-guidance Applications?
Jalaliniya, Shahram; Pederson, Thomas; Houben, Steven
2014-01-01
alternatives to Head-Mounted Displays for indicating where in the physical environment the local agent should direct her/his attention. The potential beneﬁt of the laser pointer would be reduced eye fatigue, due to the fact that the documented refocusing challenges associated with HMD use would be completely...
A Laser-Pointer-Based Spectrometer for Endpoint Detection of EDTA Titrations
Dahm, Christopher E.; Hall, James W.; Mattioni, Brian E.
2004-01-01
A laser spectrometer for the ethylenediaminetetra-acetic acid (EDTA) titration of magnesium or calcium ions that is designed around a handheld laser pointer as the source and a photoresistor as the detector is developed. Findings show that the use of the spectrometer reduces the degree of uncertainty and error in one part of the EDTA titrations,…
Laser Pointers: Low-Cost, Low-Tech Innovative, Interactive Instruction Tool
Zdravkovska, Nevenka; Cech, Maureen; Beygo, Pinar; Kackley, Bob
2010-01-01
This paper discusses the use of laser pointers at the Engineering and Physical Sciences Library, University of Maryland, College Park, as a personal response system (PRS) tool to encourage student engagement in and interactivity with one-shot, lecture-based information literacy sessions. Unlike more sophisticated personal response systems like…
Post-selected von Neumann measurement with Hermite-Gaussian and Laguerre-Gaussian pointer states
Turek, Yusuf; Akutsu, Tomotada; Sun, Chang-Pu; Shikano, Yutaka
2015-01-01
Through the von Neumann interaction followed by post-selection, we can extract not only the eigenvalue of an observable of the measured system but also the weak value. In this post-selected von Neumann measurement, the initial pointer state of the measuring device is assumed to be a fundamental Gaussian wave function. By considering the optical implementation of the post-selected von Neumann measurement, higher-order Gaussian modes can be used. In this paper, we consider the Hermite--Gaussian (HG) and Laguerre--Gaussian (LG) modes as pointer states and calculate the average shift of the pointer states of the post-selected von Neumann measurement by assuming the system observable $\\hat{A}$ with $\\hat{A}^{2}=\\hat{I}$ and $\\hat{A}^{2}=\\hat{A}$ for an arbitrary interaction strength, where $\\hat{I}$ represents the identity operator. Our results show that the HG and LG pointer states for a given coupling direction have advantages and disadvantages over the fundamental Gaussian mode in improving the signal-to-noise ...
Using Local Perturbations To Manipulate and Control Pointer States in Quantum Dot Systems
Akis, Richard; Speyer, Gil; Ferry, David; Brunner, Roland
2012-02-01
Recently, scanning gate microscopy (SGM) was used to image scarred wave functions in an open InAs quantum dot[1]. The SGM tip provides a local potential perturbation and imaging is performed by measuring changes in conductance. Scarred wave functions, long associated with quantum chaos, have been shown in open dots to correspond to pointer states[2], eigenstates that survive the decoherence process that occurs via coupling to the environment. Pointer states modulate the conductance, yielding periodic fluctuations and the scars, normally thought unstable, are stabilized by quantum Darwinism [3]. We shall show that, beyond probing, pointer states can be manipulated by local perturbations. Particularly interesting effects occur in coupled quantum dot arrays, where a pointer state localized in one dot can be shifted over into another with a perturbation in a completely different part of the system. These nonlocal effects may perhaps be exploited to give such systems an exotic functionality. [1] A. M. Burke, R. Akis, T. E. Day, Gil Speyer, D. K. Ferry, and B. R. Bennett, Phys. Rev. Lett. 104, 176801 (2010). [2] D. K. Ferry, R. Akis, and J. P. Bird, Phys. Rev. Lett. 104, 176801 (2004). [3] R. Brunner, R. Akis,D. K. Ferry, F. Kuchar,and R. Meisels, Phys. Rev. Lett. 101, 024102 (2008).
Tremor side effects of salbutamol, quantified by a laser pointer technique.
Nizet, T.; Broeders, M.E.A.C.; Folgering, H.T.M.
2004-01-01
OBJECTIVE: To study tremor side effects of salbutamol an easily applicable, quick and low-priced method is needed. A new method using a commercially available, pen-shaped laser pointer was developed. Aim of the study was to determine sensitivity, reproducibility, reference values and the agreement w
Wearable Laser Pointer Versus Head-mounted Display for Tele-guidance Applications?
Jalaliniya, Shahram; Pederson, Thomas; Houben, Steven
2014-01-01
alternatives to Head-Mounted Displays for indicating where in the physical environment the local agent should direct her/his attention. The potential beneﬁt of the laser pointer would be reduced eye fatigue, due to the fact that the documented refocusing challenges associated with HMD use would be completely...
Laser-Induced Fluorescence in Gaseous [I[subscript]2] Excited with a Green Laser Pointer
Tellinghuisen, Joel
2007-01-01
A green laser pointer could be used in a flashy demonstration of laser-induced fluorescence in the gas phase by directing the beam of the laser through a cell containing [I[subscript]2] at its room temperature vapor pressure. The experiment could be used to provide valuable insight into the requirements for laser-induced fluorescence (LIF) and the…
Directions in healthcare research: pointers from retailing and services marketing.
Van Rompay, Thomas L J; Tanja-Dijkstra, Karin
2010-01-01
Although the importance of the environment in relation to healing processes has been well established, empirical evidence for environmental effects on patient well-being and behavior is sparse. In addition, few attempts have been made to integrate insights from related fields of research such as retailing and services marketing with findings from healthcare studies. In this paper, relevant findings and insights from these domains are discussed. What insights and findings from retailing and services marketing are (potentially) of interest to the healthcare context, and how should one interpret and follow up on these results in healthcare environments? Research in retailing and services marketing indicates that physical environmental factors (i.e., music and scent) and social environmental factors (i.e., crowded conditions) may affect consumer satisfaction and well-being. In addition, environmental effects have been shown to vary with contextual factors (e.g., the type of environment) and consumer needs (e.g., the extent to which consumers value social contact or stimulation in a specific setting). Although the evidence base for environmental factors in health environments is steadily growing, few attempts have been made to integrate findings from both domains. The findings presented indicate that environmental variables such as music and scent can contribute to patient well-being and overall satisfaction. In addition, findings suggest that these variables may be used to counteract the negative effects resulting from crowded conditions in different healthcare units. Taking into account recent developments in the healthcare industry, the importance of creating memorable and pleasant patient experiences is likely to grow in the years to come. Hence, the finding that subtle and relatively inexpensive manipulations may affect patient well-being in profound ways should inspire follow-up research aimed at unraveling the specifics of environmental influences in health
Andrieux, A.; Vandanjon, P. O.; Lengelle, R.; Chabanon, C.
2010-12-01
Tyre-road estimation methods have been the objective of many research programmes throughout the world. Most of these methods aim at estimating the friction components such as tyre longitudinal slip rate κ and friction coefficient μ in the contact patch area. In order to estimate the maximum available friction coefficient μmax, these methods generally use a probabilistic relationship between the grip obtained for low tyre excitations (such as constant speed driving) and the grip obtained for high tyre excitations (such as emergency braking manoeuvre). Confirmation or invalidation of this relationship from experimental results is the purpose of this paper. Experiments have been carried out on a reference track including several test boards corresponding to a wide textural spectrum. The main advantage of these experiments lies in the use of a vehicle allowing us to accurately build point-by-point relationship between κ and μ. This relationship has been determined for different tyres and pavement textures. Finally, the curves obtained are analysed to check the validity of the relationship between the current friction coefficient used by the car during normal driving conditions and μmax.
Meter reading recognition method via the pointer region feature%基于指针区域特征的仪表读数识别算法
宋伟; 张文杰; 张家齐; 王玉平; 周庆; 石为人
2014-01-01
Reading recognition algorithms for pointer type meter are vulnerable to illumination, and recognition rate for thicker pointer is not high. In this paper, we propose a reading recognition method based on pointer region. In detail, this approach can be divided into two phases: in the first phase, the proposed method designs Gaussian homomorphic high-pass filter to enhance the adaptability to the illumination change for instrument image, and utilizes image preprocessing methods (including image subtraction, binarization, median filter, image morphology processing et al.) to reduce noise interference, suppress of background region, outstand pointer region. In the second phase, the presented method firstly rotates the meter image, and computes projection of the rotated image on the vertical axis in each rotation angle, then calculates the rotated angle which rotate the pointer of meter image to parallel to the horizontal axis. Whereafter, the deflection angle of pointer can be derived for combining the projection of rotated image on the horizontal axis which the pointer of rotated image parallels to the horizontal axis. Finally, reading recognition is identified by the relation between the pointer angle and the corresponding instruments scale. This approach is simple and stable, the error between the reading of the proposed approach and artificial reading is small, and it can better adapt to illumination changes.%针对目前指针式仪表读数识别算法易受光照影响，并且对较粗指针的仪表识别率不高的问题，本文提出一种基于指针区域的仪表读数识别算法。本文方法可分为两个阶段：第一阶段设计高斯型同态高通滤波器，增强仪表图像对光照变化的适应性；并通过减影法、二值化、中值滤波、图像形态学处理等预处理方法减缓噪声、仪表非指针区域的干扰，突出仪表图像指针区域。第二阶段本文通过计算指针图像旋转不同角度后在纵轴
Ruiz, Maria Cristina; Ayala, Victoria; Portero-Otín, Manel; Requena, Jesús R; Barja, Gustavo; Pamplona, Reinald
2005-10-01
Aging affects all organisms and its basic mechanisms are expected to be conserved across species. Oxidation of proteins has been proposed to be one of the basic mechanisms linking oxygen radicals with the basic aging process. If oxidative damage to proteins is involved in aging, long-lived animals (which age slowly) should show lower levels of markers of this kind of damage than short-lived ones. However, this possibility has not been investigated yet. In this study, steady-state levels of markers of different kinds of protein damage--oxidation (glutamic and aminoadipic semialdehydes), mixed glyco- and lipoxidation (carboxymethyl- and carboxyethyllysine), lipoxidation (malondialdehydelysine) and amino acid composition--were measured in the heart of eight mammalian species ranging in maximum life span (MLSP) from 3.5 to 46 years. Oxidation markers were directly correlated with MLSP across species. Mixed glyco- and lipoxidation markers did not correlate with MLSP. However, the lipoxidation marker malondialdehydelysine was inversely correlated with MLSP (r2=0.85; P<0.001). The amino acid compositional analysis revealed that methionine is the only amino acid strongly correlated MLSP and that such correlation is negative (r2=0.93; P<0.001). This trait may contribute to lower steady-state levels of oxidized methionine residues in cellular proteins. These results reinforce the notion that high longevity in homeothermic vertebrates is achieved in part by constitutively decreasing the sensitivity of both tissue proteins and lipids to oxidative damage. This is obtained by modifying the constituent structural components of proteins and lipids, selecting those less sensitive to oxidative modifications.
Certified Absence of Dangling Pointers in a Language with Explicit Deallocation
de Dios, Javier; Montenegro, Manuel; Peña, Ricardo
Safe is a first-order eager functional language with facilities for programmer controlled destruction of data structures. It provides also regions, i.e. disjoint parts of the heap, where the program allocates data structures, so that the runtime system does not need a garbage collector. A region is a collection of cells, each one big enough to allocate a data constructor. Deallocating cells or regions may create dangling pointers. The language is aimed at inferring and certifying memory safety properties in a Proof Carrying Code like environment. Some of its analyses have been presented elsewhere. The one relevant to this paper is a type system and a type inference algorithm guaranteeing that well-typed programs will be free of dangling pointers at runtime.
Kronbak, Jacob; Leleur, Steen
2002-01-01
When evaluating large infrastructure projects one point to be made is the importance of supplementing the more traditional effects captured by a cost-benefit analysis as presented for example by use of the benefit cost ratio (BCR) with strategic effect modelling to obtain a more comprehensive view...... of the societal importance of the examined project. The focus of this paper is to present the first results of a comparison of the use of a strategic impact measurement (the POINTER index) for road traffic on the Great Belt Link, the Øresund Link and the proposed Fehmarn Belt link. Specifically a number...... of calculation scenarios have been made to assess the possibility of the POINTER approach to indicate the changes in strategic mobility (accessibility) associated with the implementing of one or more of the three fixed links. Finally, conclusions and a research perspective are outlined....
CORBA中的C++灵巧指针%Learners' Garden C++ Smart Pointer in CORBA
杨圣云; 赖国民
2001-01-01
分析了CORBA开发中的-var灵巧指针，示例说明了它的特点与不足，最后讨论了CORBA的本地动态内存管理。%This paper introduces the smart pointer type of -var in CORBA. Some examples are illustrated to show its features and limitations. At last, dynamic management of local memory in CORBA is detailed.
Large Display Interaction via Multiple Acceleration Curves and Multifinger Pointer Control
Andrey Esakia
2014-01-01
Full Text Available Large high-resolution displays combine high pixel density with ample physical dimensions. The combination of these factors creates a multiscale workspace where interactive targeting of on-screen objects requires both high speed for distant targets and high accuracy for small targets. Modern operating systems support implicit dynamic control-display gain adjustment (i.e., a pointer acceleration curve that helps to maintain both speed and accuracy. However, large high-resolution displays require a broader range of control-display gains than a single acceleration curve can usably enable. Some interaction techniques attempt to solve the problem by utilizing multiple explicit modes of interaction, where different modes provide different levels of pointer precision. Here, we investigate the alternative hypothesis of using a single mode of interaction for continuous pointing that enables both (1 standard implicit granularity control via an acceleration curve and (2 explicit switching between multiple acceleration curves in an efficient and dynamic way. We evaluate a sample solution that augments standard touchpad accelerated pointer manipulation with multitouch capability, where the choice of acceleration curve dynamically changes depending on the number of fingers in contact with the touchpad. Specifically, users can dynamically switch among three different acceleration curves by using one, two, or three fingers on the touchpad.
SWIP Prediction: Complexity-Effective Indirect-Branch Prediction Using Pointers
Zi-Chao Xie; Dong Tong; Ming-Kai Huang; Qin-Qing Shi; Xu Cheng
2012-01-01
Predicting indirect-branch targets has become a performance bottleneck for many applications.Previous highperformance indirect-branch predictors usually require significant hardware storage or additional compiler support,which increases the complexity of the processor front-end or the compilers.This paper proposes a complexity-effective indirectbranch prediction mechanism,called the Set-Way Index Pointing (SWIP) prediction.It stores multiple indirect-branch targets in different branch target buffer (BTB) entries,whose set indices and way locations are treated as set-way index pointers.These pointers are stored in the existing branch-direction predictor.SWIP prediction reuses the branch direction predictor to provide such pointers,and then accesses the pointed BTB entries for the predicted indirect-branch target.Our evaluation shows that SWIP prediction could achieve attractive performance improvement without requiring large dedicated storage or additional compiler support.It improves the indirect-branch prediction accuracy by 36.5％ compared to that of a commonly-used BTB,resulting in average performance improvement of 18.56％.Its energy consumption is also reduced by 14.34％ over that of the baseline.
VIRTUAL POINTER UNTUK IDENTIFIKASI ISYARAT TANGAN SEBAGAI PENGENDALI GERAKAN ROBOT SECARA REAL-TIME
M. Isa Irawan
2008-01-01
Full Text Available Many previous researches have been done in relation to human-robot interface, which is an interaction between human and robot using hand gesture. Hand gesture that is used in this research is a moving hand gesture with pointing position. The most important factor to identify hand gesture is the ability to differentiate hands with other objects based on the skin colour. A method to detect hand skin colour is using Fuzzy C-Means (FCM which can refine a cluster centre and the membership value of each data iteratively by minimizing objective function. Hence, the cluster centre moves to the correct location. Recognition result with moving detection method was able to detect the movement of a moving object 91.07944% in 1 second. Skin detection using FCM was able to segment skin colour and not the skin in real time with the successful rate 90.2834%. The successful rate of the hand gesture pattern identification using rule base is 86.67%. The successful rate of virtual hand writing using LVQ artificial neural network as a command for controlling a robot is 79.2%. Abstract in Bahasa Indonesia : Banyak penelitian sebelumnya berrhubungan dengan human robot interface, interaksi manusia dengan robot menggunakan isyarat tangan sebagai bahasa tubuh manusia. Isyarat tangan yang digunakan dalam penelitian ini adalah isyarat tangan bergerak yang berposisi menunjuk untuk identifikasi isyarat tangan, faktor yang paling penting adalah kemampuan membedakan tangan dengan obyek lain berdasarkan warna kulitnya. Metode untuk mendeteksi warna kulit tangan adalah Fuzzy C-Means (FCM yang memiliki kemampuan memperbaiki pusat cluster dan nilai keanggotaan tiap data secara berulang dengan meminimumkan fungsi obyektif, sehingga pusat cluster akan bergerak menuju lokasi yang tepat. Hasil pengenalan dengan metode moving detection, mampu mendeteksi pergerakan obyek bergerak. secara baik sebesar 91.07944% dalam 1 detik. obyek Skin detection dengan Fuzzy C-Means (FCM mampu melakukan
Simpson, Matthew J. R.; Milne, Glenn A.; Huybrechts, Philippe; Long, Antony J.
2009-08-01
We constrain a three-dimensional thermomechanical model of Greenland ice sheet (GrIS) evolution from the Last Glacial Maximum (LGM, 21 ka BP) to the present-day using, primarily, observations of relative sea level (RSL) as well as field data on past ice extent. Our new model (Huy2) fits a majority of the observations and is characterised by a number of key features: (i) the ice sheet had an excess volume (relative to present) of 4.1 m ice-equivalent sea level at the LGM, which increased to reach a maximum value of 4.6 m at 16.5 ka BP; (ii) retreat from the continental shelf was not continuous around the entire margin, as there was a Younger Dryas readvance in some areas. The final episode of marine retreat was rapid and relatively late (c. 12 ka BP), leaving the ice sheet land based by 10 ka BP; (iii) in response to the Holocene Thermal Maximum (HTM) the ice margin retreated behind its present-day position by up to 80 km in the southwest, 20 km in the south and 80 km in a small area of the northeast. As a result of this retreat the modelled ice sheet reaches a minimum extent between 5 and 4 ka BP, which corresponds to a deficit volume (relative to present) of 0.17 m ice-equivalent sea level. Our results suggest that remaining discrepancies between the model and the observations are likely associated with non-Greenland ice load, differences between modelled and observed present-day ice elevation around the margin, lateral variations in Earth structure and/or the pattern of ice margin retreat.
Ngeow, Chow-Choong; Kanbur, Shashi M.; Bhardwaj, Anupam; Schrecengost, Zachariah; Singh, Harinder P.
2017-01-01
Investigation of period–color (PC) and amplitude–color (AC) relations at the maximum and minimum light can be used to probe the interaction of the hydrogen ionization front (HIF) with the photosphere and the radiation hydrodynamics of the outer envelopes of Cepheids and RR Lyraes. For example, theoretical calculations indicated that such interactions would occur at minimum light for RR Lyrae and result in a flatter PC relation. In the past, the PC and AC relations have been investigated by using either the (V ‑ R)MACHO or (V ‑ I) colors. In this work, we extend previous work to other bands by analyzing the RR Lyraes in the Sloan Digital Sky Survey Stripe 82 Region. Multi-epoch data are available for RR Lyraes located within the footprint of the Stripe 82 Region in five (ugriz) bands. We present the PC and AC relations at maximum and minimum light in four colors: (u ‑ g)0, (g ‑ r)0, (r ‑ i)0, and (i ‑ z)0, after they are corrected for extinction. We found that the PC and AC relations for this sample of RR Lyraes show a complex nature in the form of flat, linear or quadratic relations. Furthermore, the PC relations at minimum light for fundamental mode RR Lyrae stars are separated according to the Oosterhoff type, especially in the (g ‑ r)0 and (r ‑ i)0 colors. If only considering the results from linear regressions, our results are quantitatively consistent with the theory of HIF-photosphere interaction for both fundamental and first overtone RR Lyraes.
Saveljev, Vladimir; Kim, Sung-Kyu; Lee, Hyoung; Kim, Hyun-Woo; Lee, Byoungho
2016-02-08
The amplitude of the moiré patterns is estimated in relation to the opening ratio in line gratings and square grids. The theory is developed; the experimental measurements are performed. The minimum and the maximum of the amplitude are found. There is a good agreement between the theoretical and experimental data. This is additionally confirmed by the visual observation. The results can be applied to the image quality improvement in autostereoscopic 3D displays, to the measurements, and to the moiré displays.
Kronbak, Jacob; Leleur, Steen
2002-01-01
When evaluating large infrastructure projects one point to be made is the importance of supplementing the more traditional effects captured by a cost-benefit analysis as presented for example by use of the benefit cost ratio (BCR) with strategic effect modelling to obtain a more comprehensive view...... of the societal importance of the examined project. The focus of this paper is to present the first results of a comparison of the use of a strategic impact measurement (the POINTER index) for road traffic on the Great Belt Link, the Øresund Link and the proposed Fehmarn Belt link. Specifically a number...
ZHANG De'er; Demaree Gaston
2004-01-01
In the context of historical climate records of China and early meteorological measurements of Beijing discovered recently in Europe, a study is undertaken on the 1743 hottest summer of north China over the last 700 a, covering Beijing, Tianjin, and the provinces of Hebei, Shanxi and Shandong, with the highest temperature reaching 44.4℃ in July 1743 in Beijing, in excess of the maximum climate record in the 20th century. Results show that the related weather/climate features of the 1743 heat wave, e.g., flood/ drought distribution and Meiyu activity and the external forcings, such as solar activity and equatorial Pacific SST condition are the same as those of the 1942 and 1999 heat events. It is noted that the 1743 burning summer event occurs in a relatively warm climate background prior to the Industrial Revolution, with a lower level of CO2 release.
Traversal Caches: A Framework for FPGA Acceleration of Pointer Data Structures
James Coole
2010-01-01
Full Text Available Field-programmable gate arrays (FPGAs and other reconfigurable computing (RC devices have been widely shown to have numerous advantages including order of magnitude performance and power improvements compared to microprocessors for some applications. Unfortunately, FPGA usage has largely been limited to applications exhibiting sequential memory access patterns, thereby prohibiting acceleration of important applications with irregular patterns (e.g., pointer-based data structures. In this paper, we present a design pattern for RC application development that serializes irregular data structure traversals online into a traversal cache, which allows the corresponding data to be efficiently streamed to the FPGA. The paper presents a generalized framework that benefits applications with repeated traversals, which we show can achieve between 7x and 29x speedup over pointer-based software. For applications without strictly repeated traversals, we present application-specialized extensions that benefit applications with highly similar traversals by exploiting similarity to improve memory bandwidth and execute multiple traversals in parallel. We show that these extensions can achieve a speedup between 11x and 70x on a Virtex4 LX100 for Barnes-Hut n-body simulation.
Lila-Krasniqi, Zana D.; Shala, Kujtim Sh.; Pustina-Krasniqi, Teuta; Bicaj, Teuta; Dula, Linda J.; Guguvčevski, Ljuben
2015-01-01
Objective: To compare subjects from the group with fixed dentures, the group who present temporomandibular disorders (TMDs) and a control group considering centric relation (CR) and maximum intercuspation (MIC)/habitual occlusion (Hab. Occl.) and to analyze the related variables also compared and analyzed with electronic system T-scan III. Materials and Methods: A total of 54 subjects were divided into three groups; 17 subjects with fixed dentures, 14 with TMD and 23 controls-selection based on anamnesis-responded to a Fonseca questionnaire and clinical measurements analyzed with electronic system T-scan III. Occlusal force, presented by percentage (automatically by the T-scan electronic system) was analyzed in CR and in MIC. Results: Data were presented as mean ± standard deviation and differences in P 0.05 it was not significant in all three groups. Conclusion: In our study, it was concluded that there are not statistically significant differences between CR and MIC in the group of individuals without any symptom or sign of TMD although there are noticed in the group with TMD and fixed dentures disharmonic relation between the arches with overload of the occlusal force on the one side. PMID:26929698
Environmental Fate Data Base (ENVIROFATE): data pointer file (DATALOG), February 1987. Data file
Boethling, R.
1987-02-01
DATALOG is a bibliographic pointer file of environmental fate data and each record contains a chemical CAS registry number (9 bytes), one of 18 data types (9 bytes), and a reference number (6 bytes) to the full citation in the XREF file. The data types identify physical properties, degradation and transport studies, and food, occupational and ambient monitoring data. The file should be used with CASLST (in order to link the chemical name and formula) and XREF (in order to link the full reference citation). A list of possible data types and their abbreviations are: Adsorption (ADSORP), Bioconcentration (BIOCON), Biodegradation (BIODEG), Dissociation constant (DISS CON), Ecosystem (ECOS), Effluent concentrations (EFFL), Evaporation from water (EVAP), Food and crop concentrations (FOOD), Field studies (FIELD), Henry's Law constant (HENRY CON), Hydrolysis (HYDROL), Monitoring (MONIT), Occupational concentrations (OCCUP), Octanol/water partition coefficient (O/W PART), Photoxidation (PHOTOOXID), UV Spectra (UV), Vapor pressure (UP) and Water solubility (WATER SOL).
Effect of laser pointer on students\\\\\\' levels of technical errors during intraoral radiographs
Farshid Shams
2015-12-01
Full Text Available Background and Aims: According to concerns of technical errors in intraoral radiographs and reports of positive effects of central ray laser pointer on the students, technical errors and shortcomings of earlier research, this study was conducted at Department of Oral and Maxillofacial Radiology of Tehran Azad University to determine the role and use of laser pointer in the technical error made by students. Materials and Methods: This clinical trials study was carried out at the Department of Radiology, Dental School of Azad Tehran University. A total of eighty students were divided into two groups of forty. In the first group the periapical technique was taught without laser guided indicator, while in the second group laser guided indicator was applied in training the students. All students took X-rays from the phantoms. Laser guided indicator is made of one low power red laser diodes.Totally 560 X-rays were taken and the technical errors and the quality of X-rays were evaluated. Data were analyzed statistically using Mann-Whitney and Chi Square tests (&alpha=0.05. Results: The X-rays taken by students showed significant differences in cone cutting (P<0.001, and overlapping (P<0.001 between the two study groups. In the subjective review of X-ray quality, the difference was also significant (P=0.03. Conclusion: The use of laser guided indicator as a training aid seems to be practical for easier learning of central ray adjustment and reduction of technical errors in taking X-rays by dental students.
Maximum likelihood-based analysis of photon arrival trajectories in single-molecule FRET
Waligorska, Marta [Adam Mickiewicz University, Faculty of Chemistry, Grunwaldzka 6, 60-780 Poznan (Poland); Molski, Andrzej, E-mail: amolski@amu.edu.pl [Adam Mickiewicz University, Faculty of Chemistry, Grunwaldzka 6, 60-780 Poznan (Poland)
2012-07-25
Highlights: Black-Right-Pointing-Pointer We study model selection and parameter recovery from single-molecule FRET experiments. Black-Right-Pointing-Pointer We examine the maximum likelihood-based analysis of two-color photon trajectories. Black-Right-Pointing-Pointer The number of observed photons determines the performance of the method. Black-Right-Pointing-Pointer For long trajectories, one can extract mean dwell times that are comparable to inter-photon times. -- Abstract: When two fluorophores (donor and acceptor) are attached to an immobilized biomolecule, anti-correlated fluctuations of the donor and acceptor fluorescence caused by Foerster resonance energy transfer (FRET) report on the conformational kinetics of the molecule. Here we assess the maximum likelihood-based analysis of donor and acceptor photon arrival trajectories as a method for extracting the conformational kinetics. Using computer generated data we quantify the accuracy and precision of parameter estimates and the efficiency of the Akaike information criterion (AIC) and the Bayesian information criterion (BIC) in selecting the true kinetic model. We find that the number of observed photons is the key parameter determining parameter estimation and model selection. For long trajectories, one can extract mean dwell times that are comparable to inter-photon times.
Garrido, Nuno D; Silva, António J; Fernandes, Ricardo J; Barbosa, Tiago M; Costa, Aldo M; Marinho, Daniel A; Marques, Mário C
2012-06-01
The relationship between handgrip isometric strength and swimming performance was assessed in the four competitive swimming strokes in swimmers of different age groups and of both sexes. 78 national-level Portuguese swimmers (39 males, 39 females) were selected for this study. Grip strength, previously used as a marker of overall strength to predict future swimming performance, was measured using a hand dynamometer. The best competitive time at 100 and 200 m in all four swimming strokes were converted into 2010 FINA points. Non-parametric tests were used to evaluate differences between groups. Pearson product-moment correlations were computed to verify the association between variables. Handgrip maximum isometric strength was significantly correlated with swimming performance, particularly among female swimmers. Among female age group swimmers, the relationship between handgrip and 100-m freestyle was significant. Handgrip isometric strength seems to be related to swimming performance, especially to 100-m freestyle and in female swimmers. For all other distances and strokes, technique and training probably are more influential than semi-hereditary strength markers such as grip strength.
Oh, Seungkyung
2012-01-01
We perform the largest currently available set of direct N-body calculations of young star cluster models to study the dynamical influence, especially through the ejections of the most massive star in the cluster, on the current relation between the maximum-stellar-mass and the star-cluster-mass. We vary several initial parameters such as the initial half-mass radius of the cluster, the initial binary fraction, and the degree of initial mass segregation. Two different pairing methods are used to construct massive binaries for more realistic initial conditions of massive binaries. We find that lower mass clusters (= 1000 Msun), no most-massive star escapes the cluster within 3 Myr regardless of the initial conditions if clusters have initial half-mass radii, r_0.5, >= 0.8 pc. However, a few of the initially smaller sized clusters (r_0.5 = 0.3 pc), which have a higher density, eject their most massive star within 3 Myr. If clusters form with a compact size and their massive stars are born in a binary system wit...
EFSA Panel on Contaminants in the Food Chain (CONTAM
2013-12-01
Full Text Available The European Food Safety Authority (EFSA was asked to deliver a scientific opinion on the risks for public health related to a possible increase of the maximum level (ML of deoxynivalenol (DON for certain semi-processed cereal products from 750 µg/kg to 1000 µg/kg. For this statement, EFSA relied on existing occurrence data on DON in food collected between 2007 and 2012 and reported by 21 European countries. Due to the lack of appropriate occurrence data from pre-market monitoring, the impact of increasing the ML was estimated using a simulation approach, resulting in an expected increase in mean levels of the respective food products by a factor of 1.14-1.16. Based on median chronic exposure in several age classes, the percentage of consumers exceeding the group provisional maximum tolerable daily intake (PMTDI of 1 μg/kg body weight (b.w. for the sum of DON and its 3- and 15-acetyl-derivatives, established by the Joint FAO/WHO Expert Committee on Food Additives (JECFA in 2010, is approximately 2-fold higher with the suggested increased ML than with the current ML. Several acute exposure scenarios resulted in exceedance of the group acute reference dose (ARfD of 8 µg/kg b.w. established by JECFA with up to 25.9 % of the consumption days above the group ARfD. The EFSA Scientific Panel on Contaminants in the Food Chain notes that the group health based guidance values (HBGVs include 3-Ac-DON and 15-Ac-DON. The exposure from the acetyl-derivatives has not been covered in this statement, since the acetyl-derivatives are not included in the current or suggested increased ML and because only few occurrence data are available. An increase of the DON ML can be expected to be associated with an increase of the levels of DON and Ac-DONs, and can therefore increase the exposure and consequently the exceedances of the group HBGVs.
Peace and Conflict Research in the Age of the Cholera: Ten Pointers to the Future of Peace Studies.
Galtung, Johan
1996-01-01
Presents 10 pointers that can lead to constructive peace making. Covers issues such as a definition of peace; the training of peace workers; the role of the state system in creating conflict; legitimizing peace actions; and suggestions for future peace creation. Discusses the links between direct, structural, and cultural violence. (DSK)
Wilson, John J.; Palaniappan, Ramaswamy
2011-04-01
The steady state visual evoked protocol has recently become a popular paradigm in brain-computer interface (BCI) applications. Typically (regardless of function) these applications offer the user a binary selection of targets that perform correspondingly discrete actions. Such discrete control systems are appropriate for applications that are inherently isolated in nature, such as selecting numbers from a keypad to be dialled or letters from an alphabet to be spelled. However motivation exists for users to employ proportional control methods in intrinsically analogue tasks such as the movement of a mouse pointer. This paper introduces an online BCI in which control of a mouse pointer is directly proportional to a user's intent. Performance is measured over a series of pointer movement tasks and compared to the traditional discrete output approach. Analogue control allowed subjects to move the pointer faster to the cued target location compared to discrete output but suffers more undesired movements overall. Best performance is achieved when combining the threshold to movement of traditional discrete techniques with the range of movement offered by proportional control.
Peace and Conflict Research in the Age of the Cholera: Ten Pointers to the Future of Peace Studies.
Galtung, Johan
1996-01-01
Presents 10 pointers that can lead to constructive peace making. Covers issues such as a definition of peace; the training of peace workers; the role of the state system in creating conflict; legitimizing peace actions; and suggestions for future peace creation. Discusses the links between direct, structural, and cultural violence. (DSK)
Kaufmann, Tobias; Kübler, Andrea
2014-10-01
Objective. The speed of brain-computer interfaces (BCI), based on event-related potentials (ERP), is inherently limited by the commonly used one-stimulus paradigm. In this paper, we introduce a novel paradigm that can increase the spelling speed by a factor of 2, thereby extending the one-stimulus paradigm to a two-stimulus paradigm. Two different stimuli (a face and a symbol) are presented at the same time, superimposed on different characters and ERPs are classified using a multi-class classifier. Here, we present the proof-of-principle that is achieved with healthy participants. Approach. Eight participants were confronted with the novel two-stimulus paradigm and, for comparison, with two one-stimulus paradigms that used either one of the stimuli. Classification accuracies (percentage of correctly predicted letters) and elicited ERPs from the three paradigms were compared in a comprehensive offline analysis. Main results. The accuracies slightly decreased with the novel system compared to the established one-stimulus face paradigm. However, the use of two stimuli allowed for spelling at twice the maximum speed of the one-stimulus paradigms, and participants still achieved an average accuracy of 81.25%. This study introduced an alternative way of increasing the spelling speed in ERP-BCIs and illustrated that ERP-BCIs may not yet have reached their speed limit. Future research is needed in order to improve the reliability of the novel approach, as some participants displayed reduced accuracies. Furthermore, a comparison to the most recent BCI systems with individually adjusted, rapid stimulus timing is needed to draw conclusions about the practical relevance of the proposed paradigm. Significance. We introduced a novel two-stimulus paradigm that might be of high value for users who have reached the speed limit with the current one-stimulus ERP-BCI systems.
S.L. Amosun
2012-12-01
Full Text Available In response to the state mandate to improve access and equityin higher education, the admission policy of universities in South Africa (SAcurrently employs measures for the redress of past inequalities and racialinjustices. As there is no information on the processes to widen access toundergraduate physiotherapy education program in SA, the aim of this reportedstudy was to search for pointers from students’ records in one local university,situated in the Western Province of SA, that would inform the development ofstrategies that will widen the access for previously disadvantaged populationgroups and ensure successful academic outcomes. The records of six cohorts of students who earlier applied for andlater enrolled in the undergraduate physiotherapy program between the years 2000 and 2005 were retrospectivelyreviewed. Information pertaining to access, student characteristics, and academic persistence was extracted, reviewedand analysed descriptively. During the period reviewed, approximately equal numbers of Black and non-Black studentsapplied for admission to the program. The proportion of Black applicants meeting minimum admission requirementswas less than half of the White/Asian applicants. Less than 50% (105/212 of the offers made to Black applicantswere accepted. Forty one percent (43/105 of the enrolled Black students successfully completed the program withinthe minimum 4 years compared to 75.5% (145/192 of the White/Asian students. Strategies should be implementedto increase awareness and recruitment, improve enrolment rates, and improve retention and throughput for Blackstudents in the undergraduate physiotherapy program of a historically “white” SA university.
Maximum Autocorrelation Factorial Kriging
Nielsen, Allan Aasbjerg; Conradsen, Knut; Pedersen, John L.
2000-01-01
This paper describes maximum autocorrelation factor (MAF) analysis, maximum autocorrelation factorial kriging, and its application to irregularly sampled stream sediment geochemical data from South Greenland. Kriged MAF images are compared with kriged images of varimax rotated factors from...
F. TopsÃƒÂ¸e
2001-09-01
Full Text Available Abstract: In its modern formulation, the Maximum Entropy Principle was promoted by E.T. Jaynes, starting in the mid-fifties. The principle dictates that one should look for a distribution, consistent with available information, which maximizes the entropy. However, this principle focuses only on distributions and it appears advantageous to bring information theoretical thinking more prominently into play by also focusing on the "observer" and on coding. This view was brought forward by the second named author in the late seventies and is the view we will follow-up on here. It leads to the consideration of a certain game, the Code Length Game and, via standard game theoretical thinking, to a principle of Game Theoretical Equilibrium. This principle is more basic than the Maximum Entropy Principle in the sense that the search for one type of optimal strategies in the Code Length Game translates directly into the search for distributions with maximum entropy. In the present paper we offer a self-contained and comprehensive treatment of fundamentals of both principles mentioned, based on a study of the Code Length Game. Though new concepts and results are presented, the reading should be instructional and accessible to a rather wide audience, at least if certain mathematical details are left aside at a rst reading. The most frequently studied instance of entropy maximization pertains to the Mean Energy Model which involves a moment constraint related to a given function, here taken to represent "energy". This type of application is very well known from the literature with hundreds of applications pertaining to several different elds and will also here serve as important illustration of the theory. But our approach reaches further, especially regarding the study of continuity properties of the entropy function, and this leads to new results which allow a discussion of models with so-called entropy loss. These results have tempted us to speculate over
Haring, Martijn T.; Liv, Nalan; Zonnevylle, A. Christiaan; Narvaez, Angela C.; Voortman, Lenard M.; Kruit, Pieter; Hoogenboom, Jacob P.
2017-03-01
In the biological sciences, data from fluorescence and electron microscopy is correlated to allow fluorescence biomolecule identification within the cellular ultrastructure and/or ultrastructural analysis following live-cell imaging. High-accuracy (sub-100 nm) image overlay requires the addition of fiducial markers, which makes overlay accuracy dependent on the number of fiducials present in the region of interest. Here, we report an automated method for light-electron image overlay at high accuracy, i.e. below 5 nm. Our method relies on direct visualization of the electron beam position in the fluorescence detection channel using cathodoluminescence pointers. We show that image overlay using cathodoluminescence pointers corrects for image distortions, is independent of user interpretation, and does not require fiducials, allowing image correlation with molecular precision anywhere on a sample.
Haring, Martijn T.; Liv, Nalan; Zonnevylle, A. Christiaan; Narvaez, Angela C.; Voortman, Lenard M.; Kruit, Pieter; Hoogenboom, Jacob P.
2017-01-01
In the biological sciences, data from fluorescence and electron microscopy is correlated to allow fluorescence biomolecule identification within the cellular ultrastructure and/or ultrastructural analysis following live-cell imaging. High-accuracy (sub-100 nm) image overlay requires the addition of fiducial markers, which makes overlay accuracy dependent on the number of fiducials present in the region of interest. Here, we report an automated method for light-electron image overlay at high accuracy, i.e. below 5 nm. Our method relies on direct visualization of the electron beam position in the fluorescence detection channel using cathodoluminescence pointers. We show that image overlay using cathodoluminescence pointers corrects for image distortions, is independent of user interpretation, and does not require fiducials, allowing image correlation with molecular precision anywhere on a sample. PMID:28252673
Kaiadi, Mehrzad; Tunestål, Per; Johansson, Bengt
2010-01-01
High EGR rates combined with turbocharging has been identified as a promising way to increase the maximum load and efficiency of heavy duty spark ignition Natural Gas engines. With stoichiometric conditions a three way catalyst can be used which means that regulated emissions can be kept at very low levels. Most of the heavy duty NG engines are diesel engines which are converted for SI operation. These engine's components are in common with the diesel-engine which put limits on higher exh...
Son, Seungsik; Jeong, Jongpil
2014-01-01
In this paper, a mobility-aware Dual Pointer Forwarding scheme (mDPF) is applied in Proxy Mobile IPv6 (PMIPv6) networks. The movement of a Mobile Node (MN) is classified as intra-domain and inter-domain handoff. When the MN moves, this scheme can reduce the high signaling overhead for intra-handoff/inter-handoff, because the Local Mobility Anchor (LMA) and Mobile Access Gateway (MAG) are connected by pointer chains. In other words, a handoff is aware of low mobility between the previously attached MAG (pMAG) and newly attached MAG (nMAG), and another handoff between the previously attached LMA (pLMA) and newly attached LMA (nLMA) is aware of high mobility. Based on these mobility-aware binding updates, the overhead of the packet delivery can be reduced. Also, we analyse the binding update cost and packet delivery cost for route optimization, based on the mathematical analytic model. Analytical results show that our mDPF outperforms the PMIPv6 and the other pointer forwarding schemes, in terms of reducing the total cost of signaling.
Ojdanic, D.; Chen, L.; Peitgen, H.-O.
2012-02-01
User interaction during navigated surgery is often a critical issue in the overall procedure, as several complex aspects must be considered, such as sterility, workflow, field of view, and cognitive load. This work introduces a new approach for intraoperative interaction that seamlessly fits the high surgical requirements. A navigation system, typically consisting of a tracking system and a monitor for 3D virtual models, is augmented with a tracked pointer with triggering functionality and a pan-tilt mounted laser. The pointer, which is sterile and can be applied for landmark-based organ registration, is used for wireless interaction with the monitor scene. The laser system enables the calibration of the monitor, which is out of the tracking system's range. Moreover, the laser beam can focus on any organ point defined on the virtual model, which improves targeting or visual feedback during intervention. The calibration of the laser system, monitor, and triggered pointer is achieved by an effective procedure, which can be easily repeated in operating room. The mathematical background of the calibration is based on the Levenberg-Marquardt and Umeyama's algorithms.
Maximum Autocorrelation Factorial Kriging
Nielsen, Allan Aasbjerg; Conradsen, Knut; Pedersen, John L.; Steenfelt, Agnete
2000-01-01
This paper describes maximum autocorrelation factor (MAF) analysis, maximum autocorrelation factorial kriging, and its application to irregularly sampled stream sediment geochemical data from South Greenland. Kriged MAF images are compared with kriged images of varimax rotated factors from an ordinary non-spatial factor analysis, and they are interpreted in a geological context. It is demonstrated that MAF analysis contrary to ordinary non-spatial factor analysis gives an objective discrimina...
Liu Li-Li; Jiang Cheng-Bao
2011-01-01
The oxidation microstructure and maximum energy product (BH)max loss of a Sm(Co0.76,Fe0.7,Cu0.1,Zr0.04)7magnet oxidized at 500 ℃ were systematically investigated.Three different oxidation regions were formed in the oxidized magnet:a continuous external oxide scale,an internal reaction layer,and a diffusion zone.Both room-temperature and high-temperature (BH)max losses exhibited the same parabolic increase with oxidation time.An oxygen diffusion model was proposed to simulate the dependence of (BH)max loss on oxidation time.It is found that the external oxide scale has little effect on the (BH)max loss,and both the internal reaction layer and diffusion zone result in the (BH)max loss.Moreover,the diffusion zone leads to more (BH)max loss than the internal reaction layer.The values of the oxidation rate constant k for internal reaction layer and oxygen diffusion coefficient D for diffusion zone were obtained,which are about 1.91× 10-10 cm2/s and 6.54× 10-11 cm2/s,respectively.
Murata, Atsuo; Hayami, Takehito; Moriwaka, Makoto; Takahashi, Rina
2009-01-01
The aim of this paper was to explore the effects of image addition and layered structure on Web search performance on the basis of the search time and the movement trajectory of mouse pointer. The difference of search characteristics between young and older adults was also examined. Older adults tended to take more time to search for the linked item especially when the layered structure was deep. For the deep layered structure, both young and older adults allocate more time to think which ite...
Pointer Mechanism and Source Files Association's Visualization in C++%C++指针机制与源文件关联关系的可视化研究
古辉; 乔凯旋
2012-01-01
Source code as object for research in this paper, mechanism of pointer and relation of association about pointer type object (variable) within multiple source files is studied. Based on the extraction of message and packing up the results, automatically visualize the elation about source files and mechanism of pointer in C plusplus. The storage mechanism on extraction and an algorithm of automatically disposing the include-relation on files are proposed. As a supplement, an algorithm of manual adjustment for meta-graph is introduced. The actual analysis results on codes show that the research is conducive to support for program analysis and understanding of source code.%研究了C++中的指针机制、以及指针类型对象(变量)在多个源程序代码文件中关联关系.基于信息提取和结果整理,计算机可视化实现和表示C++中的指针机制和多源程序代码文件的关联关系.研究了抽取结果的存储机制和基于该机制的自动排序源文件引用关系的方法,最后提出了一种手工调整图元布局的算法,作为自动排序算法的补充.对实际代码分析的结果表明该方法利于程序分析并支持对源代码的辅助理解.
Maximum likely scale estimation
Loog, Marco; Pedersen, Kim Steenstrup; Markussen, Bo
2005-01-01
A maximum likelihood local scale estimation principle is presented. An actual implementation of the estimation principle uses second order moments of multiple measurements at a fixed location in the image. These measurements consist of Gaussian derivatives possibly taken at several scales and/or ...
Maximum information photoelectron metrology
Hockett, P; Wollenhaupt, M; Baumert, T
2015-01-01
Photoelectron interferograms, manifested in photoelectron angular distributions (PADs), are a high-information, coherent observable. In order to obtain the maximum information from angle-resolved photoionization experiments it is desirable to record the full, 3D, photoelectron momentum distribution. Here we apply tomographic reconstruction techniques to obtain such 3D distributions from multiphoton ionization of potassium atoms, and fully analyse the energy and angular content of the 3D data. The PADs obtained as a function of energy indicate good agreement with previous 2D data and detailed analysis [Hockett et. al., Phys. Rev. Lett. 112, 223001 (2014)] over the main spectral features, but also indicate unexpected symmetry-breaking in certain regions of momentum space, thus revealing additional continuum interferences which cannot otherwise be observed. These observations reflect the presence of additional ionization pathways and, most generally, illustrate the power of maximum information measurements of th...
浅析C＋＋程序设计中指针的应用%Analysis of Pointer Application in C＋＋ Programming
姜晨
2011-01-01
Pointer is the core of the C＋＋ programming,to correct and flexible use of function pointers is the premise to achieve the transfer of information,to design a efficient programs,need to combined the pointer to the ordinary variables,arrays,functions.In this paper, to the understanding of pointers,pointers＇ some common uses for C＋＋ programming in the application as well as a detailed analysis in order to show pointer is the key of C＋＋ programming.%指针是c＋＋程序设计的核心，能够正确而灵活地运用指针是实现函数间信息传递的前提，要设计高校的程序，需把指针与普通变量、数组、函数结合。本文针对指针的理解，指针在c＋＋程序设计中的应用方面以及一些常见用法进行了具体分析，从而显示出c＋＋程序设计中指针的关键性。
Alternative Multiview Maximum Entropy Discrimination.
Chao, Guoqing; Sun, Shiliang
2016-07-01
Maximum entropy discrimination (MED) is a general framework for discriminative estimation based on maximum entropy and maximum margin principles, and can produce hard-margin support vector machines under some assumptions. Recently, the multiview version of MED multiview MED (MVMED) was proposed. In this paper, we try to explore a more natural MVMED framework by assuming two separate distributions p1( Θ1) over the first-view classifier parameter Θ1 and p2( Θ2) over the second-view classifier parameter Θ2 . We name the new MVMED framework as alternative MVMED (AMVMED), which enforces the posteriors of two view margins to be equal. The proposed AMVMED is more flexible than the existing MVMED, because compared with MVMED, which optimizes one relative entropy, AMVMED assigns one relative entropy term to each of the two views, thus incorporating a tradeoff between the two views. We give the detailed solving procedure, which can be divided into two steps. The first step is solving our optimization problem without considering the equal margin posteriors from two views, and then, in the second step, we consider the equal posteriors. Experimental results on multiple real-world data sets verify the effectiveness of the AMVMED, and comparisons with MVMED are also reported.
Maximum Likelihood Associative Memories
Gripon, Vincent; Rabbat, Michael
2013-01-01
Associative memories are structures that store data in such a way that it can later be retrieved given only a part of its content -- a sort-of error/erasure-resilience property. They are used in applications ranging from caches and memory management in CPUs to database engines. In this work we study associative memories built on the maximum likelihood principle. We derive minimum residual error rates when the data stored comes from a uniform binary source. Second, we determine the minimum amo...
Maximum likely scale estimation
Loog, Marco; Pedersen, Kim Steenstrup; Markussen, Bo
2005-01-01
A maximum likelihood local scale estimation principle is presented. An actual implementation of the estimation principle uses second order moments of multiple measurements at a fixed location in the image. These measurements consist of Gaussian derivatives possibly taken at several scales and....../or having different derivative orders. Although the principle is applicable to a wide variety of image models, the main focus here is on the Brownian model and its use for scale selection in natural images. Furthermore, in the examples provided, the simplifying assumption is made that the behavior...... of the measurements is completely characterized by all moments up to second order....
C语言中指针教学的研究与探讨%The Research and Discussion on Pointer in C Programming Language Teaching
汤涛
2014-01-01
指针是C语言的一个重点和难点，初学时常会出错。该文结合作者多年的教学实践，对C语言中指针的教学过程以及如何让学生更好地掌握指针进行了较为深入的研究探讨。%Pointer is an important and difficult, C beginners often will go wrong. In this paper, in combination with the authors years of teaching practice, the pointer in C programming language teaching process and how to make students better grasp the pointer has carried on the thorough research.
Regularized maximum correntropy machine
Wang, Jim Jing-Yan
2015-02-12
In this paper we investigate the usage of regularized correntropy framework for learning of classifiers from noisy labels. The class label predictors learned by minimizing transitional loss functions are sensitive to the noisy and outlying labels of training samples, because the transitional loss functions are equally applied to all the samples. To solve this problem, we propose to learn the class label predictors by maximizing the correntropy between the predicted labels and the true labels of the training samples, under the regularized Maximum Correntropy Criteria (MCC) framework. Moreover, we regularize the predictor parameter to control the complexity of the predictor. The learning problem is formulated by an objective function considering the parameter regularization and MCC simultaneously. By optimizing the objective function alternately, we develop a novel predictor learning algorithm. The experiments on two challenging pattern classification tasks show that it significantly outperforms the machines with transitional loss functions.
The Testability of Maximum Magnitude
Clements, R.; Schorlemmer, D.; Gonzalez, A.; Zoeller, G.; Schneider, M.
2012-12-01
Recent disasters caused by earthquakes of unexpectedly large magnitude (such as Tohoku) illustrate the need for reliable assessments of the seismic hazard. Estimates of the maximum possible magnitude M at a given fault or in a particular zone are essential parameters in probabilistic seismic hazard assessment (PSHA), but their accuracy remains untested. In this study, we discuss the testability of long-term and short-term M estimates and the limitations that arise from testing such rare events. Of considerable importance is whether or not those limitations imply a lack of testability of a useful maximum magnitude estimate, and whether this should have any influence on current PSHA methodology. We use a simple extreme value theory approach to derive a probability distribution for the expected maximum magnitude in a future time interval, and we perform a sensitivity analysis on this distribution to determine if there is a reasonable avenue available for testing M estimates as they are commonly reported today: devoid of an appropriate probability distribution of their own and estimated only for infinite time (or relatively large untestable periods). Our results imply that any attempt at testing such estimates is futile, and that the distribution is highly sensitive to M estimates only under certain optimal conditions that are rarely observed in practice. In the future we suggest that PSHA modelers be brutally honest about the uncertainty of M estimates, or must find a way to decrease its influence on the estimated hazard.
Equalized near maximum likelihood detector
2012-01-01
This paper presents new detector that is used to mitigate intersymbol interference introduced by bandlimited channels. This detector is named equalized near maximum likelihood detector which combines nonlinear equalizer and near maximum likelihood detector. Simulation results show that the performance of equalized near maximum likelihood detector is better than the performance of nonlinear equalizer but worse than near maximum likelihood detector.
Could the changes in regional crop yields be a pointer of climatic change?
Trnka, M; Brázdil, R; Olesen, Jørgen E
2012-01-01
by a factor of 2–3 since the late 19th century, whereas the temperature and precipitation patterns showed trends toward warmer and drier conditions. There was a considerable increase in the inter-annual variability of the absolute yields in the districts. When the variability was evaluated in relative terms...
Clarke, Doug M.; Roche, Anne
2009-01-01
As part of individual interviews incorporating whole number and rational number tasks, 323 grade 6 children in Victoria, Australia were asked to nominate the larger of two fractions for eight pairs, giving reasons for their choice. All tasks were expected to be undertaken mentally. The relative difficulty of the pairs was found to be close to that…
Cheeseman, Peter; Stutz, John
2005-01-01
A long standing mystery in using Maximum Entropy (MaxEnt) is how to deal with constraints whose values are uncertain. This situation arises when constraint values are estimated from data, because of finite sample sizes. One approach to this problem, advocated by E.T. Jaynes [1], is to ignore this uncertainty, and treat the empirically observed values as exact. We refer to this as the classic MaxEnt approach. Classic MaxEnt gives point probabilities (subject to the given constraints), rather than probability densities. We develop an alternative approach that assumes that the uncertain constraint values are represented by a probability density {e.g: a Gaussian), and this uncertainty yields a MaxEnt posterior probability density. That is, the classic MaxEnt point probabilities are regarded as a multidimensional function of the given constraint values, and uncertainty on these values is transmitted through the MaxEnt function to give uncertainty over the MaXEnt probabilities. We illustrate this approach by explicitly calculating the generalized MaxEnt density for a simple but common case, then show how this can be extended numerically to the general case. This paper expands the generalized MaxEnt concept introduced in a previous paper [3].
Wong, Siu Ling; Mak, Se-yuen
2008-01-01
We describe the design of a simple homemade apparatus for the measurement of the refractive indices of liquids and demonstration of refraction. A circular transparent plastic tank and a lazy Susan are held concentrically. A laser pointer is mounted on the lazy Susan with its laser beam pointing radially through the centre of the plastic tank.…
Intention, emotion, and action: a neural theory based on semantic pointers.
Schröder, Tobias; Stewart, Terrence C; Thagard, Paul
2014-06-01
We propose a unified theory of intentions as neural processes that integrate representations of states of affairs, actions, and emotional evaluation. We show how this theory provides answers to philosophical questions about the concept of intention, psychological questions about human behavior, computational questions about the relations between belief and action, and neuroscientific questions about how the brain produces actions. Our theory of intention ties together biologically plausible mechanisms for belief, planning, and motor control. The computational feasibility of these mechanisms is shown by a model that simulates psychologically important cases of intention. © 2013 Cognitive Science Society, Inc.
Gruber, H E; Hanley, E N; Sun, Y
2009-12-01
The advent of digital imaging and online submission of manuscripts has created new challenges for authors using histological images. Digital images are used routinely in today's histology research lab and authors must prepare illustrations that meet standards for resolution, color modes, image size, and digital file types for successful online submission to biomedical journals. Because authors may not be familiar with these requirements, our objective here is to present practical guidelines and information for successful image submission online. Ethical issues related to digital imaging and other current topics also are discussed with reference to available online resources.
Yuan-Hong Jiang
Full Text Available OBJECTIVES: The aim of this study was to investigate the predictive values of the total International Prostate Symptom Score (IPSS-T and voiding to storage subscore ratio (IPSS-V/S in association with total prostate volume (TPV and maximum urinary flow rate (Qmax in the diagnosis of bladder outlet-related lower urinary tract dysfunction (LUTD in men with lower urinary tract symptoms (LUTS. METHODS: A total of 298 men with LUTS were enrolled. Video-urodynamic studies were used to determine the causes of LUTS. Differences in IPSS-T, IPSS-V/S ratio, TPV and Qmax between patients with bladder outlet-related LUTD and bladder-related LUTD were analyzed. The positive and negative predictive values (PPV and NPV for bladder outlet-related LUTD were calculated using these parameters. RESULTS: Of the 298 men, bladder outlet-related LUTD was diagnosed in 167 (56%. We found that IPSS-V/S ratio was significantly higher among those patients with bladder outlet-related LUTD than patients with bladder-related LUTD (2.28±2.25 vs. 0.90±0.88, p1 or >2 was factored into the equation instead of IPSS-T, PPV were 91.4% and 97.3%, respectively, and NPV were 54.8% and 49.8%, respectively. CONCLUSIONS: Combination of IPSS-T with TPV and Qmax increases the PPV of bladder outlet-related LUTD. Furthermore, including IPSS-V/S>1 or >2 into the equation results in a higher PPV than IPSS-T. IPSS-V/S>1 is a stronger predictor of bladder outlet-related LUTD than IPSS-T.
Minimum Length - Maximum Velocity
Panes, Boris
2011-01-01
We study a framework where the hypothesis of a minimum length in space-time is complemented with the notion of reference frame invariance. It turns out natural to interpret the action of the obtained reference frame transformations in the context of doubly special relativity. As a consequence of this formalism we find interesting connections between the minimum length properties and the modified velocity-energy relation for ultra-relativistic particles. For example we can predict the ratio between the minimum lengths in space and time using the results from OPERA about superluminal neutrinos.
Volatile displacement of Meghalaya coals – A pointer to explore low sulphur coals
P Behera
2007-04-01
Volatile displacement, which measures the difference between calculated and experimental volatile matter, is indicative of abnormality in coals which may be related to petrological or chemical parameters. The volatile displacement () values of Meghalaya coals were calculated from their chemical analyses. Correlations of volatile displacement () with parameters such as carbon, hydrogen, moisture, oxygen, oxygen plus sulphate sulphur, oxygen plus pyritic sulphur, oxygen plus organic sulphur and total sulphur were studied. An approximately linear relationship exists only between and moisture, and and total sulphur, and not between other parameters. Plots on Seyler’s chart indicate the coals as perhydrous to orthohydrous. The linear relationship with total sulphur indicates that the coals may have become abnormal mainly due to the marine environment of deposition and weathering.
Gargon, Elizabeth; Williamson, Paula R; Young, Bridget
2017-06-01
The objective of the study was to explore core outcome set (COS) developers' experiences of their work to inform methodological guidance on COS development and identify areas for future methodological research. Semistructured, audio-recorded interviews with a purposive sample of 32 COS developers. Analysis of transcribed interviews was informed by the constant comparative method and framework analysis. Developers found COS development to be challenging, particularly in relation to patient participation and accessing funding. Their accounts raised fundamental questions about the status of COS development and whether it is consultation or research. Developers emphasized how the absence of guidance had affected their work and identified areas where guidance or evidence about COS development would be useful including, patient participation, ethics, international development, and implementation. They particularly wanted guidance on systematic reviews, Delphi, and consensus meetings. The findings raise important questions about the funding, status, and process of COS development and indicate ways that it could be strengthened. Guidance could help developers to strengthen their work, but over specification could threaten quality in COS development. Guidance should therefore highlight common issues to consider and encourage tailoring of COS development to the context and circumstances of particular COS. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.
Objects of maximum electromagnetic chirality
Fernandez-Corbaton, Ivan
2015-01-01
We introduce a definition of the electromagnetic chirality of an object and show that it has an upper bound. The upper bound is attained if and only if the object is transparent for fields of one handedness (helicity). Additionally, electromagnetic duality symmetry, i.e. helicity preservation upon scattering, turns out to be a necessary condition for reciprocal scatterers to attain the upper bound. We use these results to provide requirements for the design of such extremal scatterers. The requirements can be formulated as constraints on the polarizability tensors for dipolar scatterers or as material constitutive relations. We also outline two applications for objects of maximum electromagnetic chirality: A twofold resonantly enhanced and background free circular dichroism measurement setup, and angle independent helicity filtering glasses.
宁戎; 洪志祥
2015-01-01
介绍了当前比较通用的一种有料车高炉的自动装料方式，并详细解释了如何通过西门子S7-400PLC指针编程的形式，使用一个关联的数据区，一个动态的巡检字，三套并联的程序结构，实现高炉自动装料的方法。%This paper introduces a common kind of blast furnace with skip car that can automatic charging, and explain how to use the SIEMENS S7-400PLC with pointer program-ming, using a relational data area, a dynamic inspection of words, three sets of parallel pro-gram structure,to achieve automatic charging of blast furnace automatically.
Aragon-choudhury, P
1992-01-01
During a sharing session which took place at a conference sponsored by the Philippine Institute for Social Studies and Action in 1991, Peruvian Victoria Villanueva and US citizen Margaret Ann Schuller discussed their work. Schuller reported on her upcoming book entitled "Freedom from Violence: Women's Strategies Around the World." In addition to proposing a definition of violence against women, the book will include 12 case studies from Malaysia, Bolivia, Mexico, India, Pakistan, Thailand, Sri Lanka, Brazil, Zimbabwe, Chile, Africa, and Alaska describing how national organizations of women are dealing with the problem. An important advance is the development of a framework to look at the connection which exists between violence and health issues. Villanueva described the work of the Movimiento Manuela Ramos, which was organized informally to deal with reproductive rights and abortion and has since expanded to parent groups of women who defend legal and medical cases as paid paralegals. Manuela Ramos uses popular media, traditional drama, and even state television to publicize its issues. Manuela Ramos has accomplished important work on rape, unsafe abortion, and maternal mortality, but most importantly, the women involved with the organization have had the opportunity to develop their self-esteem.
Jadranka Denkova
2013-07-01
Full Text Available With ambiguous authorizations and responsibilities there are no precise and measurable pointers for the efficiency and effectiveness of the public administration. The authorizations and responsibilities of the administration are measureable if there isprecise information supported with measurable indicators. The final result of such setting will influence the motivation of the public administration that is to say, increasing of its effectiveness and efficiency. The aim of this paper is to analyze the states regarding the measuring and the evaluation of the work of the civil servants, the procedures for evaluation and if there are clear indicators though which the achievements of the workers can be measured and their influence to the motivation of the employees in public sector. The main direction in the process of reformation of the administration in Macedonia is creation of professional depoliticized, effective and efficiently civil oriented administration in accordance to theprinciple parliament democracy and responsibility.Suchdetermination means strengthening of the principle of the law ruling and working according to the law. Beside that it is necessary to strengthen the formal rules and the formal working and management to press the informal public and administrative section, culture and habits, then establishment of more flexible type of management in public administration oriented to results and aims, larger autonomy as larger responsibility in order to increase the effectiveness and the efficiency of the public administration. The significant basis for achievements of these aims is the establishment of public authorizations and responsibilities and precise indicators for measurement of the work that will influence to effectiveness and motivation of the state administration.
Du, Zhen-Yu, E-mail: zdu@nifes.no [National Institute of Nutrition and Seafood Research (NIFES), N-5817 Bergen (Norway); Zhang, Jian [National Institute of Nutrition and Seafood Research (NIFES), N-5817 Bergen (Norway); Institute of Nutrition and Food Safety, Chinese Center for Disease Control and Prevention, Beijing, 100050 (China); Department of Biomedicine, University of Bergen (Norway); Wang, Chunrong; Li, Lixiang; Man, Qingqing [Institute of Nutrition and Food Safety, Chinese Center for Disease Control and Prevention, Beijing, 100050 (China); Lundebye, Anne-Katrine; Froyland, Livar [National Institute of Nutrition and Seafood Research (NIFES), N-5817 Bergen (Norway)
2012-02-01
The risks and benefits of fish from markets in Chinese cities have not previously been fully evaluated. In the present study, 24 common fish species with more than 400 individual samples were collected from markets from five big Chinese cities in 2007. The main nutrients and contaminants were measured and the risk-benefit was evaluated based on recommended nutrient intakes and risk level criteria set by relevant authorities. The comprehensive effects of nutrients and contaminants in marine oily fish were also evaluated using the data of two related human dietary intervention trials performed in dyslipidemic Chinese men and women in 2008 and 2010, respectively. The results showed that concentrations of contaminants analyzed including DDT, PCB{sub 7}, arsenic and cadmium were much lower than their corresponding maximum limits with the exception of the mercury concentration in common carp. Concentrations of POPs and n-3 LCPUFA, mainly EPA and DHA, were positively associated with the lipid content of the fish. With a daily intake of 80-100 g marine oily fish, the persistent organic pollutants in fish would not counteract the beneficial effects of n-3 LCPUFA in reducing cardiovascular disease (CVD) risk markers. Marine oily fish provided more effective protection against CVD than lean fish, particularly for the dyslipidemic populations. The risk-benefit assessment based on the present daily aquatic product intake in Chinese urban residents (44.9 and 62.3 g for the average values for all cities and big cities, respectively) indicated that fish, particularly marine oily fish, can be regularly consumed to achieve optimal nutritional benefits from n-3 LCPUFA, without causing significant contaminant-related health risks. However, the potential health threat from contaminants in fish should still be emphasized for the populations consuming large quantities of fish, particularly wild fish. - Highlights: Black-Right-Pointing-Pointer We collected 24 fish species with more than
Recognition of Instrument Color Pointer Based on Color Extracting Method%基于颜色提取方法的仪表彩色指针识别
冀辰宇; 阚一凡; 焦圣喜
2012-01-01
主要论述了利用颜色提取及图像分割技术对汽车仪表指针位置进行识别处理的原理、步骤以及具体的实现过程。通过使用彩色提取和形态学图像处理技术,成功地提取了仪表盘中红色的指针,确定圆心并利用角度法计算出指针仪表读数。颜色提取方法为HIS颜色空间的坐标区域表示法,与常见的先把仪表图像转化成灰度图再进行处理的方法比较,此方法对于彩色指针仪表的识别更简单有效。%This thesis mostly discusses the image processing and recognition means of the dial pointer position using the technology of color extracting and image segmentation. The red pointer on the dashboard has extracted successfully and confirm the central point of automobile dashboard with the technology of color extracting and image segmentation. Then calculate pointer numerical using angle method. The method of color extracting is coordinate regional representation of color space, and the common method is doing image processing after turning instrument image into grayimage. For the recognition of color pointer, coordinate regional representation of color space is more effective.
基于污点指针的二进制代码缺陷检测%Binary Code Defect Detection Based on Taint Pointer
刘杰; 王嘉捷; 欧阳永基; 王清贤
2012-01-01
Taint pointers are serious threats to the security of data flow and control flow. A method for binary defect detection is proposed, which is based on dynamic taint propagation, dynamic symbolic execution and bound constraint analysis, including introduction of the pointer propagation rules, generation of trigger condition by combing path constraints with bound constraints. It can generate inputs for four types of code defects caused by taint pointer. Test results show that this method reduces the number of test case generation effectively, and a virtual function call hijack and two pointer memory corruption defects are found in the test of Linux system tools.%污点指针严重影响二进制代码数据流和控制流的安全.为此,提出一种二进制代码缺陷检测方法.引入指针污点传播规则,结合路径约束条件和边界约束条件得到缺陷引发条件,构造能够引发4类污点指针代码缺陷的输入数据.在Linux系统下实现ELF二进制代码缺陷检测工具,测试结果表明,该方法能降低测试用例生成数量,并发现Linux系统工具的1个虚函数调用控制缺陷和2个指针内存破坏缺陷.
Robotic hand-eye calibration based on active vision with laser pointer%带激光笔的机器人主动视觉手眼标定
黄朝兴; 陈丹; 唐旭晟
2015-01-01
以固高 GRB-400机器人和摄像机组成手眼系统，在手眼关系旋转矩阵的标定方面，分析了基于主动视觉的标定方法。为实现手眼关系平移向量的标定，提出以固定于机械臂末端的激光笔来获取工件平台上特征点的基坐标，并结合已标定的旋转矩阵来标定平移向量。最后，从图像求取多个特征点之间的距离并与实际值进行误差比较，平面特征点间的长度测量误差在±0.8 mm 之间，表明手眼标定精度较高，可满足机器人进行工件定位与自动抓取的要求。%In terms of the rotation matrix calibration of the hand-eye relation, the calibration method based on active vision for eye-in-hand system which is consisted of Googol GRB-400 robot and camera is analyzed. The world coordinates of the feature point on the workpiece platform are obtained by laser pointer mounted on the mechanical arm which is proposed for the calibration of translation vector of the hand-eye relation combined with the calibrated rotation matrix. Finally , the result can be get that the measurement error range of the lengths between the plane feature points is ±0 . 8 mm, which indicates the hand-eye calibration precision is high, through calculating the distances between multiple feature points from the image and comparing the error with the actual values, and it can meet the requirements of workpiece positioning and automatic grabbing for robot.
Chao, Cherng; Tan, Justin; Castillo, Edward M; Zawaideh, Mazen; Roberts, Anne C; Kinney, Thomas B
2014-08-01
We adapted and evaluated the Microsoft Kinect (touchless interface), Hillcrest Labs Loop Pointer (gyroscopic mouse), and the Apple iPad (multi-touch tablet) for intra-procedural imaging review efficacy in a simulation using MIM Software DICOM viewers. Using each device, 29 radiologists executed five basic interactions to complete the overall task of measuring an 8.1-cm hepatic lesion: scroll, window, zoom, pan, and measure. For each interaction, participants assessed the devices on a 3-point subjective scale (3 = highest usability score). The five individual scores were summed to calculate a subjective composite usability score (max 15 points). Overall task time to completion was recorded. Each user also assessed each device for its potential to jeopardize a sterile field. The composite usability scores were as follows: Kinect 9.9 (out of 15.0; SD = 2.8), Loop Pointer 12.9 (SD = 13.5), and iPad 13.5 (SD = 1.8). Mean task completion times were as follows: Kinect 156.7 s (SD = 86.5), Loop Pointer 51.5 s (SD = 30.6), and iPad 41.1 s (SD = 25.3). The mean hepatic lesion measurements were as follows: Kinect was 7.3 cm (SD = 0.9), Loop Pointer 7.8 cm (SD = 1.1), and iPad 8.2 cm (SD = 1.2). The mean deviations from true hepatic lesion measurement were as follows: Kinect 1.0 cm and for both the Loop Pointer and iPad, 0.9 cm (SD = 0.7). The Kinect had the least and iPad had the most subjective concern for compromising the sterile field. A new intra-operative imaging review interface may be near. Most surveyed foresee these devices as useful in procedures, and most do not anticipate problems with a sterile field. An ideal device would combine iPad's usability and accuracy with the Kinect's touchless aspect.
The maximum rotation of a galactic disc
Bottema, R
1997-01-01
The observed stellar velocity dispersions of galactic discs show that the maximum rotation of a disc is on average 63% of the observed maximum rotation. This criterion can, however, not be applied to small or low surface brightness (LSB) galaxies because such systems show, in general, a continuously rising rotation curve until the outermost measured radial position. That is why a general relation has been derived, giving the maximum rotation for a disc depending on the luminosity, surface brightness, and colour of the disc. As a physical basis of this relation serves an adopted fixed mass-to-light ratio as a function of colour. That functionality is consistent with results from population synthesis models and its absolute value is determined from the observed stellar velocity dispersions. The derived maximum disc rotation is compared with a number of observed maximum rotations, clearly demonstrating the need for appreciable amounts of dark matter in the disc region and even more so for LSB galaxies. Matters h...
OECD Maximum Residue Limit Calculator
With the goal of harmonizing the calculation of maximum residue limits (MRLs) across the Organisation for Economic Cooperation and Development, the OECD has developed an MRL Calculator. View the calculator.
张磊; 李珊; 彭舰; 陈黎; 黎红友
2014-01-01
In recent years, feature-opinion pairs classification of Chinese product review is one of the most important research field in Web data mining technology. In this paper, five types of Chinese dependency relationships for product review have been concluded based on the traditional English dependency grammar. The maximum entropy model is used to predict the opinion-relevant product feature relations. To train the model, a set of feature symbol combinations have been designed by means of Chinese dependency. The experiment result shows that the recall and F-score of our approach could reach 78.68%and 75.36%respectively, which is clearly superior to Hu’s adjacent based method and Popesecu’s pattern based method.%中文产品评论特征词与关联的情感词的分类是观点挖掘的重要研究内容之一。该文改进了英文依存关系语法，总结出5种常用的中文产品评论依存关系；利用最大熵模型进行训练，设计了基于依存关系的复合特征模板。实验证明，应用该复合模板进行特征-情感对的提取，系统的查全率和F-score相比于传统方法，分别提高到78.68%和75.36%。
Maximum likelihood estimation of fractionally cointegrated systems
Lasak, Katarzyna
In this paper we consider a fractionally cointegrated error correction model and investigate asymptotic properties of the maximum likelihood (ML) estimators of the matrix of the cointe- gration relations, the degree of fractional cointegration, the matrix of the speed of adjustment...
Nakada, Masao; Okuno, Jun'ichi; Yokoyama, Yusuke
2016-02-01
Inference of globally averaged eustatic sea level (ESL) rise since the Last Glacial Maximum (LGM) highly depends on the interpretation of relative sea level (RSL) observations at Barbados and Bonaparte Gulf, Australia, which are sensitive to the viscosity structure of Earth's mantle. Here we examine the RSL changes at the LGM for Barbados and Bonaparte Gulf ({{RSL}}_{{L}}^{{{Bar}}} and {{RSL}}_{{L}}^{{{Bon}}}), differential RSL for both sites (Δ {{RSL}}_{{L}}^{{{Bar}},{{Bon}}}) and rate of change of degree-two harmonics of Earth's geopotential due to glacial isostatic adjustment (GIA) process (GIA-induced J˙2) to infer the ESL component and viscosity structure of Earth's mantle. Differential RSL, Δ {{RSL}}_{{L}}^{{{Bar}},{{Bon}}} and GIA-induced J˙2 are dominantly sensitive to the lower-mantle viscosity, and nearly insensitive to the upper-mantle rheological structure and GIA ice models with an ESL component of about (120-130) m. The comparison between the predicted and observationally derived Δ {{RSL}}_{{L}}^{{{Bar}},{{Bon}}} indicates the lower-mantle viscosity higher than ˜2 × 1022 Pa s, and the observationally derived GIA-induced J˙2 of -(6.0-6.5) × 10-11 yr-1 indicates two permissible solutions for the lower mantle, ˜1022 and (5-10) × 1022 Pa s. That is, the effective lower-mantle viscosity inferred from these two observational constraints is (5-10) × 1022 Pa s. The LGM RSL changes at both sites, {{RSL}}_{{L}}^{{{Bar}}} and {{RSL}}_{{L}}^{{{Bon}}}, are also sensitive to the ESL component and upper-mantle viscosity as well as the lower-mantle viscosity. The permissible upper-mantle viscosity increases with decreasing ESL component due to the sensitivity of the LGM sea level at Bonaparte Gulf ({{RSL}}_{{L}}^{{{Bon}}}) to the upper-mantle viscosity, and inferred upper-mantle viscosity for adopted lithospheric thicknesses of 65 and 100 km is (1-3) × 1020 Pa s for ESL˜130 m and (4-10) × 1020 Pa s for ESL˜125 m. The former solution of (1-3) × 1020
曾杰; 张永兴; 靳晓光
2011-01-01
通过分析国内外岩爆预测的判据,选择岩爆发生所需的力学条件、完整性条件、储能条件和脆性条件作为岩爆预测指标.引入岩爆预测的相对隶属度概念,计算了岩爆的相对隶属度模糊矩阵和预测指标的权重,以信息熵来描述并比较岩爆评价中的不确定性,定义了加权广义权距离来表征岩爆的差异.根据最大熵原理建立了岩爆预测的模糊最优化模型,对一些岩石地下工程实例进行了分析,预测结果与其他方法的分析结果以及实际情况基本一致.并将模型运用于葡萄山隧道岩爆预测,预测结果与实际岩爆情况符合较好.%In the analysis of rock burst criterion prediction at home and abroad, the prediction standards of rock burst are selected including the conditions of mechanics integrity, energy and brittle. The concept of relative membership degree on the rock burst prediction was introduced. The weight of standards and fuzzy matrix of relative membership degree are calculated. Uncertainty in rock burst prediction is described and compared according to the information entropy. Generalized weighted distance is also defined to characterize the differences in rock burst based on the maximum entropy principle, the establishment of a rock burst prediction fuzzy optimization model. The results from the application to practical example and comparisons with other methods are fairly good. Finally, the prediction model is applied in Putaoshan tunnel and the predictions are consistent with the actual rock burst.
Revealing the Maximum Strength in Nanotwinned Copper
Lu, L.; Chen, X.; Huang, Xiaoxu
2009-01-01
The strength of polycrystalline materials increases with decreasing grain size. Below a critical size, smaller grains might lead to softening, as suggested by atomistic simulations. The strongest size should arise at a transition in deformation mechanism from lattice dislocation activities to grain...... boundary–related processes. We investigated the maximum strength of nanotwinned copper samples with different twin thicknesses. We found that the strength increases with decreasing twin thickness, reaching a maximum at 15 nanometers, followed by a softening at smaller values that is accompanied by enhanced...
Revealing the Maximum Strength in Nanotwinned Copper
Lu, L.; Chen, X.; Huang, Xiaoxu
2009-01-01
The strength of polycrystalline materials increases with decreasing grain size. Below a critical size, smaller grains might lead to softening, as suggested by atomistic simulations. The strongest size should arise at a transition in deformation mechanism from lattice dislocation activities to grain...... boundary–related processes. We investigated the maximum strength of nanotwinned copper samples with different twin thicknesses. We found that the strength increases with decreasing twin thickness, reaching a maximum at 15 nanometers, followed by a softening at smaller values that is accompanied by enhanced...
Maximum Multiflow in Wireless Network Coding
Zhou, Jin-Yi; Jiang, Yong; Zheng, Hai-Tao
2012-01-01
In a multihop wireless network, wireless interference is crucial to the maximum multiflow (MMF) problem, which studies the maximum throughput between multiple pairs of sources and sinks. In this paper, we observe that network coding could help to decrease the impacts of wireless interference, and propose a framework to study the MMF problem for multihop wireless networks with network coding. Firstly, a network model is set up to describe the new conflict relations modified by network coding. Then, we formulate a linear programming problem to compute the maximum throughput and show its superiority over one in networks without coding. Finally, the MMF problem in wireless network coding is shown to be NP-hard and a polynomial approximation algorithm is proposed.
Research on fitting method of pointer meters scales center straight-line%指针表刻度中心直线拟合方法研究
李学聪; 汪仁煌; 刘洪江
2012-01-01
重点研究直线图像的直线方程拟合方法,提出了一种基于最小距离直线拟合新方法,用于拟合指针表刻度直线.对采集的仪表图像做预处理后,获得仪表各刻度直线的二值图像.使用最小距离直线拟合方法拟合各刻度中心直线,使用刻度拟合直线相交点对拟合直线进行最小二乘修正,提高刻度中心直线的拟合精度.仿真结果表明,该直线拟合方法能快速确定高精度指针表刻度线图像的精确直线方程,有利于提高指针表读数的精度.%The straight-line equation fitting method of the straight-line image is researched in detail, and a straight-line fitting method is proposed based on the least distance for fitting pointer meters scales center straight-line. After pre-processing of pointer meters image, the meter scale linear binary image is obtained. Firstly, the least distance straight-line fitting method is used to fitting each scale center straight-line, then the crosspoints of the scale fitting straight-line is used to fix the fitting line on the least-square correction, and the scale center straight-line fitting precision is improved. Simulation results show that the linear fitting method can rapidly determine the precise equation of the scale center straight-line of the high-precision pointer meters and improve effectively the accuracy of meter readout.
Maximum margin Bayesian network classifiers.
Pernkopf, Franz; Wohlmayr, Michael; Tschiatschek, Sebastian
2012-03-01
We present a maximum margin parameter learning algorithm for Bayesian network classifiers using a conjugate gradient (CG) method for optimization. In contrast to previous approaches, we maintain the normalization constraints on the parameters of the Bayesian network during optimization, i.e., the probabilistic interpretation of the model is not lost. This enables us to handle missing features in discriminatively optimized Bayesian networks. In experiments, we compare the classification performance of maximum margin parameter learning to conditional likelihood and maximum likelihood learning approaches. Discriminative parameter learning significantly outperforms generative maximum likelihood estimation for naive Bayes and tree augmented naive Bayes structures on all considered data sets. Furthermore, maximizing the margin dominates the conditional likelihood approach in terms of classification performance in most cases. We provide results for a recently proposed maximum margin optimization approach based on convex relaxation. While the classification results are highly similar, our CG-based optimization is computationally up to orders of magnitude faster. Margin-optimized Bayesian network classifiers achieve classification performance comparable to support vector machines (SVMs) using fewer parameters. Moreover, we show that unanticipated missing feature values during classification can be easily processed by discriminatively optimized Bayesian network classifiers, a case where discriminative classifiers usually require mechanisms to complete unknown feature values in the data first.
Maximum Entropy in Drug Discovery
Chih-Yuan Tseng
2014-07-01
Full Text Available Drug discovery applies multidisciplinary approaches either experimentally, computationally or both ways to identify lead compounds to treat various diseases. While conventional approaches have yielded many US Food and Drug Administration (FDA-approved drugs, researchers continue investigating and designing better approaches to increase the success rate in the discovery process. In this article, we provide an overview of the current strategies and point out where and how the method of maximum entropy has been introduced in this area. The maximum entropy principle has its root in thermodynamics, yet since Jaynes’ pioneering work in the 1950s, the maximum entropy principle has not only been used as a physics law, but also as a reasoning tool that allows us to process information in hand with the least bias. Its applicability in various disciplines has been abundantly demonstrated. We give several examples of applications of maximum entropy in different stages of drug discovery. Finally, we discuss a promising new direction in drug discovery that is likely to hinge on the ways of utilizing maximum entropy.
Greenslade, Thomas B., Jr.
1985-01-01
Discusses a series of experiments performed by Thomas Hope in 1805 which show the temperature at which water has its maximum density. Early data cast into a modern form as well as guidelines and recent data collected from the author provide background for duplicating Hope's experiments in the classroom. (JN)
Abolishing the maximum tension principle
Dabrowski, Mariusz P
2015-01-01
We find the series of example theories for which the relativistic limit of maximum tension $F_{max} = c^2/4G$ represented by the entropic force can be abolished. Among them the varying constants theories, some generalized entropy models applied both for cosmological and black hole horizons as well as some generalized uncertainty principle models.
Abolishing the maximum tension principle
Mariusz P. Da̧browski
2015-09-01
Full Text Available We find the series of example theories for which the relativistic limit of maximum tension Fmax=c4/4G represented by the entropic force can be abolished. Among them the varying constants theories, some generalized entropy models applied both for cosmological and black hole horizons as well as some generalized uncertainty principle models.
韩旭; 邵玉婷; 孙钦凤; 郭泾
2015-01-01
Objective To investigate the effects of orthodontic intervention on maximum intercuspation (MIC)-centric relation (CR)condylar displacement of patients with or without temporomandibular disorders (TMD).Methods A total of 31 orthodontic patients aged 16 to 45 years were selected and divided into the TMD group (n =15)and non-TMD group (n =16).Records of MIC and CR of these patients taken before and after orthodontic intervention were compared.Results The two groups had different MIC-CR displacement before and after treatment.There were more changes in the TMD group,and the changes were mostly favorable.The MIC-CR condylar displacement was correlated with the symptom checklist (SCL)score.Conclusion Orthodontic intervention has effect on condylar position of pa-tients,especially for those with TMD.Orthodontists,therefore,need to understand and pay attention to the effect of malocclusion on TMD and limitations of measurement of condylar displacement (MCD)in the diagnosis.%目的：旨在探讨正畸干预对最大牙尖交错位（MIC）-正中关系位（CR）髁突位移量的影响。方法选取16～45岁门诊正畸患者31例，其中非颞下颌关节紊乱病（TMD）组16例（NTMD 组），TMD 组15例，在正畸干预前后分别取 MIC 与 CR 位记录。分析两组患者正畸干预前后正中牙合位与正中关系位的髁突位置差异。结果正畸干预前 TMD 组与 NTMD 组 MIC-CR 之间髁突位移（MCD）量存在差异，正畸干预后两组 MIC-CR 位移量减小，TMD组较为显著，且多为有利变化，TMD 组正畸干预前后 MIC-CR 髁突位移量与 TMD 症状自评量表得分呈相关性。结论正畸干预在一定程度上对错牙合畸形患者的髁突位置产生影响，对于 TMD 患者尤甚。因此，正畸医生要认识并注意到错牙合畸形对 TMD 产生和发展的作用，并正确认识 MCD 在诊断中的局限性。
Maximum Genus of Strong Embeddings
Er-ling Wei; Yan-pei Liu; Han Ren
2003-01-01
The strong embedding conjecture states that any 2-connected graph has a strong embedding on some surface. It implies the circuit double cover conjecture: Any 2-connected graph has a circuit double cover.Conversely, it is not true. But for a 3-regular graph, the two conjectures are equivalent. In this paper, a characterization of graphs having a strong embedding with exactly 3 faces, which is the strong embedding of maximum genus, is given. In addition, some graphs with the property are provided. More generally, an upper bound of the maximum genus of strong embeddings of a graph is presented too. Lastly, it is shown that the interpolation theorem is true to planar Halin graph.
Remizov, Ivan D
2009-01-01
In this note, we represent a subdifferential of a maximum functional defined on the space of all real-valued continuous functions on a given metric compact set. For a given argument, $f$ it coincides with the set of all probability measures on the set of points maximizing $f$ on the initial compact set. This complete characterization lies in the heart of several important identities in microeconomics, such as Roy's identity, Sheppard's lemma, as well as duality theory in production and linear programming.
PREDICTION OF MAXIMUM DRY DENSITY OF LOCAL GRANULAR ...
methods. A test on a soil of relatively high solid density revealed that the developed relation looses ... where, Pd max is the laboratory maximum dry ... Addis-Jinima Road Rehabilitation. ..... data sets that differ considerably in the magnitude.
Cacti with maximum Kirchhoff index
Wang, Wen-Rui; Pan, Xiang-Feng
2015-01-01
The concept of resistance distance was first proposed by Klein and Randi\\'c. The Kirchhoff index $Kf(G)$ of a graph $G$ is the sum of resistance distance between all pairs of vertices in $G$. A connected graph $G$ is called a cactus if each block of $G$ is either an edge or a cycle. Let $Cat(n;t)$ be the set of connected cacti possessing $n$ vertices and $t$ cycles, where $0\\leq t \\leq \\lfloor\\frac{n-1}{2}\\rfloor$. In this paper, the maximum kirchhoff index of cacti are characterized, as well...
Generic maximum likely scale selection
Pedersen, Kim Steenstrup; Loog, Marco; Markussen, Bo
2007-01-01
The fundamental problem of local scale selection is addressed by means of a novel principle, which is based on maximum likelihood estimation. The principle is generally applicable to a broad variety of image models and descriptors, and provides a generic scale estimation methodology. The focus...... on second order moments of multiple measurements outputs at a fixed location. These measurements, which reflect local image structure, consist in the cases considered here of Gaussian derivatives taken at several scales and/or having different derivative orders....
Economics and Maximum Entropy Production
Lorenz, R. D.
2003-04-01
Price differentials, sales volume and profit can be seen as analogues of temperature difference, heat flow and work or entropy production in the climate system. One aspect in which economic systems exhibit more clarity than the climate is that the empirical and/or statistical mechanical tendency for systems to seek a maximum in production is very evident in economics, in that the profit motive is very clear. Noting the common link between 1/f noise, power laws and Self-Organized Criticality with Maximum Entropy Production, the power law fluctuations in security and commodity prices is not inconsistent with the analogy. There is an additional thermodynamic analogy, in that scarcity is valued. A commodity concentrated among a few traders is valued highly by the many who do not have it. The market therefore encourages via prices the spreading of those goods among a wider group, just as heat tends to diffuse, increasing entropy. I explore some empirical price-volume relationships of metals and meteorites in this context.
Maximum mutual information regularized classification
Wang, Jim Jing-Yan
2014-09-07
In this paper, a novel pattern classification approach is proposed by regularizing the classifier learning to maximize mutual information between the classification response and the true class label. We argue that, with the learned classifier, the uncertainty of the true class label of a data sample should be reduced by knowing its classification response as much as possible. The reduced uncertainty is measured by the mutual information between the classification response and the true class label. To this end, when learning a linear classifier, we propose to maximize the mutual information between classification responses and true class labels of training samples, besides minimizing the classification error and reducing the classifier complexity. An objective function is constructed by modeling mutual information with entropy estimation, and it is optimized by a gradient descend method in an iterative algorithm. Experiments on two real world pattern classification problems show the significant improvements achieved by maximum mutual information regularization.
The strong maximum principle revisited
Pucci, Patrizia; Serrin, James
In this paper we first present the classical maximum principle due to E. Hopf, together with an extended commentary and discussion of Hopf's paper. We emphasize the comparison technique invented by Hopf to prove this principle, which has since become a main mathematical tool for the study of second order elliptic partial differential equations and has generated an enormous number of important applications. While Hopf's principle is generally understood to apply to linear equations, it is in fact also crucial in nonlinear theories, such as those under consideration here. In particular, we shall treat and discuss recent generalizations of the strong maximum principle, and also the compact support principle, for the case of singular quasilinear elliptic differential inequalities, under generally weak assumptions on the quasilinear operators and the nonlinearities involved. Our principal interest is in necessary and sufficient conditions for the validity of both principles; in exposing and simplifying earlier proofs of corresponding results; and in extending the conclusions to wider classes of singular operators than previously considered. The results have unexpected ramifications for other problems, as will develop from the exposition, e.g. two point boundary value problems for singular quasilinear ordinary differential equations (Sections 3 and 4); the exterior Dirichlet boundary value problem (Section 5); the existence of dead cores and compact support solutions, i.e. dead cores at infinity (Section 7); Euler-Lagrange inequalities on a Riemannian manifold (Section 9); comparison and uniqueness theorems for solutions of singular quasilinear differential inequalities (Section 10). The case of p-regular elliptic inequalities is briefly considered in Section 11.
Maximum-biomass prediction of homofermentative Lactobacillus.
Cui, Shumao; Zhao, Jianxin; Liu, Xiaoming; Chen, Yong Q; Zhang, Hao; Chen, Wei
2016-07-01
Fed-batch and pH-controlled cultures have been widely used for industrial production of probiotics. The aim of this study was to systematically investigate the relationship between the maximum biomass of different homofermentative Lactobacillus and lactate accumulation, and to develop a prediction equation for the maximum biomass concentration in such cultures. The accumulation of the end products and the depletion of nutrients by various strains were evaluated. In addition, the minimum inhibitory concentrations (MICs) of acid anions for various strains at pH 7.0 were examined. The lactate concentration at the point of complete inhibition was not significantly different from the MIC of lactate for all of the strains, although the inhibition mechanism of lactate and acetate on Lactobacillus rhamnosus was different from the other strains which were inhibited by the osmotic pressure caused by acid anions at pH 7.0. When the lactate concentration accumulated to the MIC, the strains stopped growing. The maximum biomass was closely related to the biomass yield per unit of lactate produced (YX/P) and the MIC (C) of lactate for different homofermentative Lactobacillus. Based on the experimental data obtained using different homofermentative Lactobacillus, a prediction equation was established as follows: Xmax - X0 = (0.59 ± 0.02)·YX/P·C.
Molina, L; Elosua, R; Marrugat, J; Pons, S
1999-10-15
The relation between maximum systolic blood pressure (BP) during exercise and left ventricular (LV) mass is controversial. Physical activity also induces LV mass increase. The objective was to assess the relation between BP response to exercise and LV mass in normotensive men, taking into account physical activity practice. A cross-sectional study was performed. Three hundred eighteen healthy normotensive men, aged between 20 and 60 years, participated in this study. The Minnesota questionnaire was used to assess physical activity practice. An echocardiogram and a maximum exercise test were performed. LV mass was calculated and indexed to body surface area. LV hypertrophy was defined as a ventricular mass index > or =134 g/m2. BP was measured at the moment of maximum effort. Hypertensive response was considered when BP was > or =210 mm Hg. In the multiple linear regression model, maximum systolic BP was associated with LV mass index and correlation coefficient was 0.27 (SE 0.07). Physical activity practice and age were also associated with LV mass. An association between hypertensive response to exercise and LV hypertrophy was observed (odds ratio 3.16). Thus, BP response to exercise is associated with LV mass and men with systolic BP response > or =210 mm Hg present a 3-times higher risk of LV hypertrophy than those not reaching this limit. Physical activity practice is related to LV mass, but not to LV hypertrophy.
Maximum-entropy clustering algorithm and its global convergence analysis
无
2001-01-01
Constructing a batch of differentiable entropy functions touniformly approximate an objective function by means of the maximum-entropy principle, a new clustering algorithm, called maximum-entropy clustering algorithm, is proposed based on optimization theory. This algorithm is a soft generalization of the hard C-means algorithm and possesses global convergence. Its relations with other clustering algorithms are discussed.
Maximum entropy production in daisyworld
Maunu, Haley A.; Knuth, Kevin H.
2012-05-01
Daisyworld was first introduced in 1983 by Watson and Lovelock as a model that illustrates how life can influence a planet's climate. These models typically involve modeling a planetary surface on which black and white daisies can grow thus influencing the local surface albedo and therefore also the temperature distribution. Since then, variations of daisyworld have been applied to study problems ranging from ecological systems to global climate. Much of the interest in daisyworld models is due to the fact that they enable one to study self-regulating systems. These models are nonlinear, and as such they exhibit sensitive dependence on initial conditions, and depending on the specifics of the model they can also exhibit feedback loops, oscillations, and chaotic behavior. Many daisyworld models are thermodynamic in nature in that they rely on heat flux and temperature gradients. However, what is not well-known is whether, or even why, a daisyworld model might settle into a maximum entropy production (MEP) state. With the aim to better understand these systems, this paper will discuss what is known about the role of MEP in daisyworld models.
Maximum stellar iron core mass
F W Giacobbe
2003-03-01
An analytical method of estimating the mass of a stellar iron core, just prior to core collapse, is described in this paper. The method employed depends, in part, upon an estimate of the true relativistic mass increase experienced by electrons within a highly compressed iron core, just prior to core collapse, and is signiﬁcantly different from a more typical Chandrasekhar mass limit approach. This technique produced a maximum stellar iron core mass value of 2.69 × 1030 kg (1.35 solar masses). This mass value is very near to the typical mass values found for neutron stars in a recent survey of actual neutron star masses. Although slightly lower and higher neutron star masses may also be found, lower mass neutron stars are believed to be formed as a result of enhanced iron core compression due to the weight of non-ferrous matter overlying the iron cores within large stars. And, higher mass neutron stars are likely to be formed as a result of fallback or accretion of additional matter after an initial collapse event involving an iron core having a mass no greater than 2.69 × 1030 kg.
Maximum Matchings via Glauber Dynamics
Jindal, Anant; Pal, Manjish
2011-01-01
In this paper we study the classic problem of computing a maximum cardinality matching in general graphs $G = (V, E)$. The best known algorithm for this problem till date runs in $O(m \\sqrt{n})$ time due to Micali and Vazirani \\cite{MV80}. Even for general bipartite graphs this is the best known running time (the algorithm of Karp and Hopcroft \\cite{HK73} also achieves this bound). For regular bipartite graphs one can achieve an $O(m)$ time algorithm which, following a series of papers, has been recently improved to $O(n \\log n)$ by Goel, Kapralov and Khanna (STOC 2010) \\cite{GKK10}. In this paper we present a randomized algorithm based on the Markov Chain Monte Carlo paradigm which runs in $O(m \\log^2 n)$ time, thereby obtaining a significant improvement over \\cite{MV80}. We use a Markov chain similar to the \\emph{hard-core model} for Glauber Dynamics with \\emph{fugacity} parameter $\\lambda$, which is used to sample independent sets in a graph from the Gibbs Distribution \\cite{V99}, to design a faster algori...
2011-01-10
...: Establishing Maximum Allowable Operating Pressure or Maximum Operating Pressure Using Record Evidence, and... facilities of their responsibilities, under Federal integrity management (IM) regulations, to perform... system, especially when calculating Maximum Allowable Operating Pressure (MAOP) or Maximum Operating...
Accurate structural correlations from maximum likelihood superpositions.
Douglas L Theobald
2008-02-01
Full Text Available The cores of globular proteins are densely packed, resulting in complicated networks of structural interactions. These interactions in turn give rise to dynamic structural correlations over a wide range of time scales. Accurate analysis of these complex correlations is crucial for understanding biomolecular mechanisms and for relating structure to function. Here we report a highly accurate technique for inferring the major modes of structural correlation in macromolecules using likelihood-based statistical analysis of sets of structures. This method is generally applicable to any ensemble of related molecules, including families of nuclear magnetic resonance (NMR models, different crystal forms of a protein, and structural alignments of homologous proteins, as well as molecular dynamics trajectories. Dominant modes of structural correlation are determined using principal components analysis (PCA of the maximum likelihood estimate of the correlation matrix. The correlations we identify are inherently independent of the statistical uncertainty and dynamic heterogeneity associated with the structural coordinates. We additionally present an easily interpretable method ("PCA plots" for displaying these positional correlations by color-coding them onto a macromolecular structure. Maximum likelihood PCA of structural superpositions, and the structural PCA plots that illustrate the results, will facilitate the accurate determination of dynamic structural correlations analyzed in diverse fields of structural biology.
The Sherpa Maximum Likelihood Estimator
Nguyen, D.; Doe, S.; Evans, I.; Hain, R.; Primini, F.
2011-07-01
A primary goal for the second release of the Chandra Source Catalog (CSC) is to include X-ray sources with as few as 5 photon counts detected in stacked observations of the same field, while maintaining acceptable detection efficiency and false source rates. Aggressive source detection methods will result in detection of many false positive source candidates. Candidate detections will then be sent to a new tool, the Maximum Likelihood Estimator (MLE), to evaluate the likelihood that a detection is a real source. MLE uses the Sherpa modeling and fitting engine to fit a model of a background and source to multiple overlapping candidate source regions. A background model is calculated by simultaneously fitting the observed photon flux in multiple background regions. This model is used to determine the quality of the fit statistic for a background-only hypothesis in the potential source region. The statistic for a background-plus-source hypothesis is calculated by adding a Gaussian source model convolved with the appropriate Chandra point spread function (PSF) and simultaneously fitting the observed photon flux in each observation in the stack. Since a candidate source may be located anywhere in the field of view of each stacked observation, a different PSF must be used for each observation because of the strong spatial dependence of the Chandra PSF. The likelihood of a valid source being detected is a function of the two statistics (for background alone, and for background-plus-source). The MLE tool is an extensible Python module with potential for use by the general Chandra user.
Vestige: Maximum likelihood phylogenetic footprinting
Maxwell Peter
2005-05-01
Full Text Available Abstract Background Phylogenetic footprinting is the identification of functional regions of DNA by their evolutionary conservation. This is achieved by comparing orthologous regions from multiple species and identifying the DNA regions that have diverged less than neutral DNA. Vestige is a phylogenetic footprinting package built on the PyEvolve toolkit that uses probabilistic molecular evolutionary modelling to represent aspects of sequence evolution, including the conventional divergence measure employed by other footprinting approaches. In addition to measuring the divergence, Vestige allows the expansion of the definition of a phylogenetic footprint to include variation in the distribution of any molecular evolutionary processes. This is achieved by displaying the distribution of model parameters that represent partitions of molecular evolutionary substitutions. Examination of the spatial incidence of these effects across regions of the genome can identify DNA segments that differ in the nature of the evolutionary process. Results Vestige was applied to a reference dataset of the SCL locus from four species and provided clear identification of the known conserved regions in this dataset. To demonstrate the flexibility to use diverse models of molecular evolution and dissect the nature of the evolutionary process Vestige was used to footprint the Ka/Ks ratio in primate BRCA1 with a codon model of evolution. Two regions of putative adaptive evolution were identified illustrating the ability of Vestige to represent the spatial distribution of distinct molecular evolutionary processes. Conclusion Vestige provides a flexible, open platform for phylogenetic footprinting. Underpinned by the PyEvolve toolkit, Vestige provides a framework for visualising the signatures of evolutionary processes across the genome of numerous organisms simultaneously. By exploiting the maximum-likelihood statistical framework, the complex interplay between mutational
张彦平
2012-01-01
对步进电机式电子仪表的工作过程进行简要介绍,详细阐述其装针过程设计与控制。%The author briefly introduces the working process of electronic instrument based on stepper motor, elaborates the process design and control of its pointer installation.
Minimum length-maximum velocity
Panes, Boris
2012-03-01
We study a framework where the hypothesis of a minimum length in space-time is complemented with the notion of reference frame invariance. It turns out natural to interpret the action of the obtained reference frame transformations in the context of doubly special relativity. As a consequence of this formalism we find interesting connections between the minimum length properties and the modified velocity-energy relation for ultra-relativistic particles. For example, we can predict the ratio between the minimum lengths in space and time using the results from OPERA on superluminal neutrinos.
A Scalable IP Packet Classification Algorithm Using Indexed Pointers%一种基于索引指针的可扩展IP包分类算法
李金库; 马建峰; 张德运
2012-01-01
设计并实现了一种基于索引指针的可扩展IP包分类算法.该算法通过分析源/目的端口号和协议类型字段在实际应用中的分布特性,将这3个字段映射到一个8比特元组上,压缩了分类维数;算法依据压缩后的8比特元组将分类规则集划分为256个子集,并为每个子集建立一个索引指针,指向该子集的存贮起始地址;算法通过计算IP包中＂源/目的IP地址联合字段＂中各个比特的信息熵值,找出最优的比特序列作为根和子节点,为每个规则子集建立一棵Tries查找树,既保证了存贮空间和查找时间最小,而且不存在回溯问题.实验结果证明,该算法分类效率高.%In this paper,a scalable IP packet classification algorithm using indexed pointers has been proposed.According to the distribution of source port,destination port and protocol type fields in the real applications,the algorithm maps the three fields to an eight-bit value and divides the whole rule set into 256 subsets.It assigns each subset an indexed pointer that points to the starting address of its storage space.The algorithm finds the best bit sequence and uses them as root and child nodes by calculating each bit＇s information entropy value of the combined field of source IP address and destination IP address,then it establishes a Tries lookup tree for each rule subset.By doing so,it requires the least storage space and lookup time without retrospect.The experimental results indicate that the new algorithm is highly efficient.
Thermodynamic hardness and the maximum hardness principle
Franco-Pérez, Marco; Gázquez, José L.; Ayers, Paul W.; Vela, Alberto
2017-08-01
An alternative definition of hardness (called the thermodynamic hardness) within the grand canonical ensemble formalism is proposed in terms of the partial derivative of the electronic chemical potential with respect to the thermodynamic chemical potential of the reservoir, keeping the temperature and the external potential constant. This temperature dependent definition may be interpreted as a measure of the propensity of a system to go through a charge transfer process when it interacts with other species, and thus it keeps the philosophy of the original definition. When the derivative is expressed in terms of the three-state ensemble model, in the regime of low temperatures and up to temperatures of chemical interest, one finds that for zero fractional charge, the thermodynamic hardness is proportional to T-1(I -A ) , where I is the first ionization potential, A is the electron affinity, and T is the temperature. However, the thermodynamic hardness is nearly zero when the fractional charge is different from zero. Thus, through the present definition, one avoids the presence of the Dirac delta function. We show that the chemical hardness defined in this way provides meaningful and discernible information about the hardness properties of a chemical species exhibiting integer or a fractional average number of electrons, and this analysis allowed us to establish a link between the maximum possible value of the hardness here defined, with the minimum softness principle, showing that both principles are related to minimum fractional charge and maximum stability conditions.
Maximum Segment Sum, Monadically (distilled tutorial
Jeremy Gibbons
2011-09-01
Full Text Available The maximum segment sum problem is to compute, given a list of integers, the largest of the sums of the contiguous segments of that list. This problem specification maps directly onto a cubic-time algorithm; however, there is a very elegant linear-time solution too. The problem is a classic exercise in the mathematics of program construction, illustrating important principles such as calculational development, pointfree reasoning, algebraic structure, and datatype-genericity. Here, we take a sideways look at the datatype-generic version of the problem in terms of monadic functional programming, instead of the traditional relational approach; the presentation is tutorial in style, and leavened with exercises for the reader.
Receiver function estimated by maximum entropy deconvolution
吴庆举; 田小波; 张乃铃; 李卫平; 曾融生
2003-01-01
Maximum entropy deconvolution is presented to estimate receiver function, with the maximum entropy as the rule to determine auto-correlation and cross-correlation functions. The Toeplitz equation and Levinson algorithm are used to calculate the iterative formula of error-predicting filter, and receiver function is then estimated. During extrapolation, reflective coefficient is always less than 1, which keeps maximum entropy deconvolution stable. The maximum entropy of the data outside window increases the resolution of receiver function. Both synthetic and real seismograms show that maximum entropy deconvolution is an effective method to measure receiver function in time-domain.
Maximum Power from a Solar Panel
Michael Miller
2010-01-01
Full Text Available Solar energy has become a promising alternative to conventional fossil fuel sources. Solar panels are used to collect solar radiation and convert it into electricity. One of the techniques used to maximize the effectiveness of this energy alternative is to maximize the power output of the solar collector. In this project the maximum power is calculated by determining the voltage and the current of maximum power. These quantities are determined by finding the maximum value for the equation for power using differentiation. After the maximum values are found for each time of day, each individual quantity, voltage of maximum power, current of maximum power, and maximum power is plotted as a function of the time of day.
Research on Dynamic Memory Faults Testing Method Based on Pointer Mapping Sets%基于指针映射集的动态内存故障测试方法研究
张威; 宫云战; 卢庆龄; 万琳
2009-01-01
Dynamic memory faults are ubiquitous in the program with pointers. It is difficult to locate faults sources adopting dynamic testing method. Static analysis methods nowadays often miss some faults and produce too many false alarms. Considering of these problems, this paper puts forward the notion of pointer mapping algebra system that reflects the mapping relationship of pointer and memory completely, and gives the construction rules of pointer mapping sets for different faults class, and then establishes dynamic memory faults model. Through pointer map-ping sets and fault model, it can detect bad deallocation, memory leak and null pointer derefer-ence faults automatically and increase the testing efficiency. In the process of analyzing, it adopts synthetically control flow chart and path condition in order to increase the precision of testing re-sults. Results of experimentation show that this method can detect dynamic memory faults effec-tively. Since rule definition is general, the probability of missing faults and producing false alarms is lower.%动态内存故障在使用指针的程序中是普遍存在的,采用动态测试方法进行测试难以准确定位故障源.而现有的静态分析方法主要存在漏报和误报过多的情况.针对这些问题,提出了指针映射代数系统的概念,全面地反映了指针与内存之间的映射关系,并给出了面向不同故障的指针映射集的构造规则,以此为基础建立了动态内存故障模型.通过指针映射集和故障模型,可以自动检测内存释放异常、内存泄露和空指针引用等动态内存故障,提高了测试效率.在分析过程中,还综合应用了控制流图和路径条件,提高了测试结果的精度.实验结果表明,该方法能够有效检测动态内存故障,而且出于规则定义较为全面,漏报和误报率也较低.
The Pointer Assertion Logic Engine
Møller, Anders; Schwartzbach, Michael Ignatieff
2001-01-01
We present a new framework for verifying partial specifications of programs in order to catch type and memory errors and check data structure invariants. Our technique can verify a large class of data structures, namely all those that can be expressed as graph types. Earlier versions were restric...
The Megalithic Lunar Season Pointer
Clausen, Claus
A data set of 163 Danish megalithic passage tombs including 27 passage graves from Skåne (Sweden), believed to be built within a century, around 3200 BCE., were measured in terms of their location and orientation. The result of this work has provided important new knowledge on the Danish megalithic...... monuments. They have likely served as landscape markers, as there is documented intervisibilitet between most of the graves. The determined distribution of passage directions were the basis for developing a new and more general hypothesis concerning the orientation of megalithic monuments: Directions...... and these showed a statistical similar orientation and distribution. The current conclusion of the thesis based on these results is that the megalithic passage graves are oriented towards the moon, or a moon/ sun relationship throughout the period with megalithic monuments in Europe and adjacent areas...
高阻指针式检流计的设计%Design of the High-resistance Pointer Type Galvanometer
刘秋霞
2012-01-01
在分析传统检流计的基础上,采用恒流源式场效应管差分放大电路作为前级放大路以放大信号,2级运算放大电路作为后级放大电路,多电阻分两级调零,表头串接一电阻并在2端加装1个双向稳压管组成保护电路,传统检流计表头接保护电路完成指示功能,6节5号电池为电路供电,对传统的检流计进行了改进.结果表明,该设计提高了系统的测量精度和灵敏度,而且结构简单、稳定.%In our report, the traditional galvanometer was modified by increasing two cascade amplifying circuits and other circuits, the differential amplifying circuit was used as preamp to amplify signals, at the same time, input resistance was increased, the measured circuit was protected, and the measuring precision was improved. Two cascade op-amp circuits were used as the latter grade amplifier circuit to stabilize the circuit and raise the system sensibility. Two cascade ZA circuits were used to complete the function of ZA. The protecting circuit was made of gauge outfit, resistance and dual stability, the pointer adding the protecting circuit completes the display function, moreover, we used input batteries as power source. The results indicated that circuit is simple and visual display.
Determining Dynamical Path Distributions usingMaximum Relative Entropy
2015-05-31
θ). The selected joint posterior Pnew(x, θ) is that which maximizes the entropy1 , S[P, Pold ] = − ∫ P (x, θ) log P (x, θ) Pold (x, θ) dxdθ , (15) 1...to the appropriate constraints (parameters can be discrete as well). Pold (x, θ) contains our prior information which we call the joint prior. To be...explicit, Pold (x, θ) = Pold (x) Pold (θ|x) , (16) where Pold (x) is the traditional Bayesian prior and Pold (θ|x) is the likelihood. It is important to
The inverse maximum dynamic flow problem
BAGHERIAN; Mehri
2010-01-01
We consider the inverse maximum dynamic flow (IMDF) problem.IMDF problem can be described as: how to change the capacity vector of a dynamic network as little as possible so that a given feasible dynamic flow becomes a maximum dynamic flow.After discussing some characteristics of this problem,it is converted to a constrained minimum dynamic cut problem.Then an efficient algorithm which uses two maximum dynamic flow algorithms is proposed to solve the problem.
Maximum permissible voltage of YBCO coated conductors
Wen, J.; Lin, B.; Sheng, J.; Xu, J.; Jin, Z. [Department of Electrical Engineering, Shanghai Jiao Tong University, Shanghai (China); Hong, Z., E-mail: zhiyong.hong@sjtu.edu.cn [Department of Electrical Engineering, Shanghai Jiao Tong University, Shanghai (China); Wang, D.; Zhou, H.; Shen, X.; Shen, C. [Qingpu Power Supply Company, State Grid Shanghai Municipal Electric Power Company, Shanghai (China)
2014-06-15
Highlights: • We examine three kinds of tapes’ maximum permissible voltage. • We examine the relationship between quenching duration and maximum permissible voltage. • Continuous I{sub c} degradations under repetitive quenching where tapes reaching maximum permissible voltage. • The relationship between maximum permissible voltage and resistance, temperature. - Abstract: Superconducting fault current limiter (SFCL) could reduce short circuit currents in electrical power system. One of the most important thing in developing SFCL is to find out the maximum permissible voltage of each limiting element. The maximum permissible voltage is defined as the maximum voltage per unit length at which the YBCO coated conductors (CC) do not suffer from critical current (I{sub c}) degradation or burnout. In this research, the time of quenching process is changed and voltage is raised until the I{sub c} degradation or burnout happens. YBCO coated conductors test in the experiment are from American superconductor (AMSC) and Shanghai Jiao Tong University (SJTU). Along with the quenching duration increasing, the maximum permissible voltage of CC decreases. When quenching duration is 100 ms, the maximum permissible of SJTU CC, 12 mm AMSC CC and 4 mm AMSC CC are 0.72 V/cm, 0.52 V/cm and 1.2 V/cm respectively. Based on the results of samples, the whole length of CCs used in the design of a SFCL can be determined.
熊伟丽; 周寰; 徐保国
2009-01-01
A fully automatic calibration system for pressure gauges with pointers is presented, including a calibration management system of pressure gauges and an automatic reading system of pointer readings. The operation is very convenient with the use of visual interface in this system. The operator only needs to input several necessary coefficients such as range, accuracy and so on, the pointer reading of pressure gauges will be read automatically. The automatic reading method will be used to calibrate for almost all kinds of pointer gauges. The accuracy is decided by resolution and processing time. This automatic calibration system improves the present state of pressure calibration, and enhances the automation level of calibration to some extent.%设计并开发了一种指针式压力表自动检定系统.该系统包括压力表检定/校准管理系统和仪表示值自动判读系统两部分.系统的可视化界面使得操作非常简单,操作人员只需输入量程、精度等必要参数,压力表的指针读数即可自动读出.给出的自动判读方法可用于各种指针式仪表精度的自动检测、判读,其精度取决于图像的分辨率和限定的处理时间.本系统可较大地改善压力表人工检定工作的落后状况,在一定程度上提高了检定工作的自动化水平和检定结果的稳定性.
Bartlett, Roscoe A
2004-06-01
Dynamic memory management in C++ is one of the most common areas of difficulty and errors for amateur and expert C++ developers alike. The improper use of operator new and operator delete is arguably the most common cause of incorrect program behavior and segmentation faults in C++ programs. Here we introduce a templated concrete C++ class Teuchos::RefCountPtr<>, which is part of the Trilinos tools package Teuchos, that combines the concepts of smart pointers and reference counting to build a low-overhead but effective tool for simplifying dynamic memory management in C++. We discuss why the use of raw pointers for memory management, managed through explicit calls to operator new and operator delete, is so difficult to accomplish without making mistakes and how programs that use raw pointers for memory management can easily be modified to use RefCountPtr<>. In addition, explicit calls to operator delete is fragile and results in memory leaks in the presents of C++ exceptions. In its most basic usage, RefCountPtr<> automatically determines when operator delete should be called to free an object allocated with operator new and is not fragile in the presents of exceptions. The class also supports more sophisticated use cases as well. This document describes just the most basic usage of RefCountPtr<> to allow developers to get started using it right away. However, more detailed information on the design and advanced features of RefCountPtr<> is provided by the companion document 'Teuchos::RefCountPtr : The Trilinos Smart Reference-Counted Pointer Class for (Almost) Automatic Dynamic Memory Management in C++'.
刘迪; 毕笃彦; 李权合; 刘楠
2013-01-01
拟合人眼视觉机制提出了非均匀光照下指针仪表图像的预处理算法，在此基础上进一步设计了座舱指针式仪表自动判读算法。首先对仪表盘图像进行亮度均衡、二值化变换，再将指针细化，然后根据改进的Hough变换提取目标信息，确定座舱指针式仪表的读数。实验结果表明，该算法有效地解决了在非均匀光照情况下的飞机座舱指针式仪表自动判读，降低了判读误差。%the human visual mechanism , based on that we further design the automatic interpretation algorithm for cockpit pointer instruments .Firstly the algorithm conducts the lightness balance and binarisation transform on instruments panel images , and then makes the pointer needles thinning down;afterwards , it extracts the object information according to the modified Hough transformation and determines the readings of cockpit index instruments.Experimental results prove that the algorithm effectively solves the issue of automatic interpretation of cockpit pointer instruments in circumstance of non-uniform lightness , reduces the reading error .
Nielsen, Søren R. K.; Köyüoglu, H. U.; Cakmak, A. S.
The maximum softening concept is based on the variation of the vibrational periods of a structure during a seismic event. Maximum softening damage indicators, which measure the maximum relative stiffness reduction caused by stiffness and strength deterioration of the actual structure, are calcula......The maximum softening concept is based on the variation of the vibrational periods of a structure during a seismic event. Maximum softening damage indicators, which measure the maximum relative stiffness reduction caused by stiffness and strength deterioration of the actual structure...
Generalised maximum entropy and heterogeneous technologies
Oude Lansink, A.G.J.M.
1999-01-01
Generalised maximum entropy methods are used to estimate a dual model of production on panel data of Dutch cash crop farms over the period 1970-1992. The generalised maximum entropy approach allows a coherent system of input demand and output supply equations to be estimated for each farm in the sam
20 CFR 229.48 - Family maximum.
2010-04-01
... month on one person's earnings record is limited. This limited amount is called the family maximum. The family maximum used to adjust the social security overall minimum rate is based on the employee's Overall..., when any of the persons entitled to benefits on the insured individual's compensation would, except...
The maximum rotation of a galactic disc
Bottema, R
1997-01-01
The observed stellar velocity dispersions of galactic discs show that the maximum rotation of a disc is on average 63% of the observed maximum rotation. This criterion can, however, not be applied to small or low surface brightness (LSB) galaxies because such systems show, in general, a continuously
Duality of Maximum Entropy and Minimum Divergence
Shinto Eguchi
2014-06-01
Full Text Available We discuss a special class of generalized divergence measures by the use of generator functions. Any divergence measure in the class is separated into the difference between cross and diagonal entropy. The diagonal entropy measure in the class associates with a model of maximum entropy distributions; the divergence measure leads to statistical estimation via minimization, for arbitrarily giving a statistical model. The dualistic relationship between the maximum entropy model and the minimum divergence estimation is explored in the framework of information geometry. The model of maximum entropy distributions is characterized to be totally geodesic with respect to the linear connection associated with the divergence. A natural extension for the classical theory for the maximum likelihood method under the maximum entropy model in terms of the Boltzmann-Gibbs-Shannon entropy is given. We discuss the duality in detail for Tsallis entropy as a typical example.
Danilo F. Pereira
2005-08-01
Full Text Available Dada a importância da avicultura de corte para a economia brasileira, este trabalho buscou compreender e identificar novos indicadores de bem-estar inerentes ao animal que contribuíssem para o aumento da eficácia produtiva, estudando-se diferentes reações comportamentais de matrizes de frango de corte, em câmara climática. O experimento foi delineado como um Quadrado Latino 3x3x3, em que as variáveis temperatura do ar, ração e idade das aves foram controladas, sendo as aves de diferentes idades alojadas em boxes distintos. Foram feitas observações do comportamento das aves em dois horários do dia, sendo uma observação na parte da manhã e outra na parte da tarde, durante períodos de 15 minutos cada, por meio de câmeras de vídeo instaladas no teto da câmara climática, não havendo interferência humana no registro das informações. Constatou-se, nos dados, a influência das variáveis controladas em diversos comportamentos observados, concluindo-se que a presença do alimento é um fator experimental relevante, pois aumenta as reações comportamentais agressivas.Knowing the importance that the poultry industry represents for the Brazilian economy, this work, searched to understand and to identify new welfare pointers inherent to the animal that contributed for the increase of the productive effectiveness, studying different behavior reactions in broiler breeders, in climatic chamber. The experiment was delineated as a Latin Square 3x3x3, where the variable: temperature of air, birds ration and birds age had been controlled. The birds of different ages had been lodged in distinct boxes. Observations of the behavior of the birds in two schedules of the day had been made, being one in the morning and the other one in the afternoon, during a period of 15 minutes each through video cameras, installed in the ceiling of the climatic chamber, having no interference of human being in the register of the data. It was verified the
Berry, Vincent; Nicolas, François
2006-01-01
Given a set of evolutionary trees on a same set of taxa, the maximum agreement subtree problem (MAST), respectively, maximum compatible tree problem (MCT), consists of finding a largest subset of taxa such that all input trees restricted to these taxa are isomorphic, respectively compatible. These problems have several applications in phylogenetics such as the computation of a consensus of phylogenies obtained from different data sets, the identification of species subjected to horizontal gene transfers and, more recently, the inference of supertrees, e.g., Trees Of Life. We provide two linear time algorithms to check the isomorphism, respectively, compatibility, of a set of trees or otherwise identify a conflict between the trees with respect to the relative location of a small subset of taxa. Then, we use these algorithms as subroutines to solve MAST and MCT on rooted or unrooted trees of unbounded degree. More precisely, we give exact fixed-parameter tractable algorithms, whose running time is uniformly polynomial when the number of taxa on which the trees disagree is bounded. The improves on a known result for MAST and proves fixed-parameter tractability for MCT.
Individual Module Maximum Power Point Tracking for Thermoelectric Generator Systems
Vadstrup, Casper; Schaltz, Erik; Chen, Min
2013-07-01
In a thermoelectric generator (TEG) system the DC/DC converter is under the control of a maximum power point tracker which ensures that the TEG system outputs the maximum possible power to the load. However, if the conditions, e.g., temperature, health, etc., of the TEG modules are different, each TEG module will not produce its maximum power. If each TEG module is controlled individually, each TEG module can be operated at its maximum power point and the TEG system output power will therefore be higher. In this work a power converter based on noninverting buck-boost converters capable of handling four TEG modules is presented. It is shown that, when each module in the TEG system is operated under individual maximum power point tracking, the system output power for this specific application can be increased by up to 8.4% relative to the situation when the modules are connected in series and 16.7% relative to the situation when the modules are connected in parallel.
A dual method for maximum entropy restoration
Smith, C. B.
1979-01-01
A simple iterative dual algorithm for maximum entropy image restoration is presented. The dual algorithm involves fewer parameters than conventional minimization in the image space. Minicomputer test results for Fourier synthesis with inadequate phantom data are given.
Maximum Throughput in Multiple-Antenna Systems
Zamani, Mahdi
2012-01-01
The point-to-point multiple-antenna channel is investigated in uncorrelated block fading environment with Rayleigh distribution. The maximum throughput and maximum expected-rate of this channel are derived under the assumption that the transmitter is oblivious to the channel state information (CSI), however, the receiver has perfect CSI. First, we prove that in multiple-input single-output (MISO) channels, the optimum transmission strategy maximizing the throughput is to use all available antennas and perform equal power allocation with uncorrelated signals. Furthermore, to increase the expected-rate, multi-layer coding is applied. Analogously, we establish that sending uncorrelated signals and performing equal power allocation across all available antennas at each layer is optimum. A closed form expression for the maximum continuous-layer expected-rate of MISO channels is also obtained. Moreover, we investigate multiple-input multiple-output (MIMO) channels, and formulate the maximum throughput in the asympt...
Photoemission spectromicroscopy with MAXIMUM at Wisconsin
Ng, W.; Ray-Chaudhuri, A.K.; Cole, R.K.; Wallace, J.; Crossley, S.; Crossley, D.; Chen, G.; Green, M.; Guo, J.; Hansen, R.W.C.; Cerrina, F.; Margaritondo, G. (Dept. of Electrical Engineering, Dept. of Physics and Synchrotron Radiation Center, Univ. of Wisconsin, Madison (USA)); Underwood, J.H.; Korthright, J.; Perera, R.C.C. (Center for X-ray Optics, Accelerator and Fusion Research Div., Lawrence Berkeley Lab., CA (USA))
1990-06-01
We describe the development of the scanning photoemission spectromicroscope MAXIMUM at the Wisoncsin Synchrotron Radiation Center, which uses radiation from a 30-period undulator. The article includes a discussion of the first tests after the initial commissioning. (orig.).
Maximum-likelihood method in quantum estimation
Paris, M G A; Sacchi, M F
2001-01-01
The maximum-likelihood method for quantum estimation is reviewed and applied to the reconstruction of density matrix of spin and radiation as well as to the determination of several parameters of interest in quantum optics.
The constraint rule of the maximum entropy principle
Uffink, J.
2001-01-01
The principle of maximum entropy is a method for assigning values to probability distributions on the basis of partial information. In usual formulations of this and related methods of inference one assumes that this partial information takes the form of a constraint on allowed probability distribut
MAXIMUM-LIKELIHOOD-ESTIMATION OF THE ENTROPY OF AN ATTRACTOR
SCHOUTEN, JC; TAKENS, F; VANDENBLEEK, CM
1994-01-01
In this paper, a maximum-likelihood estimate of the (Kolmogorov) entropy of an attractor is proposed that can be obtained directly from a time series. Also, the relative standard deviation of the entropy estimate is derived; it is dependent on the entropy and on the number of samples used in the est
Maximum super angle optimization method for array antenna pattern synthesis
Wu, Ji; Roederer, A. G
1991-01-01
Different optimization criteria related to antenna pattern synthesis are discussed. Based on the maximum criteria and vector space representation, a simple and efficient optimization method is presented for array and array fed reflector power pattern synthesis. A sector pattern synthesized by a 20...
The maximum entropy technique. System's statistical description
Belashev, B Z
2002-01-01
The maximum entropy technique (MENT) is applied for searching the distribution functions of physical values. MENT takes into consideration the demand of maximum entropy, the characteristics of the system and the connection conditions, naturally. It is allowed to apply MENT for statistical description of closed and open systems. The examples in which MENT had been used for the description of the equilibrium and nonequilibrium states and the states far from the thermodynamical equilibrium are considered
19 CFR 114.23 - Maximum period.
2010-04-01
... 19 Customs Duties 1 2010-04-01 2010-04-01 false Maximum period. 114.23 Section 114.23 Customs... CARNETS Processing of Carnets § 114.23 Maximum period. (a) A.T.A. carnet. No A.T.A. carnet with a period of validity exceeding 1 year from date of issue shall be accepted. This period of validity cannot be...
Maximum-Likelihood Detection Of Noncoherent CPM
Divsalar, Dariush; Simon, Marvin K.
1993-01-01
Simplified detectors proposed for use in maximum-likelihood-sequence detection of symbols in alphabet of size M transmitted by uncoded, full-response continuous phase modulation over radio channel with additive white Gaussian noise. Structures of receivers derived from particular interpretation of maximum-likelihood metrics. Receivers include front ends, structures of which depends only on M, analogous to those in receivers of coherent CPM. Parts of receivers following front ends have structures, complexity of which would depend on N.
SEXUAL DIMORPHISM OF MAXIMUM FEMORAL LENGTH
Pandya A M
2011-04-01
Full Text Available Sexual identification from the skeletal parts has medico legal and anthropological importance. Present study aims to obtain values of maximum femoral length and to evaluate its possible usefulness in determining correct sexual identification. Study sample consisted of 184 dry, normal, adult, human femora (136 male & 48 female from skeletal collections of Anatomy department, M. P. Shah Medical College, Jamnagar, Gujarat. Maximum length of femur was considered as maximum vertical distance between upper end of head of femur and the lowest point on femoral condyle, measured with the osteometric board. Mean Values obtained were, 451.81 and 417.48 for right male and female, and 453.35 and 420.44 for left male and female respectively. Higher value in male was statistically highly significant (P< 0.001 on both sides. Demarking point (D.P. analysis of the data showed that right femora with maximum length more than 476.70 were definitely male and less than 379.99 were definitely female; while for left bones, femora with maximum length more than 484.49 were definitely male and less than 385.73 were definitely female. Maximum length identified 13.43% of right male femora, 4.35% of right female femora, 7.25% of left male femora and 8% of left female femora. [National J of Med Res 2011; 1(2.000: 67-70
Maximum efficiency of low-dissipation heat engines at arbitrary power
Holubec, Viktor; Ryabov, Artem
2016-07-01
We investigate maximum efficiency at a given power for low-dissipation heat engines. Close to maximum power, the maximum gain in efficiency scales as a square root of relative loss in power and this scaling is universal for a broad class of systems. For low-dissipation engines, we calculate the maximum gain in efficiency for an arbitrary fixed power. We show that engines working close to maximum power can operate at considerably larger efficiency compared to the efficiency at maximum power. Furthermore, we introduce universal bounds on maximum efficiency at a given power for low-dissipation heat engines. These bounds represent direct generalization of the bounds on efficiency at maximum power obtained by Esposito et al (2010 Phys. Rev. Lett. 105 150603). We derive the bounds analytically in the regime close to maximum power and for small power values. For the intermediate regime we present strong numerical evidence for the validity of the bounds.
Present and Last Glacial Maximum climates as states of maximum entropy production
Herbert, Corentin; Kageyama, Masa; Dubrulle, Berengere
2011-01-01
The Earth, like other planets with a relatively thick atmosphere, is not locally in radiative equilibrium and the transport of energy by the geophysical fluids (atmosphere and ocean) plays a fundamental role in determining its climate. Using simple energy-balance models, it was suggested a few decades ago that the meridional energy fluxes might follow a thermodynamic Maximum Entropy Production (MEP) principle. In the present study, we assess the MEP hypothesis in the framework of a minimal climate model based solely on a robust radiative scheme and the MEP principle, with no extra assumptions. Specifically, we show that by choosing an adequate radiative exchange formulation, the Net Exchange Formulation, a rigorous derivation of all the physical parameters can be performed. The MEP principle is also extended to surface energy fluxes, in addition to meridional energy fluxes. The climate model presented here is extremely fast, needs very little empirical data and does not rely on ad hoc parameterizations. We in...
Brewster, Hilary D
2009-01-01
The theory of relativity has become a cornerstone of modern physics. Over the course of time it has been scrutinized in a multitude of experiments and has always been verified with high accuracy. The correctness of this theory can no longer be called into question. Right after its discovery by Albert Einstein in 1905, special relativity was only gradually accepted because it made numerous predictions contradicting common sense, fervently castigated by Einstein, and also defied experiment for too long a time. It was only with the advent of particle or high energy physics that matter could be ac
Einstein, Albert
2013-01-01
Time magazine's ""Man of the Century"", Albert Einstein is the founder of modern physics and his theory of relativity is the most important scientific idea of the modern era. In this short book, Einstein explains, using the minimum of mathematical terms, the basic ideas and principles of the theory that has shaped the world we live in today. Unsurpassed by any subsequent books on relativity, this remains the most popular and useful exposition of Einstein's immense contribution to human knowledge.With a new foreword by Derek Raine.
Pattern formation, logistics, and maximum path probability
Kirkaldy, J. S.
1985-05-01
The concept of pattern formation, which to current researchers is a synonym for self-organization, carries the connotation of deductive logic together with the process of spontaneous inference. Defining a pattern as an equivalence relation on a set of thermodynamic objects, we establish that a large class of irreversible pattern-forming systems, evolving along idealized quasisteady paths, approaches the stable steady state as a mapping upon the formal deductive imperatives of a propositional function calculus. In the preamble the classical reversible thermodynamics of composite systems is analyzed as an externally manipulated system of space partitioning and classification based on ideal enclosures and diaphragms. The diaphragms have discrete classification capabilities which are designated in relation to conserved quantities by descriptors such as impervious, diathermal, and adiabatic. Differentiability in the continuum thermodynamic calculus is invoked as equivalent to analyticity and consistency in the underlying class or sentential calculus. The seat of inference, however, rests with the thermodynamicist. In the transition to an irreversible pattern-forming system the defined nature of the composite reservoirs remains, but a given diaphragm is replaced by a pattern-forming system which by its nature is a spontaneously evolving volume partitioner and classifier of invariants. The seat of volition or inference for the classification system is thus transferred from the experimenter or theoretician to the diaphragm, and with it the full deductive facility. The equivalence relations or partitions associated with the emerging patterns may thus be associated with theorems of the natural pattern-forming calculus. The entropy function, together with its derivatives, is the vehicle which relates the logistics of reservoirs and diaphragms to the analog logistics of the continuum. Maximum path probability or second-order differentiability of the entropy in isolation are
Proscriptive Bayesian Programming and Maximum Entropy: a Preliminary Study
Koike, Carla Cavalcante
2008-11-01
Some problems found in robotics systems, as avoiding obstacles, can be better described using proscriptive commands, where only prohibited actions are indicated in contrast to prescriptive situations, which demands that a specific command be specified. An interesting question arises regarding the possibility to learn automatically if proscriptive commands are suitable and which parametric function could be better applied. Lately, a great variety of problems in robotics domain are object of researches using probabilistic methods, including the use of Maximum Entropy in automatic learning for robot control systems. This works presents a preliminary study on automatic learning of proscriptive robot control using maximum entropy and using Bayesian Programming. It is verified whether Maximum entropy and related methods can favour proscriptive commands in an obstacle avoidance task executed by a mobile robot.
Quality, precision and accuracy of the maximum No. 40 anemometer
Obermeir, J. [Otech Engineering, Davis, CA (United States); Blittersdorf, D. [NRG Systems Inc., Hinesburg, VT (United States)
1996-12-31
This paper synthesizes available calibration data for the Maximum No. 40 anemometer. Despite its long history in the wind industry, controversy surrounds the choice of transfer function for this anemometer. Many users are unaware that recent changes in default transfer functions in data loggers are producing output wind speed differences as large as 7.6%. Comparison of two calibration methods used for large samples of Maximum No. 40 anemometers shows a consistent difference of 4.6% in output speeds. This difference is significantly larger than estimated uncertainty levels. Testing, initially performed to investigate related issues, reveals that Gill and Maximum cup anemometers change their calibration transfer functions significantly when calibrated in the open atmosphere compared with calibration in a laminar wind tunnel. This indicates that atmospheric turbulence changes the calibration transfer function of cup anemometers. These results call into question the suitability of standard wind tunnel calibration testing for cup anemometers. 6 refs., 10 figs., 4 tabs.
Maximum permissible voltage of YBCO coated conductors
Wen, J.; Lin, B.; Sheng, J.; Xu, J.; Jin, Z.; Hong, Z.; Wang, D.; Zhou, H.; Shen, X.; Shen, C.
2014-06-01
Superconducting fault current limiter (SFCL) could reduce short circuit currents in electrical power system. One of the most important thing in developing SFCL is to find out the maximum permissible voltage of each limiting element. The maximum permissible voltage is defined as the maximum voltage per unit length at which the YBCO coated conductors (CC) do not suffer from critical current (Ic) degradation or burnout. In this research, the time of quenching process is changed and voltage is raised until the Ic degradation or burnout happens. YBCO coated conductors test in the experiment are from American superconductor (AMSC) and Shanghai Jiao Tong University (SJTU). Along with the quenching duration increasing, the maximum permissible voltage of CC decreases. When quenching duration is 100 ms, the maximum permissible of SJTU CC, 12 mm AMSC CC and 4 mm AMSC CC are 0.72 V/cm, 0.52 V/cm and 1.2 V/cm respectively. Based on the results of samples, the whole length of CCs used in the design of a SFCL can be determined.
Computing Rooted and Unrooted Maximum Consistent Supertrees
van Iersel, Leo
2009-01-01
A chief problem in phylogenetics and database theory is the computation of a maximum consistent tree from a set of rooted or unrooted trees. A standard input are triplets, rooted binary trees on three leaves, or quartets, unrooted binary trees on four leaves. We give exact algorithms constructing rooted and unrooted maximum consistent supertrees in time O(2^n n^5 m^2 log(m)) for a set of m triplets (quartets), each one distinctly leaf-labeled by some subset of n labels. The algorithms extend to weighted triplets (quartets). We further present fast exact algorithms for constructing rooted and unrooted maximum consistent trees in polynomial space. Finally, for a set T of m rooted or unrooted trees with maximum degree D and distinctly leaf-labeled by some subset of a set L of n labels, we compute, in O(2^{mD} n^m m^5 n^6 log(m)) time, a tree distinctly leaf-labeled by a maximum-size subset X of L that all trees in T, when restricted to X, are consistent with.
Maximum magnitude earthquakes induced by fluid injection
McGarr, Arthur F.
2014-01-01
Analysis of numerous case histories of earthquake sequences induced by fluid injection at depth reveals that the maximum magnitude appears to be limited according to the total volume of fluid injected. Similarly, the maximum seismic moment seems to have an upper bound proportional to the total volume of injected fluid. Activities involving fluid injection include (1) hydraulic fracturing of shale formations or coal seams to extract gas and oil, (2) disposal of wastewater from these gas and oil activities by injection into deep aquifers, and (3) the development of enhanced geothermal systems by injecting water into hot, low-permeability rock. Of these three operations, wastewater disposal is observed to be associated with the largest earthquakes, with maximum magnitudes sometimes exceeding 5. To estimate the maximum earthquake that could be induced by a given fluid injection project, the rock mass is assumed to be fully saturated, brittle, to respond to injection with a sequence of earthquakes localized to the region weakened by the pore pressure increase of the injection operation and to have a Gutenberg-Richter magnitude distribution with a b value of 1. If these assumptions correctly describe the circumstances of the largest earthquake, then the maximum seismic moment is limited to the volume of injected liquid times the modulus of rigidity. Observations from the available case histories of earthquakes induced by fluid injection are consistent with this bound on seismic moment. In view of the uncertainties in this analysis, however, this should not be regarded as an absolute physical limit.
Maximum magnitude earthquakes induced by fluid injection
McGarr, A.
2014-02-01
Analysis of numerous case histories of earthquake sequences induced by fluid injection at depth reveals that the maximum magnitude appears to be limited according to the total volume of fluid injected. Similarly, the maximum seismic moment seems to have an upper bound proportional to the total volume of injected fluid. Activities involving fluid injection include (1) hydraulic fracturing of shale formations or coal seams to extract gas and oil, (2) disposal of wastewater from these gas and oil activities by injection into deep aquifers, and (3) the development of enhanced geothermal systems by injecting water into hot, low-permeability rock. Of these three operations, wastewater disposal is observed to be associated with the largest earthquakes, with maximum magnitudes sometimes exceeding 5. To estimate the maximum earthquake that could be induced by a given fluid injection project, the rock mass is assumed to be fully saturated, brittle, to respond to injection with a sequence of earthquakes localized to the region weakened by the pore pressure increase of the injection operation and to have a Gutenberg-Richter magnitude distribution with a b value of 1. If these assumptions correctly describe the circumstances of the largest earthquake, then the maximum seismic moment is limited to the volume of injected liquid times the modulus of rigidity. Observations from the available case histories of earthquakes induced by fluid injection are consistent with this bound on seismic moment. In view of the uncertainties in this analysis, however, this should not be regarded as an absolute physical limit.
Predicting species' maximum dispersal distances from simple plant traits.
Tamme, Riin; Götzenberger, Lars; Zobel, Martin; Bullock, James M; Hooftman, Danny A P; Kaasik, Ants; Pärtel, Meelis
2014-02-01
Many studies have shown plant species' dispersal distances to be strongly related to life-history traits, but how well different traits can predict dispersal distances is not yet known. We used cross-validation techniques and a global data set (576 plant species) to measure the predictive power of simple plant traits to estimate species' maximum dispersal distances. Including dispersal syndrome (wind, animal, ant, ballistic, and no special syndrome), growth form (tree, shrub, herb), seed mass, seed release height, and terminal velocity in different combinations as explanatory variables we constructed models to explain variation in measured maximum dispersal distances and evaluated their power to predict maximum dispersal distances. Predictions are more accurate, but also limited to a particular set of species, if data on more specific traits, such as terminal velocity, are available. The best model (R2 = 0.60) included dispersal syndrome, growth form, and terminal velocity as fixed effects. Reasonable predictions of maximum dispersal distance (R2 = 0.53) are also possible when using only the simplest and most commonly measured traits; dispersal syndrome and growth form together with species taxonomy data. We provide a function (dispeRsal) to be run in the software package R. This enables researchers to estimate maximum dispersal distances with confidence intervals for plant species using measured traits as predictors. Easily obtainable trait data, such as dispersal syndrome (inferred from seed morphology) and growth form, enable predictions to be made for a large number of species.
王瑞; 李琦; 方彦军
2013-01-01
利用计算机进行指针式仪表的检定技术日趋成熟，其中仪表数值的判读算法还有一定缺陷，对读数精度造成影响。针对基于机器视觉的指针式仪表检定对高精度的要求，着重研究了计算机自动读数过程中的读数算法环节，提出了基于角度法和距离法的改进角度法，该方法结合了角度法和距离法的优点，进一步提高了基于机器视觉的指针式仪表自动读数的精度。%The verification technology of pointer instruments, which uses computers, has become ripe. The interpretation arithmetic of instruments still has some drawbacks, which causes bad effects to the accuracy. This paper focused on the demand for the high accuracy of pointer instrument′s verification based on computer vision. It studies the reading arithmetic of automatic reding, and puts forwards the improved angle method which is based on the angle method and distance method. The improved angel method includes the advantages of angle method and distance method, which improve the accuracy of pointer instrument′s automatic reading based on computer vision.
Chen, Jincan; Yan, Zijun; Wu, Liqing
1996-06-01
Considering a thermoelectric generator as a heat engine cycle, the general differential equations of the temperature field inside thermoelectric elements are established by means of nonequilibrium thermodynamics. These equations are used to study the influence of heat leak, Joule's heat, and Thomson heat on the performance of the thermoelectric generator. New expressions are derived for the power output and the efficiency of the thermoelectric generator. The maximum power output is calculated and the optimal matching condition of load is determined. The maximum efficiency is discussed by a representative numerical example. The aim of this research is to provide some novel conclusions and redress some errors existing in a related investigation.
康凯; 郭伟; 吴诗其
2002-01-01
Mobility management is one of the key problems in the mobile communication techniques. The performance of the mobile communication networks is heavily depended on the effectiveness of the mobility management. In this paper, we have presented a protocol of the mobility management in the mobile IP networks, based on the distributed RSs. When the mobile user in the inactive connection handoffs, some pointers are constructed and modified by the RSs and routers. It can realize the local registration; only when a call arrives, the possible update operations of the mobile database are needed. Consequently, the access of the RS is limited, and the cost of the mobility management is also reduced.
The Wiener maximum quadratic assignment problem
Cela, Eranda; Woeginger, Gerhard J
2011-01-01
We investigate a special case of the maximum quadratic assignment problem where one matrix is a product matrix and the other matrix is the distance matrix of a one-dimensional point set. We show that this special case, which we call the Wiener maximum quadratic assignment problem, is NP-hard in the ordinary sense and solvable in pseudo-polynomial time. Our approach also yields a polynomial time solution for the following problem from chemical graph theory: Find a tree that maximizes the Wiener index among all trees with a prescribed degree sequence. This settles an open problem from the literature.
Maximum confidence measurements via probabilistic quantum cloning
Zhang Wen-Hai; Yu Long-Bao; Cao Zhuo-Liang; Ye Liu
2013-01-01
Probabilistic quantum cloning (PQC) cannot copy a set of linearly dependent quantum states.In this paper,we show that if incorrect copies are allowed to be produced,linearly dependent quantum states may also be cloned by the PQC.By exploiting this kind of PQC to clone a special set of three linearly dependent quantum states,we derive the upper bound of the maximum confidence measure of a set.An explicit transformation of the maximum confidence measure is presented.
Maximum floodflows in the conterminous United States
Crippen, John R.; Bue, Conrad D.
1977-01-01
Peak floodflows from thousands of observation sites within the conterminous United States were studied to provide a guide for estimating potential maximum floodflows. Data were selected from 883 sites with drainage areas of less than 10,000 square miles (25,900 square kilometers) and were grouped into regional sets. Outstanding floods for each region were plotted on graphs, and envelope curves were computed that offer reasonable limits for estimates of maximum floods. The curves indicate that floods may occur that are two to three times greater than those known for most streams.
The Maximum Resource Bin Packing Problem
Boyar, J.; Epstein, L.; Favrholdt, L.M.
2006-01-01
Usually, for bin packing problems, we try to minimize the number of bins used or in the case of the dual bin packing problem, maximize the number or total size of accepted items. This paper presents results for the opposite problems, where we would like to maximize the number of bins used...... algorithms, First-Fit-Increasing and First-Fit-Decreasing for the maximum resource variant of classical bin packing. For the on-line variant, we define maximum resource variants of classical and dual bin packing. For dual bin packing, no on-line algorithm is competitive. For classical bin packing, we find...
Maximum entropy analysis of EGRET data
Pohl, M.; Strong, A.W.
1997-01-01
EGRET data are usually analysed on the basis of the Maximum-Likelihood method \\cite{ma96} in a search for point sources in excess to a model for the background radiation (e.g. \\cite{hu97}). This method depends strongly on the quality of the background model, and thus may have high systematic unce...... uncertainties in region of strong and uncertain background like the Galactic Center region. Here we show images of such regions obtained by the quantified Maximum-Entropy method. We also discuss a possible further use of MEM in the analysis of problematic regions of the sky....
Maximum phytoplankton concentrations in the sea
Jackson, G.A.; Kiørboe, Thomas
2008-01-01
A simplification of plankton dynamics using coagulation theory provides predictions of the maximum algal concentration sustainable in aquatic systems. These predictions have previously been tested successfully against results from iron fertilization experiments. We extend the test to data collected...... in the North Atlantic as part of the Bermuda Atlantic Time Series program as well as data collected off Southern California as part of the Southern California Bight Study program. The observed maximum particulate organic carbon and volumetric particle concentrations are consistent with the predictions...
赵永翔; 王金诺; 高庆
2001-01-01
拓展经典极大似然法到Langer模型，提出了估计三参数、Langer和Basquin三种常用疲劳应力-寿命模型P-S-N曲线及其置信限的统一方法。方法用于处理极大似然法疲劳试验得到的S-N数据。该试验在特别关注的参考载荷试验一组试样，其余试样在不同载荷下试验。以参考载荷试验数据的统计参量为基础，按照每个模型中材料常数协同处于相同概率水平原则，将曲线表示为对数疲劳寿命均值和均方差线的广义形式，至多4个材料常数。曲线中的材料常数按极大似然原理采用数学规划法求出。45＃碳钢缺口试样(kt＝20）对称循环加载试验数据的分析说明了方法的有效性。分析还揭示合理模型有必要通过比较拟合效果、预计误差和应用安全性来确定。三参数模型的拟合效果最好，Langer模型稍差，Basquin模型较差。从拟合效果、预计误差和应用安全性角度，Basquin模型不适于描述该套数据。此外，经典极大似然法估计结果可能因受局部统计参量影响而给出非安全估计，有必要发展改进的可以最大限度减小这种影响的方法。%A unified classical maximum likelihood approach for estimating P-S-N curves of the three commonly used fatigue stress-life relations, namely three parameter, Langer and Basquin, is presented by extrapolating the classical maximum likelihood method to the Langer relation. This approach is applied to deal with the S-N data obtained from a so-called maximum likelihood method-fatigue test. In the test, a group of specimens are tested at a so-called reference load, which is specially taken care of by practice, and residual specimens are individually fatigued at different loads. The approach takes a basis of the local statistical parameters of the logarithms of fatigue lives at the reference load. According to an assumption that the material constants in each relation are concurrently in
Analysis of Photovoltaic Maximum Power Point Trackers
Veerachary, Mummadi
The photovoltaic generator exhibits a non-linear i-v characteristic and its maximum power point (MPP) varies with solar insolation. An intermediate switch-mode dc-dc converter is required to extract maximum power from the photovoltaic array. In this paper buck, boost and buck-boost topologies are considered and a detailed mathematical analysis, both for continuous and discontinuous inductor current operation, is given for MPP operation. The conditions on the connected load values and duty ratio are derived for achieving the satisfactory maximum power point operation. Further, it is shown that certain load values, falling out of the optimal range, will drive the operating point away from the true maximum power point. Detailed comparison of various topologies for MPPT is given. Selection of the converter topology for a given loading is discussed. Detailed discussion on circuit-oriented model development is given and then MPPT effectiveness of various converter systems is verified through simulations. Proposed theory and analysis is validated through experimental investigations.
On maximum cycle packings in polyhedral graphs
Peter Recht
2014-04-01
Full Text Available This paper addresses upper and lower bounds for the cardinality of a maximum vertex-/edge-disjoint cycle packing in a polyhedral graph G. Bounds on the cardinality of such packings are provided, that depend on the size, the order or the number of faces of G, respectively. Polyhedral graphs are constructed, that attain these bounds.
Hard graphs for the maximum clique problem
Hoede, Cornelis
1988-01-01
The maximum clique problem is one of the NP-complete problems. There are graphs for which a reduction technique exists that transforms the problem for these graphs into one for graphs with specific properties in polynomial time. The resulting graphs do not grow exponentially in order and number. Gra
Maximum Likelihood Estimation of Search Costs
J.L. Moraga-Gonzalez (José Luis); M.R. Wildenbeest (Matthijs)
2006-01-01
textabstractIn a recent paper Hong and Shum (forthcoming) present a structural methodology to estimate search cost distributions. We extend their approach to the case of oligopoly and present a maximum likelihood estimate of the search cost distribution. We apply our method to a data set of online p
Weak Scale From the Maximum Entropy Principle
Hamada, Yuta; Kawana, Kiyoharu
2015-01-01
The theory of multiverse and wormholes suggests that the parameters of the Standard Model are fixed in such a way that the radiation of the $S^{3}$ universe at the final stage $S_{rad}$ becomes maximum, which we call the maximum entropy principle. Although it is difficult to confirm this principle generally, for a few parameters of the Standard Model, we can check whether $S_{rad}$ actually becomes maximum at the observed values. In this paper, we regard $S_{rad}$ at the final stage as a function of the weak scale ( the Higgs expectation value ) $v_{h}$, and show that it becomes maximum around $v_{h}={\\cal{O}}(300\\text{GeV})$ when the dimensionless couplings in the Standard Model, that is, the Higgs self coupling, the gauge couplings, and the Yukawa couplings are fixed. Roughly speaking, we find that the weak scale is given by \\begin{equation} v_{h}\\sim\\frac{T_{BBN}^{2}}{M_{pl}y_{e}^{5}},\
Weak scale from the maximum entropy principle
Hamada, Yuta; Kawai, Hikaru; Kawana, Kiyoharu
2015-03-01
The theory of the multiverse and wormholes suggests that the parameters of the Standard Model (SM) are fixed in such a way that the radiation of the S3 universe at the final stage S_rad becomes maximum, which we call the maximum entropy principle. Although it is difficult to confirm this principle generally, for a few parameters of the SM, we can check whether S_rad actually becomes maximum at the observed values. In this paper, we regard S_rad at the final stage as a function of the weak scale (the Higgs expectation value) vh, and show that it becomes maximum around vh = {{O}} (300 GeV) when the dimensionless couplings in the SM, i.e., the Higgs self-coupling, the gauge couplings, and the Yukawa couplings are fixed. Roughly speaking, we find that the weak scale is given by vh ˜ T_{BBN}2 / (M_{pl}ye5), where ye is the Yukawa coupling of electron, T_BBN is the temperature at which the Big Bang nucleosynthesis starts, and M_pl is the Planck mass.
Global characterization of the Holocene Thermal Maximum
Renssen, H.; Seppä, H.; Crosta, X.; Goosse, H.; Roche, D.M.V.A.P.
2012-01-01
We analyze the global variations in the timing and magnitude of the Holocene Thermal Maximum (HTM) and their dependence on various forcings in transient simulations covering the last 9000 years (9 ka), performed with a global atmosphere-ocean-vegetation model. In these experiments, we consider the i
Instance Optimality of the Adaptive Maximum Strategy
L. Diening; C. Kreuzer; R. Stevenson
2016-01-01
In this paper, we prove that the standard adaptive finite element method with a (modified) maximum marking strategy is instance optimal for the total error, being the square root of the squared energy error plus the squared oscillation. This result will be derived in the model setting of Poisson’s e
Maximum phonation time: variability and reliability.
Speyer, Renée; Bogaardt, Hans C A; Passos, Valéria Lima; Roodenburg, Nel P H D; Zumach, Anne; Heijnen, Mariëlle A M; Baijens, Laura W J; Fleskens, Stijn J H M; Brunings, Jan W
2010-05-01
The objective of the study was to determine maximum phonation time reliability as a function of the number of trials, days, and raters in dysphonic and control subjects. Two groups of adult subjects participated in this reliability study: a group of outpatients with functional or organic dysphonia versus a group of healthy control subjects matched by age and gender. Over a period of maximally 6 weeks, three video recordings were made of five subjects' maximum phonation time trials. A panel of five experts were responsible for all measurements, including a repeated measurement of the subjects' first recordings. Patients showed significantly shorter maximum phonation times compared with healthy controls (on average, 6.6 seconds shorter). The averaged interclass correlation coefficient (ICC) over all raters per trial for the first day was 0.998. The averaged reliability coefficient per rater and per trial for repeated measurements of the first day's data was 0.997, indicating high intrarater reliability. The mean reliability coefficient per day for one trial was 0.939. When using five trials, the reliability increased to 0.987. The reliability over five trials for a single day was 0.836; for 2 days, 0.911; and for 3 days, 0.935. To conclude, the maximum phonation time has proven to be a highly reliable measure in voice assessment. A single rater is sufficient to provide highly reliable measurements.
Maximum Phonation Time: Variability and Reliability
R. Speyer; H.C.A. Bogaardt; V.L. Passos; N.P.H.D. Roodenburg; A. Zumach; M.A.M. Heijnen; L.W.J. Baijens; S.J.H.M. Fleskens; J.W. Brunings
2010-01-01
The objective of the study was to determine maximum phonation time reliability as a function of the number of trials, days, and raters in dysphonic and control subjects. Two groups of adult subjects participated in this reliability study: a group of outpatients with functional or organic dysphonia v
Maximum likelihood estimation for integrated diffusion processes
Baltazar-Larios, Fernando; Sørensen, Michael
EM-algorithm to obtain maximum likelihood estimates of the parameters in the diffusion model. As part of the algorithm, we use a recent simple method for approximate simulation of diffusion bridges. In simulation studies for the Ornstein-Uhlenbeck process and the CIR process the proposed method works...
Maximum gain of Yagi-Uda arrays
Bojsen, J.H.; Schjær-Jacobsen, Hans; Nilsson, E.
1971-01-01
Numerical optimisation techniques have been used to find the maximum gain of some specific parasitic arrays. The gain of an array of infinitely thin, equispaced dipoles loaded with arbitrary reactances has been optimised. The results show that standard travelling-wave design methods are not optimum....... Yagi–Uda arrays with equal and unequal spacing have also been optimised with experimental verification....
Terezinha Petrucia da Nóbrega
2004-12-01
the educational process, and such conception is still predominant. Our reflection attempts to point to other ways of understanding the body in education, starting from an attitude that seeks to overcome instrumentalism and to expand educational references by considering the phenomenology of the body and its relation to sensitive knowledge as capable of amplifying the corporeal texture of the knowledge processes. Considering the body experience in the Butoh dance, we present pointers to reflect on education related to the aesthetic experience. Among them we highlight: the plasticity of the body, its incessant production of resignifications, its openness to innovation, its mutant condition, its rupture with gestural mechanization, its avoidance of the man-world and thought-feeling fragmentations, all those aspects that bring together the recursive, integrative and creative knowledge of the body are indicators to help us think on education, because their represent new possibilities for reading the real starting from the language of the gesture, effecting dialogues between knowledges and practices in the corporeal experience.
Transferability between Isolated Joint Torques and a Maximum Polyarticular Task: A Preliminary Study
Costes Antony
2016-04-01
Full Text Available The aims of this study were to determine if isolated maximum joint torques and joint torques during a maximum polyarticular task (i.e. cycling at maximum power are correlated despite joint angle and velocity discrepancies, and to assess if an isolated joint-specific torque production capability at slow angular velocity is related to cycling power. Nine cyclists completed two different evaluations of their lower limb maximum joint torques. Maximum Isolated Torques were assessed on isolated joint movements using an isokinetic ergometer and Maximum Pedalling Torques were calculated at the ankle, knee and hip for flexion and extension by inverse dynamics during cycling at maximum power. A correlation analysis was made between Maximum Isolated Torques and respective Maximum Pedalling Torques [3 joints x (flexion + extension], showing no significant relationship. Only one significant relationship was found between cycling maximum power and knee extension Maximum Isolated Torque (r=0.68, p<0.05. Lack of correlations between isolated joint torques measured at slow angular velocity and the same joint torques involved in a polyarticular task shows that transfers between both are not direct due to differences in joint angular velocities and in mono-articular versus poly articular joint torque production capabilities. However, this study confirms that maximum power in cycling is correlated with slow angular velocity mono-articular maximum knee extension torque.
Incorporating Linguistic Structure into Maximum Entropy Language Models
FANG GaoLin(方高林); GAO Wen(高文); WANG ZhaoQi(王兆其)
2003-01-01
In statistical language models, how to integrate diverse linguistic knowledge in a general framework for long-distance dependencies is a challenging issue. In this paper, an improved language model incorporating linguistic structure into maximum entropy framework is presented.The proposed model combines trigram with the structure knowledge of base phrase in which trigram is used to capture the local relation between words, while the structure knowledge of base phrase is considered to represent the long-distance relations between syntactical structures. The knowledge of syntax, semantics and vocabulary is integrated into the maximum entropy framework.Experimental results show that the proposed model improves by 24% for language model perplexity and increases about 3% for sign language recognition rate compared with the trigram model.
Payoff-monotonic game dynamics and the maximum clique problem.
Pelillo, Marcello; Torsello, Andrea
2006-05-01
Evolutionary game-theoretic models and, in particular, the so-called replicator equations have recently proven to be remarkably effective at approximately solving the maximum clique and related problems. The approach is centered around a classic result from graph theory that formulates the maximum clique problem as a standard (continuous) quadratic program and exploits the dynamical properties of these models, which, under a certain symmetry assumption, possess a Lyapunov function. In this letter, we generalize previous work along these lines in several respects. We introduce a wide family of game-dynamic equations known as payoff-monotonic dynamics, of which replicator dynamics are a special instance, and show that they enjoy precisely the same dynamical properties as standard replicator equations. These properties make any member of this family a potential heuristic for solving standard quadratic programs and, in particular, the maximum clique problem. Extensive simulations, performed on random as well as DIMACS benchmark graphs, show that this class contains dynamics that are considerably faster than and at least as accurate as replicator equations. One problem associated with these models, however, relates to their inability to escape from poor local solutions. To overcome this drawback, we focus on a particular subclass of payoff-monotonic dynamics used to model the evolution of behavior via imitation processes and study the stability of their equilibria when a regularization parameter is allowed to take on negative values. A detailed analysis of these properties suggests a whole class of annealed imitation heuristics for the maximum clique problem, which are based on the idea of varying the parameter during the imitation optimization process in a principled way, so as to avoid unwanted inefficient solutions. Experiments show that the proposed annealing procedure does help to avoid poor local optima by initially driving the dynamics toward promising regions in
Jorge Valenzuela Carreño
2007-12-01
Full Text Available En el presente trabajo se recogen las distintas aproximaciones al fenómeno de la motivación, y en dicho contexto, se constata y cuestiona aquella conceptualización de la Motivación Escolar que pone su acento en el nivel de la tarea, haciendo de ésta una motiva-ción por realizar ciertas actividades que le son demandadas por la Escuela, pero no necesariamente, una motivación por aprender. En este marco, se proponen pistas sobre algunos elementos o variables que debieran ser considerados en el constructo de la Motivación Escolar. Estas pistas tienen que ver específicamente con la incorporación, dentro del constructo, de los motivos que los alumnos tienen para aprender en la Escuela y que dan sentido a su aprendizaje escolar; y que complementan las variables vinculadas a la tarea. Así, sin olvidar que el alumno debe poner su empeño en realizar las tareas diseñadas para su aprendizaje, debemos volver a mirar las características de la Motivación Escolar, es decir, en aquello que impulsa, conduce y mantiene el esfuerzo del alumno y activa recursos cognitivos para aprender. Aprendizaje que, como es de consenso entre los educadores, no puede limitarse a la mera adquisición de información sino que debe procesada. Así, el conocimiento construido (Pozo, 2003 se hace significativo (Ausubel, 1978 y profundo (Beas, 1994; Beas et al., 2001.The work investigates the various approaches to the issue of motivation, and in this context observes and questions the idea of School Motivation that emphasizes the role of tasks, turning the completion of the latter into the motivation to carry out the activities required by the school, instead of stressing the motivation to learn. The author indicates some aspects or variables that should be considered when building up School Motivation. These pointers are specifically related to the incorporation inside this construct of the reasons pupils have to attend school, and which give meaning to their school
Model Selection Through Sparse Maximum Likelihood Estimation
Banerjee, Onureena; D'Aspremont, Alexandre
2007-01-01
We consider the problem of estimating the parameters of a Gaussian or binary distribution in such a way that the resulting undirected graphical model is sparse. Our approach is to solve a maximum likelihood problem with an added l_1-norm penalty term. The problem as formulated is convex but the memory requirements and complexity of existing interior point methods are prohibitive for problems with more than tens of nodes. We present two new algorithms for solving problems with at least a thousand nodes in the Gaussian case. Our first algorithm uses block coordinate descent, and can be interpreted as recursive l_1-norm penalized regression. Our second algorithm, based on Nesterov's first order method, yields a complexity estimate with a better dependence on problem size than existing interior point methods. Using a log determinant relaxation of the log partition function (Wainwright & Jordan (2006)), we show that these same algorithms can be used to solve an approximate sparse maximum likelihood problem for...
Maximum-entropy description of animal movement.
Fleming, Chris H; Subaşı, Yiğit; Calabrese, Justin M
2015-03-01
We introduce a class of maximum-entropy states that naturally includes within it all of the major continuous-time stochastic processes that have been applied to animal movement, including Brownian motion, Ornstein-Uhlenbeck motion, integrated Ornstein-Uhlenbeck motion, a recently discovered hybrid of the previous models, and a new model that describes central-place foraging. We are also able to predict a further hierarchy of new models that will emerge as data quality improves to better resolve the underlying continuity of animal movement. Finally, we also show that Langevin equations must obey a fluctuation-dissipation theorem to generate processes that fall from this class of maximum-entropy distributions when the constraints are purely kinematic.
Pareto versus lognormal: a maximum entropy test.
Bee, Marco; Riccaboni, Massimo; Schiavo, Stefano
2011-08-01
It is commonly found that distributions that seem to be lognormal over a broad range change to a power-law (Pareto) distribution for the last few percentiles. The distributions of many physical, natural, and social events (earthquake size, species abundance, income and wealth, as well as file, city, and firm sizes) display this structure. We present a test for the occurrence of power-law tails in statistical distributions based on maximum entropy. This methodology allows one to identify the true data-generating processes even in the case when it is neither lognormal nor Pareto. The maximum entropy approach is then compared with other widely used methods and applied to different levels of aggregation of complex systems. Our results provide support for the theory that distributions with lognormal body and Pareto tail can be generated as mixtures of lognormally distributed units.
Maximum Variance Hashing via Column Generation
Lei Luo
2013-01-01
item search. Recently, a number of data-dependent methods have been developed, reflecting the great potential of learning for hashing. Inspired by the classic nonlinear dimensionality reduction algorithm—maximum variance unfolding, we propose a novel unsupervised hashing method, named maximum variance hashing, in this work. The idea is to maximize the total variance of the hash codes while preserving the local structure of the training data. To solve the derived optimization problem, we propose a column generation algorithm, which directly learns the binary-valued hash functions. We then extend it using anchor graphs to reduce the computational cost. Experiments on large-scale image datasets demonstrate that the proposed method outperforms state-of-the-art hashing methods in many cases.
The Maximum Resource Bin Packing Problem
Boyar, J.; Epstein, L.; Favrholdt, L.M.
2006-01-01
algorithms, First-Fit-Increasing and First-Fit-Decreasing for the maximum resource variant of classical bin packing. For the on-line variant, we define maximum resource variants of classical and dual bin packing. For dual bin packing, no on-line algorithm is competitive. For classical bin packing, we find......Usually, for bin packing problems, we try to minimize the number of bins used or in the case of the dual bin packing problem, maximize the number or total size of accepted items. This paper presents results for the opposite problems, where we would like to maximize the number of bins used...... the competitive ratio of various natural algorithms. We study the general versions of the problems as well as the parameterized versions where there is an upper bound of on the item sizes, for some integer k....
Nonparametric Maximum Entropy Estimation on Information Diagrams
Martin, Elliot A; Meinke, Alexander; Děchtěrenko, Filip; Davidsen, Jörn
2016-01-01
Maximum entropy estimation is of broad interest for inferring properties of systems across many different disciplines. In this work, we significantly extend a technique we previously introduced for estimating the maximum entropy of a set of random discrete variables when conditioning on bivariate mutual informations and univariate entropies. Specifically, we show how to apply the concept to continuous random variables and vastly expand the types of information-theoretic quantities one can condition on. This allows us to establish a number of significant advantages of our approach over existing ones. Not only does our method perform favorably in the undersampled regime, where existing methods fail, but it also can be dramatically less computationally expensive as the cardinality of the variables increases. In addition, we propose a nonparametric formulation of connected informations and give an illustrative example showing how this agrees with the existing parametric formulation in cases of interest. We furthe...
Zipf's law, power laws and maximum entropy
Visser, Matt
2013-04-01
Zipf's law, and power laws in general, have attracted and continue to attract considerable attention in a wide variety of disciplines—from astronomy to demographics to software structure to economics to linguistics to zoology, and even warfare. A recent model of random group formation (RGF) attempts a general explanation of such phenomena based on Jaynes' notion of maximum entropy applied to a particular choice of cost function. In the present paper I argue that the specific cost function used in the RGF model is in fact unnecessarily complicated, and that power laws can be obtained in a much simpler way by applying maximum entropy ideas directly to the Shannon entropy subject only to a single constraint: that the average of the logarithm of the observable quantity is specified.
Zipf's law, power laws, and maximum entropy
Visser, Matt
2012-01-01
Zipf's law, and power laws in general, have attracted and continue to attract considerable attention in a wide variety of disciplines - from astronomy to demographics to economics to linguistics to zoology, and even warfare. A recent model of random group formation [RGF] attempts a general explanation of such phenomena based on Jaynes' notion of maximum entropy applied to a particular choice of cost function. In the present article I argue that the cost function used in the RGF model is in fact unnecessarily complicated, and that power laws can be obtained in a much simpler way by applying maximum entropy ideas directly to the Shannon entropy subject only to a single constraint: that the average of the logarithm of the observable quantity is specified.
Regions of constrained maximum likelihood parameter identifiability
Lee, C.-H.; Herget, C. J.
1975-01-01
This paper considers the parameter identification problem of general discrete-time, nonlinear, multiple-input/multiple-output dynamic systems with Gaussian-white distributed measurement errors. Knowledge of the system parameterization is assumed to be known. Regions of constrained maximum likelihood (CML) parameter identifiability are established. A computation procedure employing interval arithmetic is proposed for finding explicit regions of parameter identifiability for the case of linear systems. It is shown that if the vector of true parameters is locally CML identifiable, then with probability one, the vector of true parameters is a unique maximal point of the maximum likelihood function in the region of parameter identifiability and the CML estimation sequence will converge to the true parameters.
A Maximum Radius for Habitable Planets.
Alibert, Yann
2015-09-01
We compute the maximum radius a planet can have in order to fulfill two constraints that are likely necessary conditions for habitability: 1- surface temperature and pressure compatible with the existence of liquid water, and 2- no ice layer at the bottom of a putative global ocean, that would prevent the operation of the geologic carbon cycle to operate. We demonstrate that, above a given radius, these two constraints cannot be met: in the Super-Earth mass range (1-12 Mearth), the overall maximum that a planet can have varies between 1.8 and 2.3 Rearth. This radius is reduced when considering planets with higher Fe/Si ratios, and taking into account irradiation effects on the structure of the gas envelope.
Maximum Profit Configurations of Commercial Engines
Yiran Chen
2011-01-01
An investigation of commercial engines with finite capacity low- and high-price economic subsystems and a generalized commodity transfer law [n ∝ Δ (P m)] in commodity flow processes, in which effects of the price elasticities of supply and demand are introduced, is presented in this paper. Optimal cycle configurations of commercial engines for maximum profit are obtained by applying optimal control theory. In some special cases, the eventual state—market equilibrium—is solely determined by t...
A stochastic maximum principle via Malliavin calculus
Øksendal, Bernt; Zhou, Xun Yu; Meyer-Brandis, Thilo
2008-01-01
This paper considers a controlled It\\^o-L\\'evy process where the information available to the controller is possibly less than the overall information. All the system coefficients and the objective performance functional are allowed to be random, possibly non-Markovian. Malliavin calculus is employed to derive a maximum principle for the optimal control of such a system where the adjoint process is explicitly expressed.
Tissue radiation response with maximum Tsallis entropy.
Sotolongo-Grau, O; Rodríguez-Pérez, D; Antoranz, J C; Sotolongo-Costa, Oscar
2010-10-08
The expression of survival factors for radiation damaged cells is currently based on probabilistic assumptions and experimentally fitted for each tumor, radiation, and conditions. Here, we show how the simplest of these radiobiological models can be derived from the maximum entropy principle of the classical Boltzmann-Gibbs expression. We extend this derivation using the Tsallis entropy and a cutoff hypothesis, motivated by clinical observations. The obtained expression shows a remarkable agreement with the experimental data found in the literature.
Maximum Estrada Index of Bicyclic Graphs
Wang, Long; Wang, Yi
2012-01-01
Let $G$ be a simple graph of order $n$, let $\\lambda_1(G),\\lambda_2(G),...,\\lambda_n(G)$ be the eigenvalues of the adjacency matrix of $G$. The Esrada index of $G$ is defined as $EE(G)=\\sum_{i=1}^{n}e^{\\lambda_i(G)}$. In this paper we determine the unique graph with maximum Estrada index among bicyclic graphs with fixed order.
Maximum privacy without coherence, zero-error
Leung, Debbie; Yu, Nengkun
2016-09-01
We study the possible difference between the quantum and the private capacities of a quantum channel in the zero-error setting. For a family of channels introduced by Leung et al. [Phys. Rev. Lett. 113, 030512 (2014)], we demonstrate an extreme difference: the zero-error quantum capacity is zero, whereas the zero-error private capacity is maximum given the quantum output dimension.
Automatic maximum entropy spectral reconstruction in NMR.
Mobli, Mehdi; Maciejewski, Mark W; Gryk, Michael R; Hoch, Jeffrey C
2007-10-01
Developments in superconducting magnets, cryogenic probes, isotope labeling strategies, and sophisticated pulse sequences together have enabled the application, in principle, of high-resolution NMR spectroscopy to biomolecular systems approaching 1 megadalton. In practice, however, conventional approaches to NMR that utilize the fast Fourier transform, which require data collected at uniform time intervals, result in prohibitively lengthy data collection times in order to achieve the full resolution afforded by high field magnets. A variety of approaches that involve nonuniform sampling have been proposed, each utilizing a non-Fourier method of spectrum analysis. A very general non-Fourier method that is capable of utilizing data collected using any of the proposed nonuniform sampling strategies is maximum entropy reconstruction. A limiting factor in the adoption of maximum entropy reconstruction in NMR has been the need to specify non-intuitive parameters. Here we describe a fully automated system for maximum entropy reconstruction that requires no user-specified parameters. A web-accessible script generator provides the user interface to the system.
Maximum entropy analysis of cosmic ray composition
Nosek, Dalibor; Vícha, Jakub; Trávníček, Petr; Nosková, Jana
2016-01-01
We focus on the primary composition of cosmic rays with the highest energies that cause extensive air showers in the Earth's atmosphere. A way of examining the two lowest order moments of the sample distribution of the depth of shower maximum is presented. The aim is to show that useful information about the composition of the primary beam can be inferred with limited knowledge we have about processes underlying these observations. In order to describe how the moments of the depth of shower maximum depend on the type of primary particles and their energies, we utilize a superposition model. Using the principle of maximum entropy, we are able to determine what trends in the primary composition are consistent with the input data, while relying on a limited amount of information from shower physics. Some capabilities and limitations of the proposed method are discussed. In order to achieve a realistic description of the primary mass composition, we pay special attention to the choice of the parameters of the sup...
A Maximum Resonant Set of Polyomino Graphs
Zhang Heping
2016-05-01
Full Text Available A polyomino graph P is a connected finite subgraph of the infinite plane grid such that each finite face is surrounded by a regular square of side length one and each edge belongs to at least one square. A dimer covering of P corresponds to a perfect matching. Different dimer coverings can interact via an alternating cycle (or square with respect to them. A set of disjoint squares of P is a resonant set if P has a perfect matching M so that each one of those squares is M-alternating. In this paper, we show that if K is a maximum resonant set of P, then P − K has a unique perfect matching. We further prove that the maximum forcing number of a polyomino graph is equal to the cardinality of a maximum resonant set. This confirms a conjecture of Xu et al. [26]. We also show that if K is a maximal alternating set of P, then P − K has a unique perfect matching.
The maximum rate of mammal evolution
Evans, Alistair R.; Jones, David; Boyer, Alison G.; Brown, James H.; Costa, Daniel P.; Ernest, S. K. Morgan; Fitzgerald, Erich M. G.; Fortelius, Mikael; Gittleman, John L.; Hamilton, Marcus J.; Harding, Larisa E.; Lintulaakso, Kari; Lyons, S. Kathleen; Okie, Jordan G.; Saarinen, Juha J.; Sibly, Richard M.; Smith, Felisa A.; Stephens, Patrick R.; Theodor, Jessica M.; Uhen, Mark D.
2012-03-01
How fast can a mammal evolve from the size of a mouse to the size of an elephant? Achieving such a large transformation calls for major biological reorganization. Thus, the speed at which this occurs has important implications for extensive faunal changes, including adaptive radiations and recovery from mass extinctions. To quantify the pace of large-scale evolution we developed a metric, clade maximum rate, which represents the maximum evolutionary rate of a trait within a clade. We applied this metric to body mass evolution in mammals over the last 70 million years, during which multiple large evolutionary transitions occurred in oceans and on continents and islands. Our computations suggest that it took a minimum of 1.6, 5.1, and 10 million generations for terrestrial mammal mass to increase 100-, and 1,000-, and 5,000-fold, respectively. Values for whales were down to half the length (i.e., 1.1, 3, and 5 million generations), perhaps due to the reduced mechanical constraints of living in an aquatic environment. When differences in generation time are considered, we find an exponential increase in maximum mammal body mass during the 35 million years following the Cretaceous-Paleogene (K-Pg) extinction event. Our results also indicate a basic asymmetry in macroevolution: very large decreases (such as extreme insular dwarfism) can happen at more than 10 times the rate of increases. Our findings allow more rigorous comparisons of microevolutionary and macroevolutionary patterns and processes.
Minimal Length, Friedmann Equations and Maximum Density
Awad, Adel
2014-01-01
Inspired by Jacobson's thermodynamic approach[gr-qc/9504004], Cai et al [hep-th/0501055,hep-th/0609128] have shown the emergence of Friedmann equations from the first law of thermodynamics. We extend Akbar--Cai derivation [hep-th/0609128] of Friedmann equations to accommodate a general entropy-area law. Studying the resulted Friedmann equations using a specific entropy-area law, which is motivated by the generalized uncertainty principle (GUP), reveals the existence of a maximum energy density closed to Planck density. Allowing for a general continuous pressure $p(\\rho,a)$ leads to bounded curvature invariants and a general nonsingular evolution. In this case, the maximum energy density is reached in a finite time and there is no cosmological evolution beyond this point which leaves the big bang singularity inaccessible from a spacetime prospective. The existence of maximum energy density and a general nonsingular evolution is independent of the equation of state and the spacial curvature $k$. As an example w...
Maximum saliency bias in binocular fusion
Lu, Yuhao; Stafford, Tom; Fox, Charles
2016-07-01
Subjective experience at any instant consists of a single ("unitary"), coherent interpretation of sense data rather than a "Bayesian blur" of alternatives. However, computation of Bayes-optimal actions has no role for unitary perception, instead being required to integrate over every possible action-percept pair to maximise expected utility. So what is the role of unitary coherent percepts, and how are they computed? Recent work provided objective evidence for non-Bayes-optimal, unitary coherent, perception and action in humans; and further suggested that the percept selected is not the maximum a posteriori percept but is instead affected by utility. The present study uses a binocular fusion task first to reproduce the same effect in a new domain, and second, to test multiple hypotheses about exactly how utility may affect the percept. After accounting for high experimental noise, it finds that both Bayes optimality (maximise expected utility) and the previously proposed maximum-utility hypothesis are outperformed in fitting the data by a modified maximum-salience hypothesis, using unsigned utility magnitudes in place of signed utilities in the bias function.
The maximum rate of mammal evolution
Evans, Alistair R.; Jones, David; Boyer, Alison G.; Brown, James H.; Costa, Daniel P.; Ernest, S. K. Morgan; Fitzgerald, Erich M. G.; Fortelius, Mikael; Gittleman, John L.; Hamilton, Marcus J.; Harding, Larisa E.; Lintulaakso, Kari; Lyons, S. Kathleen; Okie, Jordan G.; Saarinen, Juha J.; Sibly, Richard M.; Smith, Felisa A.; Stephens, Patrick R.; Theodor, Jessica M.; Uhen, Mark D.
2012-01-01
How fast can a mammal evolve from the size of a mouse to the size of an elephant? Achieving such a large transformation calls for major biological reorganization. Thus, the speed at which this occurs has important implications for extensive faunal changes, including adaptive radiations and recovery from mass extinctions. To quantify the pace of large-scale evolution we developed a metric, clade maximum rate, which represents the maximum evolutionary rate of a trait within a clade. We applied this metric to body mass evolution in mammals over the last 70 million years, during which multiple large evolutionary transitions occurred in oceans and on continents and islands. Our computations suggest that it took a minimum of 1.6, 5.1, and 10 million generations for terrestrial mammal mass to increase 100-, and 1,000-, and 5,000-fold, respectively. Values for whales were down to half the length (i.e., 1.1, 3, and 5 million generations), perhaps due to the reduced mechanical constraints of living in an aquatic environment. When differences in generation time are considered, we find an exponential increase in maximum mammal body mass during the 35 million years following the Cretaceous–Paleogene (K–Pg) extinction event. Our results also indicate a basic asymmetry in macroevolution: very large decreases (such as extreme insular dwarfism) can happen at more than 10 times the rate of increases. Our findings allow more rigorous comparisons of microevolutionary and macroevolutionary patterns and processes. PMID:22308461
The maximum rate of mammal evolution.
Evans, Alistair R; Jones, David; Boyer, Alison G; Brown, James H; Costa, Daniel P; Ernest, S K Morgan; Fitzgerald, Erich M G; Fortelius, Mikael; Gittleman, John L; Hamilton, Marcus J; Harding, Larisa E; Lintulaakso, Kari; Lyons, S Kathleen; Okie, Jordan G; Saarinen, Juha J; Sibly, Richard M; Smith, Felisa A; Stephens, Patrick R; Theodor, Jessica M; Uhen, Mark D
2012-03-13
How fast can a mammal evolve from the size of a mouse to the size of an elephant? Achieving such a large transformation calls for major biological reorganization. Thus, the speed at which this occurs has important implications for extensive faunal changes, including adaptive radiations and recovery from mass extinctions. To quantify the pace of large-scale evolution we developed a metric, clade maximum rate, which represents the maximum evolutionary rate of a trait within a clade. We applied this metric to body mass evolution in mammals over the last 70 million years, during which multiple large evolutionary transitions occurred in oceans and on continents and islands. Our computations suggest that it took a minimum of 1.6, 5.1, and 10 million generations for terrestrial mammal mass to increase 100-, and 1,000-, and 5,000-fold, respectively. Values for whales were down to half the length (i.e., 1.1, 3, and 5 million generations), perhaps due to the reduced mechanical constraints of living in an aquatic environment. When differences in generation time are considered, we find an exponential increase in maximum mammal body mass during the 35 million years following the Cretaceous-Paleogene (K-Pg) extinction event. Our results also indicate a basic asymmetry in macroevolution: very large decreases (such as extreme insular dwarfism) can happen at more than 10 times the rate of increases. Our findings allow more rigorous comparisons of microevolutionary and macroevolutionary patterns and processes.
Maximum kinetic energy considerations in proton stereotactic radiosurgery.
Sengbusch, Evan R; Mackie, Thomas R
2011-04-12
The purpose of this study was to determine the maximum proton kinetic energy required to treat a given percentage of patients eligible for stereotactic radiosurgery (SRS) with coplanar arc-based proton therapy, contingent upon the number and location of gantry angles used. Treatment plans from 100 consecutive patients treated with SRS at the University of Wisconsin Carbone Cancer Center between June of 2007 and March of 2010 were analyzed. For each target volume within each patient, in-house software was used to place proton pencil beam spots over the distal surface of the target volume from 51 equally-spaced gantry angles of up to 360°. For each beam spot, the radiological path length from the surface of the patient to the distal boundary of the target was then calculated along a ray from the gantry location to the location of the beam spot. This data was used to generate a maximum proton energy requirement for each patient as a function of the arc length that would be spanned by the gantry angles used in a given treatment. If only a single treatment angle is required, 100% of the patients included in the study could be treated by a proton beam with a maximum kinetic energy of 118 MeV. As the length of the treatment arc is increased to 90°, 180°, 270°, and 360°, the maximum energy requirement increases to 127, 145, 156, and 179 MeV, respectively. A very high percentage of SRS patients could be treated at relatively low proton energies if the gantry angles used in the treatment plan do not span a large treatment arc. Maximum proton kinetic energy requirements increase linearly with size of the treatment arc.
Conjugate variables in continuous maximum-entropy inference.
Davis, Sergio; Gutiérrez, Gonzalo
2012-11-01
For a continuous maximum-entropy distribution (obtained from an arbitrary number of simultaneous constraints), we derive a general relation connecting the Lagrange multipliers and the expectation values of certain particularly constructed functions of the states of the system. From this relation, an estimator for a given Lagrange multiplier can be constructed from derivatives of the corresponding constraining function. These estimators sometimes lead to the determination of the Lagrange multipliers by way of solving a linear system, and, in general, they provide another tool to widen the applicability of Jaynes's formalism. This general relation, especially well suited for computer simulation techniques, also provides some insight into the interpretation of the hypervirial relations known in statistical mechanics and the recently derived microcanonical dynamical temperature. We illustrate the usefulness of these new relations with several applications in statistics.
Maximum power operation of interacting molecular motors
Golubeva, Natalia; Imparato, Alberto
2013-01-01
We study the mechanical and thermodynamic properties of different traffic models for kinesin which are relevant in biological and experimental contexts. We find that motor-motor interactions play a fundamental role by enhancing the thermodynamic efficiency at maximum power of the motors......, as compared to the non-interacting system, in a wide range of biologically compatible scenarios. We furthermore consider the case where the motor-motor interaction directly affects the internal chemical cycle and investigate the effect on the system dynamics and thermodynamics....
Maximum a posteriori decoder for digital communications
Altes, Richard A. (Inventor)
1997-01-01
A system and method for decoding by identification of the most likely phase coded signal corresponding to received data. The present invention has particular application to communication with signals that experience spurious random phase perturbations. The generalized estimator-correlator uses a maximum a posteriori (MAP) estimator to generate phase estimates for correlation with incoming data samples and for correlation with mean phases indicative of unique hypothesized signals. The result is a MAP likelihood statistic for each hypothesized transmission, wherein the highest value statistic identifies the transmitted signal.
Kernel-based Maximum Entropy Clustering
JIANG Wei; QU Jiao; LI Benxi
2007-01-01
With the development of Support Vector Machine (SVM),the "kernel method" has been studied in a general way.In this paper,we present a novel Kernel-based Maximum Entropy Clustering algorithm (KMEC).By using mercer kernel functions,the proposed algorithm is firstly map the data from their original space to high dimensional space where the data are expected to be more separable,then perform MEC clustering in the feature space.The experimental results show that the proposed method has better performance in the non-hyperspherical and complex data structure.
The sun and heliosphere at solar maximum.
Smith, E J; Marsden, R G; Balogh, A; Gloeckler, G; Geiss, J; McComas, D J; McKibben, R B; MacDowall, R J; Lanzerotti, L J; Krupp, N; Krueger, H; Landgraf, M
2003-11-14
Recent Ulysses observations from the Sun's equator to the poles reveal fundamental properties of the three-dimensional heliosphere at the maximum in solar activity. The heliospheric magnetic field originates from a magnetic dipole oriented nearly perpendicular to, instead of nearly parallel to, the Sun's rotation axis. Magnetic fields, solar wind, and energetic charged particles from low-latitude sources reach all latitudes, including the polar caps. The very fast high-latitude wind and polar coronal holes disappear and reappear together. Solar wind speed continues to be inversely correlated with coronal temperature. The cosmic ray flux is reduced symmetrically at all latitudes.
Conductivity maximum in a charged colloidal suspension
Bastea, S
2009-01-27
Molecular dynamics simulations of a charged colloidal suspension in the salt-free regime show that the system exhibits an electrical conductivity maximum as a function of colloid charge. We attribute this behavior to two main competing effects: colloid effective charge saturation due to counterion 'condensation' and diffusion slowdown due to the relaxation effect. In agreement with previous observations, we also find that the effective transported charge is larger than the one determined by the Stern layer and suggest that it corresponds to the boundary fluid layer at the surface of the colloidal particles.
Maximum entropy signal restoration with linear programming
Mastin, G.A.; Hanson, R.J.
1988-05-01
Dantzig's bounded-variable method is used to express the maximum entropy restoration problem as a linear programming problem. This is done by approximating the nonlinear objective function with piecewise linear segments, then bounding the variables as a function of the number of segments used. The use of a linear programming approach allows equality constraints found in the traditional Lagrange multiplier method to be relaxed. A robust revised simplex algorithm is used to implement the restoration. Experimental results from 128- and 512-point signal restorations are presented.
COMPARISON BETWEEN FORMULAS OF MAXIMUM SHIP SQUAT
PETRU SERGIU SERBAN
2016-06-01
Full Text Available Ship squat is a combined effect of ship’s draft and trim increase due to ship motion in limited navigation conditions. Over time, researchers conducted tests on models and ships to find a mathematical formula that can define squat. Various forms of calculating squat can be found in the literature. Among those most commonly used are of Barrass, Millward, Eryuzlu or ICORELS. This paper presents a comparison between the squat formulas to see the differences between them and which one provides the most satisfactory results. In this respect a cargo ship at different speeds was considered as a model for maximum squat calculations in canal navigation conditions.
Multi-Channel Maximum Likelihood Pitch Estimation
Christensen, Mads Græsbøll
2012-01-01
In this paper, a method for multi-channel pitch estimation is proposed. The method is a maximum likelihood estimator and is based on a parametric model where the signals in the various channels share the same fundamental frequency but can have different amplitudes, phases, and noise characteristics....... This essentially means that the model allows for different conditions in the various channels, like different signal-to-noise ratios, microphone characteristics and reverberation. Moreover, the method does not assume that a certain array structure is used but rather relies on a more general model and is hence...
Maximum entropy PDF projection: A review
Baggenstoss, Paul M.
2017-06-01
We review maximum entropy (MaxEnt) PDF projection, a method with wide potential applications in statistical inference. The method constructs a sampling distribution for a high-dimensional vector x based on knowing the sampling distribution p(z) of a lower-dimensional feature z = T (x). Under mild conditions, the distribution p(x) having highest possible entropy among all distributions consistent with p(z) may be readily found. Furthermore, the MaxEnt p(x) may be sampled, making the approach useful in Monte Carlo methods. We review the theorem and present a case study in model order selection and classification for handwritten character recognition.
CORA: Emission Line Fitting with Maximum Likelihood
Ness, Jan-Uwe; Wichmann, Rainer
2011-12-01
CORA analyzes emission line spectra with low count numbers and fits them to a line using the maximum likelihood technique. CORA uses a rigorous application of Poisson statistics. From the assumption of Poissonian noise, the software derives the probability for a model of the emission line spectrum to represent the measured spectrum. The likelihood function is used as a criterion for optimizing the parameters of the theoretical spectrum and a fixed point equation is derived allowing an efficient way to obtain line fluxes. CORA has been applied to an X-ray spectrum with the Low Energy Transmission Grating Spectrometer (LETGS) on board the Chandra observatory.
Dynamical maximum entropy approach to flocking
Cavagna, Andrea; Giardina, Irene; Ginelli, Francesco; Mora, Thierry; Piovani, Duccio; Tavarone, Raffaele; Walczak, Aleksandra M.
2014-04-01
We derive a new method to infer from data the out-of-equilibrium alignment dynamics of collectively moving animal groups, by considering the maximum entropy model distribution consistent with temporal and spatial correlations of flight direction. When bird neighborhoods evolve rapidly, this dynamical inference correctly learns the parameters of the model, while a static one relying only on the spatial correlations fails. When neighbors change slowly and the detailed balance is satisfied, we recover the static procedure. We demonstrate the validity of the method on simulated data. The approach is applicable to other systems of active matter.
Maximum Temperature Detection System for Integrated Circuits
Frankiewicz, Maciej; Kos, Andrzej
2015-03-01
The paper describes structure and measurement results of the system detecting present maximum temperature on the surface of an integrated circuit. The system consists of the set of proportional to absolute temperature sensors, temperature processing path and a digital part designed in VHDL. Analogue parts of the circuit where designed with full-custom technique. The system is a part of temperature-controlled oscillator circuit - a power management system based on dynamic frequency scaling method. The oscillator cooperates with microprocessor dedicated for thermal experiments. The whole system is implemented in UMC CMOS 0.18 μm (1.8 V) technology.
Zipf's law and maximum sustainable growth
Malevergne, Y; Sornette, D
2010-01-01
Zipf's law states that the number of firms with size greater than S is inversely proportional to S. Most explanations start with Gibrat's rule of proportional growth but require additional constraints. We show that Gibrat's rule, at all firm levels, yields Zipf's law under a balance condition between the effective growth rate of incumbent firms (which includes their possible demise) and the growth rate of investments in entrant firms. Remarkably, Zipf's law is the signature of the long-term optimal allocation of resources that ensures the maximum sustainable growth rate of an economy.
Maximum-entropy closure of hydrodynamic moment hierarchies including correlations.
Hughes, Keith H; Burghardt, Irene
2012-06-07
Generalized hydrodynamic moment hierarchies are derived which explicitly include nonequilibrium two-particle and higher-order correlations. The approach is adapted to strongly correlated media and nonequilibrium processes on short time scales which necessitate an explicit treatment of time-evolving correlations. Closure conditions for the extended moment hierarchies are formulated by a maximum-entropy approach, generalizing related closure procedures for kinetic equations. A self-consistent set of nonperturbative dynamical equations are thus obtained for a chosen set of single-particle and two-particle (and possibly higher-order) moments. Analytical results are derived for generalized Gaussian closures including the dynamic pair distribution function and a two-particle correction to the current density. The maximum-entropy closure conditions are found to involve the Kirkwood superposition approximation.
Resolution of overlapping ambiguity strings based on maximum entropy model
ZHANG Feng; FAN Xiao-zhong
2006-01-01
The resolution of overlapping ambiguity strings (OAS) is studied based on the maximum entropy model.There are two model outputs,where either the first two characters form a word or the last two characters form a word.The features of the model include one word in context of OAS,the current OAS and word probability relation of two kinds of segmentation results.OAS in training text is found by the combination of the FMM and BMM segmentation method.After feature tagging they are used to train the maximum entropy model.The People Daily corpus of January 1998 is used in training and testing.Experimental results show a closed test precision of 98.64% and an open test precision of 95.01%.The open test precision is 3,76% better compared with that of the precision of common word probability method.
Efficiency at maximum power of thermally coupled heat engines.
Apertet, Y; Ouerdane, H; Goupil, C; Lecoeur, Ph
2012-04-01
We study the efficiency at maximum power of two coupled heat engines, using thermoelectric generators (TEGs) as engines. Assuming that the heat and electric charge fluxes in the TEGs are strongly coupled, we simulate numerically the dependence of the behavior of the global system on the electrical load resistance of each generator in order to obtain the working condition that permits maximization of the output power. It turns out that this condition is not unique. We derive a simple analytic expression giving the relation between the electrical load resistance of each generator permitting output power maximization. We then focus on the efficiency at maximum power (EMP) of the whole system to demonstrate that the Curzon-Ahlborn efficiency may not always be recovered: The EMP varies with the specific working conditions of each generator but remains in the range predicted by irreversible thermodynamics theory. We discuss our results in light of nonideal Carnot engine behavior.
Maximum entropy production and the fluctuation theorem
Dewar, R C [Unite EPHYSE, INRA Centre de Bordeaux-Aquitaine, BP 81, 33883 Villenave d' Ornon Cedex (France)
2005-05-27
Recently the author used an information theoretical formulation of non-equilibrium statistical mechanics (MaxEnt) to derive the fluctuation theorem (FT) concerning the probability of second law violating phase-space paths. A less rigorous argument leading to the variational principle of maximum entropy production (MEP) was also given. Here a more rigorous and general mathematical derivation of MEP from MaxEnt is presented, and the relationship between MEP and the FT is thereby clarified. Specifically, it is shown that the FT allows a general orthogonality property of maximum information entropy to be extended to entropy production itself, from which MEP then follows. The new derivation highlights MEP and the FT as generic properties of MaxEnt probability distributions involving anti-symmetric constraints, independently of any physical interpretation. Physically, MEP applies to the entropy production of those macroscopic fluxes that are free to vary under the imposed constraints, and corresponds to selection of the most probable macroscopic flux configuration. In special cases MaxEnt also leads to various upper bound transport principles. The relationship between MaxEnt and previous theories of irreversible processes due to Onsager, Prigogine and Ziegler is also clarified in the light of these results. (letter to the editor)
Maximum Likelihood Analysis in the PEN Experiment
Lehman, Martin
2013-10-01
The experimental determination of the π+ -->e+ ν (γ) decay branching ratio currently provides the most accurate test of lepton universality. The PEN experiment at PSI, Switzerland, aims to improve the present world average experimental precision of 3 . 3 ×10-3 to 5 ×10-4 using a stopped beam approach. During runs in 2008-10, PEN has acquired over 2 ×107 πe 2 events. The experiment includes active beam detectors (degrader, mini TPC, target), central MWPC tracking with plastic scintillator hodoscopes, and a spherical pure CsI electromagnetic shower calorimeter. The final branching ratio will be calculated using a maximum likelihood analysis. This analysis assigns each event a probability for 5 processes (π+ -->e+ ν , π+ -->μ+ ν , decay-in-flight, pile-up, and hadronic events) using Monte Carlo verified probability distribution functions of our observables (energies, times, etc). A progress report on the PEN maximum likelihood analysis will be presented. Work supported by NSF grant PHY-0970013.
STUDY ON MAXIMUM SPECIFIC SLUDGE ACIVITY OF DIFFERENT ANAEROBIC GRANULAR SLUDGE BY BATCH TESTS
无
2001-01-01
The maximum specific sludge activity of granular sludge from large-scale UASB, IC and Biobed anaerobic reactors were investigated by batch tests. The limitation factors related to maximum specific sludge activity (diffusion, substrate sort, substrate concentration and granular size) were studied. The general principle and procedure for the precise measurement of maximum specific sludge activity were suggested. The potential capacity of loading rate of the IC and Biobed anaerobic reactors were analyzed and compared by use of the batch tests results.
Lake Basin Fetch and Maximum Length/Width
Minnesota Department of Natural Resources — Linear features representing the Fetch, Maximum Length and Maximum Width of a lake basin. Fetch, maximum length and average width are calcuated from the lake polygon...
王淳; 谢妮慧; 林喆
2014-01-01
在目前的空间指向镜位置伺服系统中，数字控制器输出脉宽调制（PWM）的量化误差会导致指向出现小幅角度振荡，使得指向的稳定性降低，进而导致光学遥感成像质量发生退化。针对航天器工程实现的特点，将 Delta-Sigma 调制器引入到数控软件中，在不改变硬件电路与系统参数的前提下，显著的降低了输出量化误差对伺服系统的影响。Matlab仿真结果表明，该方法可使指向镜的指向角度振荡幅值衰减至原来的1/5以下，有效地提高了指向镜的指向稳定性。%The quantization error of DPWM causes slight angle oscillation and stability decrease in pointer mirror position servo system, which results in degradation of optical imaging quality in space remote sensor. On the premise of changing no circuit and system parameters, Delta-Sigma modulator is applied to the software of products in this paper, which significantly lessens the impact of DPWM quantization error on the servo system. A Matlab simulation of pointer mirror position servo system is presented, with the result of more than 5 times attenuation of the amplitude of the angle oscillation by using Delta-Sigma modulator.
Maximum entropy principle and texture formation
Arminjon, M; Arminjon, Mayeul; Imbault, Didier
2006-01-01
The macro-to-micro transition in a heterogeneous material is envisaged as the selection of a probability distribution by the Principle of Maximum Entropy (MAXENT). The material is made of constituents, e.g. given crystal orientations. Each constituent is itself made of a large number of elementary constituents. The relevant probability is the volume fraction of the elementary constituents that belong to a given constituent and undergo a given stimulus. Assuming only obvious constraints in MAXENT means describing a maximally disordered material. This is proved to have the same average stimulus in each constituent. By adding a constraint in MAXENT, a new model, potentially interesting e.g. for texture prediction, is obtained.
MLDS: Maximum Likelihood Difference Scaling in R
Kenneth Knoblauch
2008-01-01
Full Text Available The MLDS package in the R programming language can be used to estimate perceptual scales based on the results of psychophysical experiments using the method of difference scaling. In a difference scaling experiment, observers compare two supra-threshold differences (a,b and (c,d on each trial. The approach is based on a stochastic model of how the observer decides which perceptual difference (or interval (a,b or (c,d is greater, and the parameters of the model are estimated using a maximum likelihood criterion. We also propose a method to test the model by evaluating the self-consistency of the estimated scale. The package includes an example in which an observer judges the differences in correlation between scatterplots. The example may be readily adapted to estimate perceptual scales for arbitrary physical continua.
Maximum Profit Configurations of Commercial Engines
Yiran Chen
2011-06-01
Full Text Available An investigation of commercial engines with finite capacity low- and high-price economic subsystems and a generalized commodity transfer law [n ∝ Δ (P m] in commodity flow processes, in which effects of the price elasticities of supply and demand are introduced, is presented in this paper. Optimal cycle configurations of commercial engines for maximum profit are obtained by applying optimal control theory. In some special cases, the eventual state—market equilibrium—is solely determined by the initial conditions and the inherent characteristics of two subsystems; while the different ways of transfer affect the model in respects of the specific forms of the paths of prices and the instantaneous commodity flow, i.e., the optimal configuration.
Maximum Information and Quantum Prediction Algorithms
McElwaine, J N
1997-01-01
This paper describes an algorithm for selecting a consistent set within the consistent histories approach to quantum mechanics and investigates its properties. The algorithm uses a maximum information principle to select from among the consistent sets formed by projections defined by the Schmidt decomposition. The algorithm unconditionally predicts the possible events in closed quantum systems and ascribes probabilities to these events. A simple spin model is described and a complete classification of all exactly consistent sets of histories formed from Schmidt projections in the model is proved. This result is used to show that for this example the algorithm selects a physically realistic set. Other tentative suggestions in the literature for set selection algorithms using ideas from information theory are discussed.
Maximum process problems in optimal control theory
Goran Peskir
2005-01-01
Full Text Available Given a standard Brownian motion (Btt≥0 and the equation of motion dXt=vtdt+2dBt, we set St=max0≤s≤tXs and consider the optimal control problem supvE(Sτ−Cτ, where c>0 and the supremum is taken over all admissible controls v satisfying vt∈[μ0,μ1] for all t up to τ=inf{t>0|Xt∉(ℓ0,ℓ1} with μ0g∗(St, where s↦g∗(s is a switching curve that is determined explicitly (as the unique solution to a nonlinear differential equation. The solution found demonstrates that the problem formulations based on a maximum functional can be successfully included in optimal control theory (calculus of variations in addition to the classic problem formulations due to Lagrange, Mayer, and Bolza.
Maximum Spectral Luminous Efficacy of White Light
Murphy, T W
2013-01-01
As lighting efficiency improves, it is useful to understand the theoretical limits to luminous efficacy for light that we perceive as white. Independent of the efficiency with which photons are generated, there exists a spectrally-imposed limit to the luminous efficacy of any source of photons. We find that, depending on the acceptable bandpass and---to a lesser extent---the color temperature of the light, the ideal white light source achieves a spectral luminous efficacy of 250--370 lm/W. This is consistent with previous calculations, but here we explore the maximum luminous efficacy as a function of photopic sensitivity threshold, color temperature, and color rendering index; deriving peak performance as a function of all three parameters. We also present example experimental spectra from a variety of light sources, quantifying the intrinsic efficacy of their spectral distributions.
Maximum entropy model for business cycle synchronization
Xi, Ning; Muneepeerakul, Rachata; Azaele, Sandro; Wang, Yougui
2014-11-01
The global economy is a complex dynamical system, whose cyclical fluctuations can mainly be characterized by simultaneous recessions or expansions of major economies. Thus, the researches on the synchronization phenomenon are key to understanding and controlling the dynamics of the global economy. Based on a pairwise maximum entropy model, we analyze the business cycle synchronization of the G7 economic system. We obtain a pairwise-interaction network, which exhibits certain clustering structure and accounts for 45% of the entire structure of the interactions within the G7 system. We also find that the pairwise interactions become increasingly inadequate in capturing the synchronization as the size of economic system grows. Thus, higher-order interactions must be taken into account when investigating behaviors of large economic systems.
Quantum gravity momentum representation and maximum energy
Moffat, J. W.
2016-11-01
We use the idea of the symmetry between the spacetime coordinates xμ and the energy-momentum pμ in quantum theory to construct a momentum space quantum gravity geometry with a metric sμν and a curvature tensor Pλ μνρ. For a closed maximally symmetric momentum space with a constant 3-curvature, the volume of the p-space admits a cutoff with an invariant maximum momentum a. A Wheeler-DeWitt-type wave equation is obtained in the momentum space representation. The vacuum energy density and the self-energy of a charged particle are shown to be finite, and modifications of the electromagnetic radiation density and the entropy density of a system of particles occur for high frequencies.
Video segmentation using Maximum Entropy Model
QIN Li-juan; ZHUANG Yue-ting; PAN Yun-he; WU Fei
2005-01-01
Detecting objects of interest from a video sequence is a fundamental and critical task in automated visual surveillance.Most current approaches only focus on discriminating moving objects by background subtraction whether or not the objects of interest can be moving or stationary. In this paper, we propose layers segmentation to detect both moving and stationary target objects from surveillance video. We extend the Maximum Entropy (ME) statistical model to segment layers with features, which are collected by constructing a codebook with a set of codewords for each pixel. We also indicate how the training models are used for the discrimination of target objects in surveillance video. Our experimental results are presented in terms of the success rate and the segmenting precision.
Evaluation of pliers' grip spans in the maximum gripping task and sub-maximum cutting task.
Kim, Dae-Min; Kong, Yong-Ku
2016-12-01
A total of 25 males participated to investigate the effects of the grip spans of pliers on the total grip force, individual finger forces and muscle activities in the maximum gripping task and wire-cutting tasks. In the maximum gripping task, results showed that the 50-mm grip span had significantly higher total grip strength than the other grip spans. In the cutting task, the 50-mm grip span also showed significantly higher grip strength than the 65-mm and 80-mm grip spans, whereas the muscle activities showed a higher value at 80-mm grip span. The ratios of cutting force to maximum grip strength were also investigated. Ratios of 30.3%, 31.3% and 41.3% were obtained by grip spans of 50-mm, 65-mm, and 80-mm, respectively. Thus, the 50-mm grip span for pliers might be recommended to provide maximum exertion in gripping tasks, as well as lower maximum-cutting force ratios in the cutting tasks.
Cosmic shear measurement with maximum likelihood and maximum a posteriori inference
Hall, Alex
2016-01-01
We investigate the problem of noise bias in maximum likelihood and maximum a posteriori estimators for cosmic shear. We derive the leading and next-to-leading order biases and compute them in the context of galaxy ellipticity measurements, extending previous work on maximum likelihood inference for weak lensing. We show that a large part of the bias on these point estimators can be removed using information already contained in the likelihood when a galaxy model is specified, without the need for external calibration. We test these bias-corrected estimators on simulated galaxy images similar to those expected from planned space-based weak lensing surveys, with very promising results. We find that the introduction of an intrinsic shape prior mitigates noise bias, such that the maximum a posteriori estimate can be made less biased than the maximum likelihood estimate. Second-order terms offer a check on the convergence of the estimators, but are largely sub-dominant. We show how biases propagate to shear estima...
Boundary condition effects on maximum groundwater withdrawal in coastal aquifers.
Lu, Chunhui; Chen, Yiming; Luo, Jian
2012-01-01
Prevention of sea water intrusion in coastal aquifers subject to groundwater withdrawal requires optimization of well pumping rates to maximize the water supply while avoiding sea water intrusion. Boundary conditions and the aquifer domain size have significant influences on simulating flow and concentration fields and estimating maximum pumping rates. In this study, an analytical solution is derived based on the potential-flow theory for evaluating maximum groundwater pumping rates in a domain with a constant hydraulic head landward boundary. An empirical correction factor, which was introduced by Pool and Carrera (2011) to account for mixing in the case with a constant recharge rate boundary condition, is found also applicable for the case with a constant hydraulic head boundary condition, and therefore greatly improves the usefulness of the sharp-interface analytical solution. Comparing with the solution for a constant recharge rate boundary, we find that a constant hydraulic head boundary often yields larger estimations of the maximum pumping rate and when the domain size is five times greater than the distance between the well and the coastline, the effect of setting different landward boundary conditions becomes insignificant with a relative difference between two solutions less than 2.5%. These findings can serve as a preliminary guidance for conducting numerical simulations and designing tank-scale laboratory experiments for studying groundwater withdrawal problems in coastal aquifers with minimized boundary condition effects.
Mitigation of maximum world oil production: Shortage scenarios
Hirsch, Robert L. [Management Information Services, Inc., 723 Fords Landing Way, Alexandria, VA 22314 (United States)
2008-02-15
A framework is developed for planning the mitigation of the oil shortages that will be caused by world oil production reaching a maximum and going into decline. To estimate potential economic impacts, a reasonable relationship between percent decline in world oil supply and percent decline in world GDP was determined to be roughly 1:1. As a limiting case for decline rates, giant fields were examined. Actual oil production from Europe and North America indicated significant periods of relatively flat oil production (plateaus). However, before entering its plateau period, North American oil production went through a sharp peak and steep decline. Examination of a number of future world oil production forecasts showed multi-year rollover/roll-down periods, which represent pseudoplateaus. Consideration of resource nationalism posits an Oil Exporter Withholding Scenario, which could potentially overwhelm all other considerations. Three scenarios for mitigation planning resulted from this analysis: (1) A Best Case, where maximum world oil production is followed by a multi-year plateau before the onset of a monatomic decline rate of 2-5% per year; (2) A Middling Case, where world oil production reaches a maximum, after which it drops into a long-term, 2-5% monotonic annual decline; and finally (3) A Worst Case, where the sharp peak of the Middling Case is degraded by oil exporter withholding, leading to world oil shortages growing potentially more rapidly than 2-5% per year, creating the most dire world economic impacts. (author)
Minimizing Maximum Response Time and Delay Factor in Broadcast Scheduling
Chekuri, Chandra; Moseley, Benjamin
2009-01-01
We consider online algorithms for pull-based broadcast scheduling. In this setting there are n pages of information at a server and requests for pages arrive online. When the server serves (broadcasts) a page p, all outstanding requests for that page are satisfied. We study two related metrics, namely maximum response time (waiting time) and maximum delay-factor and their weighted versions. We obtain the following results in the worst-case online competitive model. - We show that FIFO (first-in first-out) is 2-competitive even when the page sizes are different. Previously this was known only for unit-sized pages [10] via a delicate argument. Our proof differs from [10] and is perhaps more intuitive. - We give an online algorithm for maximum delay-factor that is O(1/eps^2)-competitive with (1+\\eps)-speed for unit-sized pages and with (2+\\eps)-speed for different sized pages. This improves on the algorithm in [12] which required (2+\\eps)-speed and (4+\\eps)-speed respectively. In addition we show that the algori...
Use of Maximum Entropy Modeling in Wildlife Research
Roger A. Baldwin
2009-11-01
Full Text Available Maximum entropy (Maxent modeling has great potential for identifying distributions and habitat selection of wildlife given its reliance on only presence locations. Recent studies indicate Maxent is relatively insensitive to spatial errors associated with location data, requires few locations to construct useful models, and performs better than other presence-only modeling approaches. Further advances are needed to better define model thresholds, to test model significance, and to address model selection. Additionally, development of modeling approaches is needed when using repeated sampling of known individuals to assess habitat selection. These advancements would strengthen the utility of Maxent for wildlife research and management.
Mean square convergence rates for maximum quasi-likelihood estimator
Arnoud V. den Boer
2015-03-01
Full Text Available In this note we study the behavior of maximum quasilikelihood estimators (MQLEs for a class of statistical models, in which only knowledge about the first two moments of the response variable is assumed. This class includes, but is not restricted to, generalized linear models with general link function. Our main results are related to guarantees on existence, strong consistency and mean square convergence rates of MQLEs. The rates are obtained from first principles and are stronger than known a.s. rates. Our results find important application in sequential decision problems with parametric uncertainty arising in dynamic pricing.
Maximum mass, moment of inertia and compactness of relativistic stars
Breu, Cosima
2016-01-01
A number of recent works have highlighted that it is possible to express the properties of general-relativistic stellar equilibrium configurations in terms of functions that do not depend on the specific equation of state employed to describe matter at nuclear densities. These functions are normally referred to as "universal relations" and have been found to apply, within limits, both to static or stationary isolated stars, as well as to fully dynamical and merging binary systems. Further extending the idea that universal relations can be valid also away from stability, we show that a universal relation is exhibited also by equilibrium solutions that are not stable. In particular, the mass of rotating configurations on the turning-point line shows a universal behaviour when expressed in terms of the normalised Keplerian angular momentum. In turn, this allows us to compute the maximum mass allowed by uniform rotation, M_{max}, simply in terms of the maximum mass of the nonrotating configuration, M_{TOV}, findi...
20 CFR 211.14 - Maximum creditable compensation.
2010-04-01
... 20 Employees' Benefits 1 2010-04-01 2010-04-01 false Maximum creditable compensation. 211.14... CREDITABLE RAILROAD COMPENSATION § 211.14 Maximum creditable compensation. Maximum creditable compensation... Employment Accounts shall notify each employer of the amount of maximum creditable compensation applicable...
49 CFR 230.24 - Maximum allowable stress.
2010-10-01
... 49 Transportation 4 2010-10-01 2010-10-01 false Maximum allowable stress. 230.24 Section 230.24... Allowable Stress § 230.24 Maximum allowable stress. (a) Maximum allowable stress value. The maximum allowable stress value on any component of a steam locomotive boiler shall not exceed 1/4 of the ultimate...
Theoretical Estimate of Maximum Possible Nuclear Explosion
Bethe, H. A.
1950-01-31
The maximum nuclear accident which could occur in a Na-cooled, Be moderated, Pu and power producing reactor is estimated theoretically. (T.R.H.) 2O82 Results of nuclear calculations for a variety of compositions of fast, heterogeneous, sodium-cooled, U-235-fueled, plutonium- and power-producing reactors are reported. Core compositions typical of plate-, pin-, or wire-type fuel elements and with uranium as metal, alloy, and oxide were considered. These compositions included atom ratios in the following range: U-23B to U-235 from 2 to 8; sodium to U-235 from 1.5 to 12; iron to U-235 from 5 to 18; and vanadium to U-235 from 11 to 33. Calculations were performed to determine the effect of lead and iron reflectors between the core and blanket. Both natural and depleted uranium were evaluated as the blanket fertile material. Reactors were compared on a basis of conversion ratio, specific power, and the product of both. The calculated results are in general agreement with the experimental results from fast reactor assemblies. An analysis of the effect of new cross-section values as they became available is included. (auth)
Proposed principles of maximum local entropy production.
Ross, John; Corlan, Alexandru D; Müller, Stefan C
2012-07-12
Articles have appeared that rely on the application of some form of "maximum local entropy production principle" (MEPP). This is usually an optimization principle that is supposed to compensate for the lack of structural information and measurements about complex systems, even systems as complex and as little characterized as the whole biosphere or the atmosphere of the Earth or even of less known bodies in the solar system. We select a number of claims from a few well-known papers that advocate this principle and we show that they are in error with the help of simple examples of well-known chemical and physical systems. These erroneous interpretations can be attributed to ignoring well-established and verified theoretical results such as (1) entropy does not necessarily increase in nonisolated systems, such as "local" subsystems; (2) macroscopic systems, as described by classical physics, are in general intrinsically deterministic-there are no "choices" in their evolution to be selected by using supplementary principles; (3) macroscopic deterministic systems are predictable to the extent to which their state and structure is sufficiently well-known; usually they are not sufficiently known, and probabilistic methods need to be employed for their prediction; and (4) there is no causal relationship between the thermodynamic constraints and the kinetics of reaction systems. In conclusion, any predictions based on MEPP-like principles should not be considered scientifically founded.
Maximum entropy production and plant optimization theories.
Dewar, Roderick C
2010-05-12
Plant ecologists have proposed a variety of optimization theories to explain the adaptive behaviour and evolution of plants from the perspective of natural selection ('survival of the fittest'). Optimization theories identify some objective function--such as shoot or canopy photosynthesis, or growth rate--which is maximized with respect to one or more plant functional traits. However, the link between these objective functions and individual plant fitness is seldom quantified and there remains some uncertainty about the most appropriate choice of objective function to use. Here, plants are viewed from an alternative thermodynamic perspective, as members of a wider class of non-equilibrium systems for which maximum entropy production (MEP) has been proposed as a common theoretical principle. I show how MEP unifies different plant optimization theories that have been proposed previously on the basis of ad hoc measures of individual fitness--the different objective functions of these theories emerge as examples of entropy production on different spatio-temporal scales. The proposed statistical explanation of MEP, that states of MEP are by far the most probable ones, suggests a new and extended paradigm for biological evolution--'survival of the likeliest'--which applies from biomacromolecules to ecosystems, not just to individuals.
Maximum likelihood continuity mapping for fraud detection
Hogden, J.
1997-05-01
The author describes a novel time-series analysis technique called maximum likelihood continuity mapping (MALCOM), and focuses on one application of MALCOM: detecting fraud in medical insurance claims. Given a training data set composed of typical sequences, MALCOM creates a stochastic model of sequence generation, called a continuity map (CM). A CM maximizes the probability of sequences in the training set given the model constraints, CMs can be used to estimate the likelihood of sequences not found in the training set, enabling anomaly detection and sequence prediction--important aspects of data mining. Since MALCOM can be used on sequences of categorical data (e.g., sequences of words) as well as real valued data, MALCOM is also a potential replacement for database search tools such as N-gram analysis. In a recent experiment, MALCOM was used to evaluate the likelihood of patient medical histories, where ``medical history`` is used to mean the sequence of medical procedures performed on a patient. Physicians whose patients had anomalous medical histories (according to MALCOM) were evaluated for fraud by an independent agency. Of the small sample (12 physicians) that has been evaluated, 92% have been determined fraudulent or abusive. Despite the small sample, these results are encouraging.
Maximum life spiral bevel reduction design
Savage, M.; Prasanna, M. G.; Coe, H. H.
1992-07-01
Optimization is applied to the design of a spiral bevel gear reduction for maximum life at a given size. A modified feasible directions search algorithm permits a wide variety of inequality constraints and exact design requirements to be met with low sensitivity to initial values. Gear tooth bending strength and minimum contact ratio under load are included in the active constraints. The optimal design of the spiral bevel gear reduction includes the selection of bearing and shaft proportions in addition to gear mesh parameters. System life is maximized subject to a fixed back-cone distance of the spiral bevel gear set for a specified speed ratio, shaft angle, input torque, and power. Significant parameters in the design are: the spiral angle, the pressure angle, the numbers of teeth on the pinion and gear, and the location and size of the four support bearings. Interpolated polynomials expand the discrete bearing properties and proportions into continuous variables for gradient optimization. After finding the continuous optimum, a designer can analyze near optimal designs for comparison and selection. Design examples show the influence of the bearing lives on the gear parameters in the optimal configurations. For a fixed back-cone distance, optimal designs with larger shaft angles have larger service lives.
CORA - emission line fitting with Maximum Likelihood
Ness, J.-U.; Wichmann, R.
2002-07-01
The advent of pipeline-processed data both from space- and ground-based observatories often disposes of the need of full-fledged data reduction software with its associated steep learning curve. In many cases, a simple tool doing just one task, and doing it right, is all one wishes. In this spirit we introduce CORA, a line fitting tool based on the maximum likelihood technique, which has been developed for the analysis of emission line spectra with low count numbers and has successfully been used in several publications. CORA uses a rigorous application of Poisson statistics. From the assumption of Poissonian noise we derive the probability for a model of the emission line spectrum to represent the measured spectrum. The likelihood function is used as a criterion for optimizing the parameters of the theoretical spectrum and a fixed point equation is derived allowing an efficient way to obtain line fluxes. As an example we demonstrate the functionality of the program with an X-ray spectrum of Capella obtained with the Low Energy Transmission Grating Spectrometer (LETGS) on board the Chandra observatory and choose the analysis of the Ne IX triplet around 13.5 Å.
Finding maximum JPEG image block code size
Lakhani, Gopal
2012-07-01
We present a study of JPEG baseline coding. It aims to determine the minimum storage needed to buffer the JPEG Huffman code bits of 8-bit image blocks. Since DC is coded separately, and the encoder represents each AC coefficient by a pair of run-length/AC coefficient level, the net problem is to perform an efficient search for the optimal run-level pair sequence. We formulate it as a two-dimensional, nonlinear, integer programming problem and solve it using a branch-and-bound based search method. We derive two types of constraints to prune the search space. The first one is given as an upper-bound for the sum of squares of AC coefficients of a block, and it is used to discard sequences that cannot represent valid DCT blocks. The second type constraints are based on some interesting properties of the Huffman code table, and these are used to prune sequences that cannot be part of optimal solutions. Our main result is that if the default JPEG compression setting is used, space of minimum of 346 bits and maximum of 433 bits is sufficient to buffer the AC code bits of 8-bit image blocks. Our implementation also pruned the search space extremely well; the first constraint reduced the initial search space of 4 nodes down to less than 2 nodes, and the second set of constraints reduced it further by 97.8%.
Maximum likelihood estimates of pairwise rearrangement distances.
Serdoz, Stuart; Egri-Nagy, Attila; Sumner, Jeremy; Holland, Barbara R; Jarvis, Peter D; Tanaka, Mark M; Francis, Andrew R
2017-06-21
Accurate estimation of evolutionary distances between taxa is important for many phylogenetic reconstruction methods. Distances can be estimated using a range of different evolutionary models, from single nucleotide polymorphisms to large-scale genome rearrangements. Corresponding corrections for genome rearrangement distances fall into 3 categories: Empirical computational studies, Bayesian/MCMC approaches, and combinatorial approaches. Here, we introduce a maximum likelihood estimator for the inversion distance between a pair of genomes, using a group-theoretic approach to modelling inversions introduced recently. This MLE functions as a corrected distance: in particular, we show that because of the way sequences of inversions interact with each other, it is quite possible for minimal distance and MLE distance to differently order the distances of two genomes from a third. The second aspect tackles the problem of accounting for the symmetries of circular arrangements. While, generally, a frame of reference is locked, and all computation made accordingly, this work incorporates the action of the dihedral group so that distance estimates are free from any a priori frame of reference. The philosophy of accounting for symmetries can be applied to any existing correction method, for which examples are offered. Copyright © 2017 Elsevier Ltd. All rights reserved.
Boedeker, Peter
2017-01-01
Hierarchical linear modeling (HLM) is a useful tool when analyzing data collected from groups. There are many decisions to be made when constructing and estimating a model in HLM including which estimation technique to use. Three of the estimation techniques available when analyzing data with HLM are maximum likelihood, restricted maximum…
Estimating landscape carrying capacity through maximum clique analysis.
Donovan, Therese M; Warrington, Gregory S; Schwenk, W Scott; Dinitz, Jeffrey H
2012-12-01
Habitat suitability (HS) maps are widely used tools in wildlife science and establish a link between wildlife populations and landscape pattern. Although HS maps spatially depict the distribution of optimal resources for a species, they do not reveal the population size a landscape is capable of supporting--information that is often crucial for decision makers and managers. We used a new approach, "maximum clique analysis," to demonstrate how HS maps for territorial species can be used to estimate the carrying capacity, N(k), of a given landscape. We estimated the N(k) of Ovenbirds (Seiurus aurocapillus) and bobcats (Lynx rufus) in an 1153-km2 study area in Vermont, USA. These two species were selected to highlight different approaches in building an HS map as well as computational challenges that can arise in a maximum clique analysis. We derived 30-m2 HS maps for each species via occupancy modeling (Ovenbird) and by resource utilization modeling (bobcats). For each species, we then identified all pixel locations on the map (points) that had sufficient resources in the surrounding area to maintain a home range (termed a "pseudo-home range"). These locations were converted to a mathematical graph, where any two points were linked if two pseudo-home ranges could exist on the landscape without violating territory boundaries. We used the program Cliquer to find the maximum clique of each graph. The resulting estimates of N(k) = 236 Ovenbirds and N(k) = 42 female bobcats were sensitive to different assumptions and model inputs. Estimates of N(k) via alternative, ad hoc methods were 1.4 to > 30 times greater than the maximum clique estimate, suggesting that the alternative results may be upwardly biased. The maximum clique analysis was computationally intensive but could handle problems with < 1500 total pseudo-home ranges (points). Given present computational constraints, it is best suited for species that occur in clustered distributions (where the problem can be
Maximum likelihood molecular clock comb: analytic solutions.
Chor, Benny; Khetan, Amit; Snir, Sagi
2006-04-01
Maximum likelihood (ML) is increasingly used as an optimality criterion for selecting evolutionary trees, but finding the global optimum is a hard computational task. Because no general analytic solution is known, numeric techniques such as hill climbing or expectation maximization (EM), are used in order to find optimal parameters for a given tree. So far, analytic solutions were derived only for the simplest model--three taxa, two state characters, under a molecular clock. Four taxa rooted trees have two topologies--the fork (two subtrees with two leaves each) and the comb (one subtree with three leaves, the other with a single leaf). In a previous work, we devised a closed form analytic solution for the ML molecular clock fork. In this work, we extend the state of the art in the area of analytic solutions ML trees to the family of all four taxa trees under the molecular clock assumption. The change from the fork topology to the comb incurs a major increase in the complexity of the underlying algebraic system and requires novel techniques and approaches. We combine the ultrametric properties of molecular clock trees with the Hadamard conjugation to derive a number of topology dependent identities. Employing these identities, we substantially simplify the system of polynomial equations. We finally use tools from algebraic geometry (e.g., Gröbner bases, ideal saturation, resultants) and employ symbolic algebra software to obtain analytic solutions for the comb. We show that in contrast to the fork, the comb has no closed form solutions (expressed by radicals in the input data). In general, four taxa trees can have multiple ML points. In contrast, we can now prove that under the molecular clock assumption, the comb has a unique (local and global) ML point. (Such uniqueness was previously shown for the fork.).
The Application of Maximum Principle in Supply Chain Cost Optimization
Zhou Ling
2013-09-01
Full Text Available In this paper, using the maximum principle for analyzing dynamic cost, we propose a new two-stage supply chain model of the manufacturing-assembly mode for high-tech perishable products supply chain and obtain the optimal conditions and results. On this basis, we further research the effect of localization of CODP on the total cost and the relation of CODP, inventory policy and demand type through the data simulation. The results of simulation show that CODP locates in the downstream of the product life cycle, is a linear function of the product life cycle. The result indicates that the demand forecast is the main factors influencing the total cost; meanwhile the mode of production according to the demand forecast is the deciding factor of the total cost. Also the model can reflect the relation between the total cost of two-stage supply chain and inventory, demand.
Maximum solid solubility of transition metals in vanadium solvent
ZHANG Jin-long; FANG Shou-shi; ZHOU Zi-qiang; LIN Gen-wen; GE Jian-sheng; FENG Feng
2005-01-01
Maximum solid solubility (Cmax) of different transition metals in metal solvent can be described by a semi-empirical equation using function Zf that contains electronegativity difference, atomic diameter and electron concentration. The relation between Cmax and these parameters of transition metals in vanadium solvent was studied.It is shown that the relation of Cmax and function Zf can be expressed as ln Cmax = Zf = 7. 316 5-2. 780 5 (△X)2 -71. 278δ2 -0. 855 56n2/3. The factor of atomic size parameter has the largest effect on the Cmax of the V binary alloy;followed by the factor of electronegativity difference; the electrons concentration has the smallest effect among the three bond parameters. Function Zf is used for predicting the unknown Cmax of the transition metals in vanadium solvent. The results are compared with Darken-Gurry theorem, which can be deduced by the obtained function Zf in this work.
Optimal specific wavelength for maximum thrust production in undulatory propulsion.
Nangia, Nishant; Bale, Rahul; Chen, Nelson; Hanna, Yohanna; Patankar, Neelesh A
2017-01-01
What wavelengths do undulatory swimmers use during propulsion? In this work we find that a wide range of body/caudal fin (BCF) swimmers, from larval zebrafish and herring to fully-grown eels, use specific wavelength (ratio of wavelength to tail amplitude of undulation) values that fall within a relatively narrow range. The possible emergence of this constraint is interrogated using numerical simulations of fluid-structure interaction. Based on these, it was found that there is an optimal specific wavelength (OSW) that maximizes the swimming speed and thrust generated by an undulatory swimmer. The observed values of specific wavelength for BCF animals are relatively close to this OSW. The mechanisms underlying the maximum propulsive thrust for BCF swimmers are quantified and are found to be consistent with the mechanisms hypothesized in prior work. The adherence to an optimal value of specific wavelength in most natural hydrodynamic propulsors gives rise to empirical design criteria for man-made propulsors.
Shanna Lara Miglioranzi
2011-12-01
Full Text Available OBJETIVO: verificar a relação entre capacidade vital (CV, tempos máximos de fonação de /e/ fechado emitido de forma áfona (TMF/ė/ e de /s/ (TMF/s/ e estatura em mulheres adultas. MÉTODO: 48 indivíduos do sexo feminino, entre 18 e 44 anos, com ausência de fatores intervenientes nas medidas de interesse (tabagistas, atletas, cantores, alterações pulmonares, articulatórias, tiveram suas medidas de CV, TMF/ė/ e TMF/s/ coletadas, três vezes cada, selecionando-se o maior valor obtido para cada variável, além da estatura auto-referida. Os valores das quatro variáveis do grupo foram comparados entre si por meio de análise estatística. Utilizou-se o coeficiente de correlação de Spearman para verificar sua relação; o teste de Wilcoxon para amostras relacionadas para comparar os TMF/s/ e TMF/ė/, além do cálculo do coeficiente de variação para comparar a homogeneidade dessas variáveis. RESULTADOS: correlação positiva significante entre: CV e TMF/s/ (r=0,326; P=0,024; CV e TMF/ė/ (r=0,379; P=0,008; TMF/s/ e TMF/ė/ (r=0,360; P=0,012; e CV e estatura (r=0,432; P=0,002. TMF/s/ significantemente maior do que TMF/ė/. TMF/ė/ da amostra (10,43s significantemente menor que os valores de referência (PPURPOSE: to check the relation among the values of vital capacity (CV, maximum phonation times (MPT of closed voiceless /e/ (/ė/ and of /s/ and height in adult normal women. METHOD: 48 females, between 18 and 44 years, with no intervening factors in measures of interest (smoking, sport practicing, singing, lung disorder, articulation disorder collected their measures of VC, MPT/ė/ and MPT/s/, three times each, and the highest produced values for each variable were selected for analysis, beyond the self-reported height. All four variables were compared. Spearman's correlation coefficient was used to check the relationship; Wilcoxon test for related samples was used to compare MPT/s/ and MPT/ė/, such as the coefficient of variation
New downshifted maximum in stimulated electromagnetic emission spectra
Sergeev, Evgeny; Grach, Savely
A new spectral maximum in spectra of stimulated electromagnetic emission of the ionosphere (SEE, [1]) was detected in experiments at the SURA facility in 2008 for the pump frequencies f0 4.4-4.5 MHz, most stably for f0 = 4.3 MHz, the lowest possible pump frequency at the SURA facility. The new maximum is situated at frequency shifts ∆f -6 kHz from the pump wave frequency f0 , ∆f = fSEE - f0 , somewhat closer to the f0 than the well known [2,3] Downshifted Maximum in the SEE spectrum at ∆f -9 kHz. The detection and detailed study of the new feature (which we tentatively called the New Downshifted Maximum, NDM) became possible due to high frequency resolution in spectral analysis. The following properties of the NDM are established. (i) The NDM appears in the SEE spectra simultaneously with the DM and UM features after the pump turn on (recall that the less intensive Upshifted Maximum, UM, is situated at ∆f +(6-8) kHz [2,3]). The NDM can't be attributed to 1 DM [4] or Narrow Continuum Maximum (NCM, 2 [5]) SEE features, as well as to splitted DM near gyroharmonics [2]. (ii) The NDM is observed as prominent feature for maximum pump power of the SURA facility P ≈ 120 MW ERP, for which the DM is almost covered by the Broad Continuum SEE feature [2,3]. For P ˜ 30-60 MW ERP the DM and NDM have comparable intensities. For the lesser pump power the DM prevails in the SEE spectrum, while the NDM becomes invisible being covered by the thermal Narrow Continuum feature [2]. (iii) The NDM is exactly symmetrical for the UM relatively to f0 when the former one is observed, although the UM frequency offset increases up to ∆fUM ≈ +9 kHz with a decrease of the pump power up to P ≈ 4 MW ERP. The DM formation in the SEE spectrum is attributed to a three-wave interaction between the upper and lower hybrid waves in the ionosphere, and the lower hybrid frequency ( 7 kHz) determines the frequency offset of the DM high frequency flank [2,6]. The detection of the NDM with
蒋应玲; 刘倩; 冯珏; 肖艳平; 郭庆; 乔艾春; 邓玲华
2013-01-01
Objective To investigate the cephalic vein puncture on hand in three acupressure pointer therapy intervention on the comfortable degree of infusion nursing. Methods In 2010 March to 2012 March in hospitalized patients with standard 160 patients according to the time of admission, were randomLy divided into experimental group and control group, 80 cases in each group. The control group in the"new"basis of nursing science"for the standard, using conventional intravenous infusion method. In experimental group received puncture side hand in three acupressure pointer therapy after routine intravenous infusion, observe two groups of patients in the cephalic vein puncture disposable puncture success rate, pain response and transfusion puncture nursing satisfaction were compared. Results The patients in the experimental group received acupuncture needle therapy, cephalic vein puncture success rate of puncture, disposable transfusion puncture pain reaction and nursing satisfaction compared with the control group had statistical significance (P<0.05). Conclusion The cephalic vein puncture on hand in three acupressure pointer therapy intervention, can significantly reduce the puncture pain, so that patients can cooperate smoothly infusion operation, improve the success rate of puncture and intravenous transfusion of patients satisfaction.% 目的探讨头静脉穿刺前行手三里穴位指针疗法干预对输液护理舒适度的影响.方法将2010年3月至2012年3月住院患者中符合标准的160例患者按照入院时间，随机分成实验组和对照组，每组80例.对照组以《新编护理学基础》为标准，采用常规静脉输液法.实验组在接受穿刺侧手三里穴位指针疗法后按常规静脉输液法，观察两组患者在头静脉穿刺的一次性穿刺成功率、疼痛反应及输液穿刺满意度等进行比较.结果实验组患者在接受穴位指针疗法后，头静脉穿刺的一次性穿刺成功率、疼痛反应及
The Prediction of Maximum Amplitudes of Solar Cycles and the Maximum Amplitude of Solar Cycle 24
无
2002-01-01
We present a brief review of predictions of solar cycle maximum ampli-tude with a lead time of 2 years or more. It is pointed out that a precise predictionof the maximum amplitude with such a lead-time is still an open question despiteprogress made since the 1960s. A method of prediction using statistical character-istics of solar cycles is developed: the solar cycles are divided into two groups, ahigh rising velocity (HRV) group and a low rising velocity (LRV) group, dependingon the rising velocity in the ascending phase for a given duration of the ascendingphase. The amplitude of Solar Cycle 24 can be predicted after the start of thecycle using the formula derived in this paper. Now, about 5 years before the startof the cycle, we can make a preliminary prediction of 83.2-119.4 for its maximumamplitude.
Organizing and Delivering Curriculum for Maximum Impact.
Spady, William G.
This paper presents ideas relating to curriculum organization from the Far West Laboratory for Educational Research and Development publication, "Excellence in Our Schools: Making It Happen," which attempted to synthesize nine reports written in response to the National Commission on Excellence in Education report, "A Nation at…
The biochemical composition of plankton in a subsurface chlorophyll maximum
Dortch, Quay
1987-06-01
The biochemical composition of plankton at a station with a deep, subsurface chlorophyll maximum (SCM) below a nitrogen-depleted surface layer off the Washington coast was determined in order to answer long-standing questions about the nature and causes of SCM. The chlorophyll maximum did not correspond to a protein-biomass maximum, and chlorophyll: protein ratios indicate that only in the SCM were phytoplankton a major constituent of the total biomass. Ratios of free amino acids: protein in the particulate matter were high at all depths in the euphotic zone. From this it can be concluded that phytoplankton in the SCM are N-sufficient, since they make up 80-90% of the biomass there. Above and below the SCM, where non-phytoplankton predominate, the state of N deficiency or sufficiency of the phytoplankton cannot be ascertained until more is known about how the chemical composition of phytoplankton, zooplankton and bacteria are related. However, if it is assumed that very N-sufficient zooplankton and bacteria would not coexist with very N-deficient phytoplankton, then it seems likely that the phytoplankton were also N-sufficient or nearly so. Thus, the biochemical indicators do not support the hypothesis that the SCM forms because it represents the only layer in the water column with adequate N and light for phytoplankton growth. Comparison of the chlorophyll: protein ratios with those from cultures and from other regions suggests that oligotrophic areas have a much higher proportion of non-phytoplankton biomass than do eutrophic areas.
Physics-based estimates of maximum magnitude of induced earthquakes
Ampuero, Jean-Paul; Galis, Martin; Mai, P. Martin
2016-04-01
In this study, we present new findings when integrating earthquake physics and rupture dynamics into estimates of maximum magnitude of induced seismicity (Mmax). Existing empirical relations for Mmax lack a physics-based relation between earthquake size and the characteristics of the triggering stress perturbation. To fill this gap, we extend our recent work on the nucleation and arrest of dynamic ruptures derived from fracture mechanics theory. There, we derived theoretical relations between the area and overstress of overstressed asperity and the ability of ruptures to either stop spontaneously (sub-critical ruptures) or runaway (super-critical ruptures). These relations were verified by comparison with simulation and laboratory results, namely 3D dynamic rupture simulations on faults governed by slip-weakening friction, and laboratory experiments of frictional sliding nucleated by localized stresses. Here, we apply and extend these results to situations that are representative for the induced seismicity environment. We present physics-based predictions of Mmax on a fault intersecting cylindrical reservoir. We investigate Mmax dependence on pore-pressure variations (by varying reservoir parameters), frictional parameters and stress conditions of the fault. We also derive Mmax as a function of injected volume. Our approach provides results that are consistent with observations but suggests different scaling with injected volume than that of empirical relation by McGarr, 2014.
Maximum likelihood pedigree reconstruction using integer linear programming.
Cussens, James; Bartlett, Mark; Jones, Elinor M; Sheehan, Nuala A
2013-01-01
Large population biobanks of unrelated individuals have been highly successful in detecting common genetic variants affecting diseases of public health concern. However, they lack the statistical power to detect more modest gene-gene and gene-environment interaction effects or the effects of rare variants for which related individuals are ideally required. In reality, most large population studies will undoubtedly contain sets of undeclared relatives, or pedigrees. Although a crude measure of relatedness might sometimes suffice, having a good estimate of the true pedigree would be much more informative if this could be obtained efficiently. Relatives are more likely to share longer haplotypes around disease susceptibility loci and are hence biologically more informative for rare variants than unrelated cases and controls. Distant relatives are arguably more useful for detecting variants with small effects because they are less likely to share masking environmental effects. Moreover, the identification of relatives enables appropriate adjustments of statistical analyses that typically assume unrelatedness. We propose to exploit an integer linear programming optimisation approach to pedigree learning, which is adapted to find valid pedigrees by imposing appropriate constraints. Our method is not restricted to small pedigrees and is guaranteed to return a maximum likelihood pedigree. With additional constraints, we can also search for multiple high-probability pedigrees and thus account for the inherent uncertainty in any particular pedigree reconstruction. The true pedigree is found very quickly by comparison with other methods when all individuals are observed. Extensions to more complex problems seem feasible.
王云霁; 陈嵩; 石安田; 吴艳; 邓潇
2013-01-01
Objective To investigate the relationship between the degree of displacement of centric relation-maximum intercuspation (CR-MI) and temporomandibular dysfunction (TMD) in Class Ⅱ patients. Methods The questionnaire and clinical examination were administered in 107 Class Ⅱ patients and students, using the Helkimo index (Di and Ai). The differences in condylar position between CR and MI in all three spatial planes were measured using the Condyle Position Indication (CPI). Results Di positively correlated with degree of CR-MI displacement in all five displacement (P<0. 05), while Ai positively correlated with degree of CR-MI displacement except horizontal displacement. Di and Ai significantly correlated with degree of CR-MI discrepancy in all five displacement (P< 0. 05) ; With the increasing of Di and Ai rank, the degree of CR-MI displacement and CR-MI discrepancy increased. Also, symptoms of TMD were significantly correlated with the degree of CR-MI displacement and CR-MI discrepancy except horizontal displacement (P<0. 05). Severer CR-MI displacement and CR-MI discrepancy were observed in patients who had TMD symptoms. Conclusion In Class Ⅱ patients, degree of CR-MI displacement is an important factor of TMD, and correlated with the severity of TMD.%目的 横向调查安氏Ⅱ类错(牙合)畸形正畸初诊人群正中关系位-最大牙尖交错位(CR-MI)的位移程度与颞下颌关节紊乱病(TMD)之间的相关性,为制定临床治疗计划提供参考.方法 选取符合纳入标准的安氏Ⅱ类错(牙合)患者107例.对所有受检者进行TMD症状、体征及相关情况的专科检查和问卷调查,所得结果用Helkimo指数[临床功能紊乱指数(Di)、既往功能紊乱指数(Ai)]进行分析.受检者正中关系位(CR)与最大牙尖交错位(MI)在三维方向上的差异用髁突位置测量仪(Condyle Position Indication,CPI)记录及测量.结果 所有受检者Di与5个方向、Ai与4个方向(除横向偏斜)上
Gacias, Mar; Perez-Marti, Albert; Pujol-Vidal, Magdalena; Marrero, Pedro F. [Department of Biochemistry and Molecular Biology, School of Pharmacy and the Institute of Biomedicine of the University of Barcelona (IBUB) (Spain); Haro, Diego, E-mail: dharo@ub.edu [Department of Biochemistry and Molecular Biology, School of Pharmacy and the Institute of Biomedicine of the University of Barcelona (IBUB) (Spain); Relat, Joana [Department of Biochemistry and Molecular Biology, School of Pharmacy and the Institute of Biomedicine of the University of Barcelona (IBUB) (Spain)
2012-07-13
Highlights: Black-Right-Pointing-Pointer The Cact gene is induced in mouse skeletal muscle after 24 h of fasting. Black-Right-Pointing-Pointer The Cact gene contains a functional consensus sequence for ERR. Black-Right-Pointing-Pointer This sequence binds ERR{alpha} both in vivo and in vitro. Black-Right-Pointing-Pointer This ERRE is required for the activation of Cact expression by the PGC-1/ERR axis. Black-Right-Pointing-Pointer Our results add Cact as a genuine gene target of these transcriptional regulators. -- Abstract: Carnitine/acylcarnitine translocase (CACT) is a mitochondrial-membrane carrier proteins that mediates the transport of acylcarnitines into the mitochondrial matrix for their oxidation by the mitochondrial fatty acid-oxidation pathway. CACT deficiency causes a variety of pathological conditions, such as hypoketotic hypoglycemia, cardiac arrest, hepatomegaly, hepatic dysfunction and muscle weakness, and it can be fatal in newborns and infants. Here we report that expression of the Cact gene is induced in mouse skeletal muscle after 24 h of fasting. To gain insight into the control of Cact gene expression, we examine the transcriptional regulation of the mouse Cact gene. We show that the 5 Prime -flanking region of this gene is transcriptionally active and contains a consensus sequence for the estrogen-related receptor (ERR), a member of the nuclear receptor family of transcription factors. This sequence binds ERR{alpha}in vivo and in vitro and is required for the activation of Cact expression by the peroxisome proliferator-activated receptor gamma coactivator (PGC)-1/ERR axis. We also demonstrate that XTC790, the inverse agonist of ERR{alpha}, specifically blocks Cact activation by PGC-1{beta} in C2C12 cells.
Tests of maximum oxygen intake. A critical review.
Shephard, R J
1984-01-01
The determinants of endurance effort vary, depending upon the extent of the muscle mass that is activated. Large muscle work, such as treadmill running, is halted by impending circulatory failure; lack of venous return may compound the basic problem of an excessive cardiac work-load. If the task calls for use of a smaller muscle mass, there is ultimately difficulty in perfusing the active muscles, and glycolysis is halted by an accumulation of acid metabolites. Simple field tests of endurance, such as Cooper's 12-minute run and the Canadian Home Fitness Test, have some value in the rapid screening of large populations, but like other submaximal tests of human performance they lack the precision needed to advise the individual. The directly measured maximum oxygen intake (VO2 max) varies with the type of exercise. The highest values are obtained during uphill treadmill running, but well trained athletes often approach these values during performance of sport-specific tasks. Limitations of methodology and wide interindividual variations of constitutional potential limit the interpretation of maximum oxygen intake data in terms of personal fitness, exercise prescription and the monitoring of training responses. The main practical value of VO2 max measurement is in the functional assessment of patients with cardiorespiratory disease, since changes are then large relative to the precision of the test.
Distribution of phytoplankton groups within the deep chlorophyll maximum
Latasa, Mikel
2016-11-01
The fine vertical distribution of phytoplankton groups within the deep chlorophyll maximum (DCM) was studied in the NE Atlantic during summer stratification. A simple but unconventional sampling strategy allowed examining the vertical structure with ca. 2 m resolution. The distribution of Prochlorococcus, Synechococcus, chlorophytes, pelagophytes, small prymnesiophytes, coccolithophores, diatoms, and dinoflagellates was investigated with a combination of pigment-markers, flow cytometry and optical and FISH microscopy. All groups presented minimum abundances at the surface and a maximum in the DCM layer. The cell distribution was not vertically symmetrical around the DCM peak and cells tended to accumulate in the upper part of the DCM layer. The more symmetrical distribution of chlorophyll than cells around the DCM peak was due to the increase of pigment per cell with depth. We found a vertical alignment of phytoplankton groups within the DCM layer indicating preferences for different ecological niches in a layer with strong gradients of light and nutrients. Prochlorococcus occupied the shallowest and diatoms the deepest layers. Dinoflagellates, Synechococcus and small prymnesiophytes preferred shallow DCM layers, and coccolithophores, chlorophytes and pelagophytes showed a preference for deep layers. Cell size within groups changed with depth in a pattern related to their mean size: the cell volume of the smallest group increased the most with depth while the cell volume of the largest group decreased the most. The vertical alignment of phytoplankton groups confirms that the DCM is not a homogeneous entity and indicates groups’ preferences for different ecological niches within this layer.
Noise and physical limits to maximum resolution of PET images
Herraiz, J.L.; Espana, S. [Dpto. Fisica Atomica, Molecular y Nuclear, Facultad de Ciencias Fisicas, Universidad Complutense de Madrid, Avda. Complutense s/n, E-28040 Madrid (Spain); Vicente, E.; Vaquero, J.J.; Desco, M. [Unidad de Medicina y Cirugia Experimental, Hospital GU ' Gregorio Maranon' , E-28007 Madrid (Spain); Udias, J.M. [Dpto. Fisica Atomica, Molecular y Nuclear, Facultad de Ciencias Fisicas, Universidad Complutense de Madrid, Avda. Complutense s/n, E-28040 Madrid (Spain)], E-mail: jose@nuc2.fis.ucm.es
2007-10-01
In this work we show that there is a limit for the maximum resolution achievable with a high resolution PET scanner, as well as for the best signal-to-noise ratio, which are ultimately related to the physical effects involved in the emission and detection of the radiation and thus they cannot be overcome with any particular reconstruction method. These effects prevent the spatial high frequency components of the imaged structures to be recorded by the scanner. Therefore, the information encoded in these high frequencies cannot be recovered by any reconstruction technique. Within this framework, we have determined the maximum resolution achievable for a given acquisition as a function of data statistics and scanner parameters, like the size of the crystals or the inter-crystal scatter. In particular, the noise level in the data as a limitation factor to yield high-resolution images in tomographs with small crystal sizes is outlined. These results have implications regarding how to decide the optimal number of voxels of the reconstructed image or how to design better PET scanners.
On the Threshold of Maximum-Distance Separable Codes
Kindarji, Bruno; Chabanne, Hervé
2010-01-01
Starting from a practical use of Reed-Solomon codes in a cryptographic scheme published in Indocrypt'09, this paper deals with the threshold of linear $q$-ary error-correcting codes. The security of this scheme is based on the intractability of polynomial reconstruction when there is too much noise in the vector. Our approach switches from this paradigm to an Information Theoretical point of view: is there a class of elements that are so far away from the code that the list size is always superpolynomial? Or, dually speaking, is Maximum-Likelihood decoding almost surely impossible? We relate this issue to the decoding threshold of a code, and show that when the minimal distance of the code is high enough, the threshold effect is very sharp. In a second part, we explicit lower-bounds on the threshold of Maximum-Distance Separable codes such as Reed-Solomon codes, and compute the threshold for the toy example that motivates this study.
Predicting the solar maximum with the rising rate
Du, Z L
2011-01-01
The growth rate of solar activity in the early phase of a solar cycle has been known to be well correlated with the subsequent amplitude (solar maximum). It provides very useful information for a new solar cycle as its variation reflects the temporal evolution of the dynamic process of solar magnetic activities from the initial phase to the peak phase of the cycle. The correlation coefficient between the solar maximum (Rmax) and the rising rate ({\\beta}a) at {\\Delta}m months after the solar minimum (Rmin) is studied and shown to increase as the cycle progresses with an inflection point (r = 0.83) at about {\\Delta}m = 20 months. The prediction error of Rmax based on {\\beta}a is found within estimation at the 90% level of confidence and the relative prediction error will be less than 20% when {\\Delta}m \\geq 20. From the above relationship, the current cycle (24) is preliminarily predicted to peak around October 2013 with a size of Rmax =84 \\pm 33 at the 90% level of confidence.
Radiation engineering of optical antennas for maximum field enhancement.
Seok, Tae Joon; Jamshidi, Arash; Kim, Myungki; Dhuey, Scott; Lakhani, Amit; Choo, Hyuck; Schuck, Peter James; Cabrini, Stefano; Schwartzberg, Adam M; Bokor, Jeffrey; Yablonovitch, Eli; Wu, Ming C
2011-07-13
Optical antennas have generated much interest in recent years due to their ability to focus optical energy beyond the diffraction limit, benefiting a broad range of applications such as sensitive photodetection, magnetic storage, and surface-enhanced Raman spectroscopy. To achieve the maximum field enhancement for an optical antenna, parameters such as the antenna dimensions, loading conditions, and coupling efficiency have been previously studied. Here, we present a framework, based on coupled-mode theory, to achieve maximum field enhancement in optical antennas through optimization of optical antennas' radiation characteristics. We demonstrate that the optimum condition is achieved when the radiation quality factor (Q(rad)) of optical antennas is matched to their absorption quality factor (Q(abs)). We achieve this condition experimentally by fabricating the optical antennas on a dielectric (SiO(2)) coated ground plane (metal substrate) and controlling the antenna radiation through optimizing the dielectric thickness. The dielectric thickness at which the matching condition occurs is approximately half of the quarter-wavelength thickness, typically used to achieve constructive interference, and leads to ∼20% higher field enhancement relative to a quarter-wavelength thick dielectric layer.
Maximum bubble pressure rheology of low molecular mass organogels.
Fei, Pengzhan; Wood, Steven J; Chen, Yan; Cavicchi, Kevin A
2015-01-13
Maximum bubble pressure rheology is used to characterize organogels of 0.25 wt % 12-hydroxystearic acid (12-HSA) in mineral oil, 3 wt % (1,3:2,4) dibenzylidene sorbitol (DBS) in poly(ethylene glycol), and 1 wt % 1,3:2,4-bis(3,4-dimethylbenzylidene) sorbitol (DMDBS) in poly(ethylene glycol). The maximum pressure required to inflate a bubble at the end of capillary inserted in a gel is measured. This pressure is related to the gel modulus in the case of elastic cavitation and the gel modulus and toughness in the case of irreversible fracture. The 12-HSA/mineral oil gels are used to demonstrate that this is a facile technique useful for studying time-dependent gel formation and aging and the thermal transition from a gel to a solution. Comparison is made to both qualitative gel tilting measurements and quantitative oscillatory shear rheology to highlight the utility of this measurement and its complementary nature to oscillatory shear rheology. The DBS and DMDBS demonstrate the generality of this measurement to measure gel transition temperatures.
Cardiorespiratory Fitness of Inmates of a Maximum Security Prison ...
USER
Maximum Security Prison; and also to determine the effects of age, gender, and period of incarceration on CRF. A total of 247 apparently healthy inmates of Maiduguri Maximum Security ... with different types of cardiovascular and metabolic.
Maximum likelihood polynomial regression for robust speech recognition
LU Yong; WU Zhenyang
2011-01-01
The linear hypothesis is the main disadvantage of maximum likelihood linear re- gression （MLLR）. This paper applies the polynomial regression method to model adaptation and establishes a nonlinear model adaptation algorithm using maximum likelihood polyno
An Improved Maximum C/I Scheduling Algorithm Combined with HARQ
无
2003-01-01
It is well known that traffic in downlink will be much greater than that in uplink in 3 G and that beyond. High Speed Downlink Packet Access(HSDPA) is the solution to transmission for high-speed downlink packet service in UMTS, of which Maximum C/I scheduling is one of the important algorithms related to performance enhancement. An improved scheme, Thorough Maximum C/I scheduling algorithm, is presented in this article, in which every transmitted frame has the maximum C/I. The simulation results show that the new Maximum C/I scheme outperforms the conventional scheme in throughput performance and delay performance, and that the FER decreases faster as the maximum number of the retransmission increases.
M. Mihelich
2014-11-01
Full Text Available We derive rigorous results on the link between the principle of maximum entropy production and the principle of maximum Kolmogorov–Sinai entropy using a Markov model of the passive scalar diffusion called the Zero Range Process. We show analytically that both the entropy production and the Kolmogorov–Sinai entropy seen as functions of f admit a unique maximum denoted fmaxEP and fmaxKS. The behavior of these two maxima is explored as a function of the system disequilibrium and the system resolution N. The main result of this article is that fmaxEP and fmaxKS have the same Taylor expansion at first order in the deviation of equilibrium. We find that fmaxEP hardly depends on N whereas fmaxKS depends strongly on N. In particular, for a fixed difference of potential between the reservoirs, fmaxEP(N tends towards a non-zero value, while fmaxKS(N tends to 0 when N goes to infinity. For values of N typical of that adopted by Paltridge and climatologists (N ≈ 10 ~ 100, we show that fmaxEP and fmaxKS coincide even far from equilibrium. Finally, we show that one can find an optimal resolution N* such that fmaxEP and fmaxKS coincide, at least up to a second order parameter proportional to the non-equilibrium fluxes imposed to the boundaries. We find that the optimal resolution N* depends on the non equilibrium fluxes, so that deeper convection should be represented on finer grids. This result points to the inadequacy of using a single grid for representing convection in climate and weather models. Moreover, the application of this principle to passive scalar transport parametrization is therefore expected to provide both the value of the optimal flux, and of the optimal number of degrees of freedom (resolution to describe the system.
20 CFR 617.14 - Maximum amount of TRA.
2010-04-01
... 20 Employees' Benefits 3 2010-04-01 2010-04-01 false Maximum amount of TRA. 617.14 Section 617.14... FOR WORKERS UNDER THE TRADE ACT OF 1974 Trade Readjustment Allowances (TRA) § 617.14 Maximum amount of TRA. (a) General rule. Except as provided under paragraph (b) of this section, the maximum amount of...
40 CFR 94.107 - Determination of maximum test speed.
2010-07-01
... specified in 40 CFR 1065.510. These data points form the lug curve. It is not necessary to generate the... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Determination of maximum test speed... Determination of maximum test speed. (a) Overview. This section specifies how to determine maximum test...
14 CFR 25.1505 - Maximum operating limit speed.
2010-01-01
... 14 Aeronautics and Space 1 2010-01-01 2010-01-01 false Maximum operating limit speed. 25.1505... Operating Limitations § 25.1505 Maximum operating limit speed. The maximum operating limit speed (V MO/M MO airspeed or Mach Number, whichever is critical at a particular altitude) is a speed that may not...
Maximum Performance Tests in Children with Developmental Spastic Dysarthria.
Wit, J.; And Others
1993-01-01
Three Maximum Performance Tasks (Maximum Sound Prolongation, Fundamental Frequency Range, and Maximum Repetition Rate) were administered to 11 children (ages 6-11) with spastic dysarthria resulting from cerebral palsy and 11 controls. Despite intrasubject and intersubject variability in normal and pathological speakers, the tasks were found to be…
Maximum physical capacity testing in cancer patients undergoing chemotherapy
Knutsen, L.; Quist, M; Midtgaard, J
2006-01-01
BACKGROUND: Over the past few years there has been a growing interest in the field of physical exercise in rehabilitation of cancer patients, leading to requirements for objective maximum physical capacity measurement (maximum oxygen uptake (VO(2max)) and one-repetition maximum (1RM)) to determine...
Neuromuscular determinants of maximum walking speed in well-functioning older adults.
Clark, David J; Manini, Todd M; Fielding, Roger A; Patten, Carolynn
2013-03-01
Maximum walking speed may offer an advantage over usual walking speed for clinical assessment of age-related declines in mobility function that are due to neuromuscular impairment. The objective of this study was to determine the extent to which maximum walking speed is affected by neuromuscular function of the lower extremities in older adults. We recruited two groups of healthy, well functioning older adults who differed primarily on maximum walking speed. We hypothesized that individuals with slower maximum walking speed would exhibit reduced lower extremity muscle size and impaired plantarflexion force production and neuromuscular activation during a rapid contraction of the triceps surae muscle group (soleus (SO) and gastrocnemius (MG)). All participants were required to have usual 10-meter walking speed of >1.0m/s. If the difference between usual and maximum 10m walking speed was 0.6m/s, the individual was assigned to the "Faster" group (n=12). Peak rate of force development (RFD) and rate of neuromuscular activation (rate of EMG rise) of the triceps surae muscle group were assessed during a rapid plantarflexion movement. Muscle cross sectional area of the right triceps surae, quadriceps and hamstrings muscle groups was determined by magnetic resonance imaging. Across participants, the difference between usual and maximal walking speed was predominantly dictated by maximum walking speed (r=.85). We therefore report maximum walking speed (1.76 and 2.17m/s in Slower and Faster, ptriceps surae (p=.44), quadriceps (p=.76) and hamstrings (p=.98). MG rate of EMG rise was positively associated with RFD and maximum 10m walking speed, but not the usual 10m walking speed. These findings support the conclusion that maximum walking speed is limited by impaired neuromuscular force and activation of the triceps surae muscle group. Future research should further evaluate the utility of maximum walking speed for use in clinical assessment to detect and monitor age-related
Maximum likelihood method and Fisher's information in physics and econophysics
Syska, Jacek
2012-01-01
Three steps in the development of the maximum likelihood (ML) method are presented. At first, the application of the ML method and Fisher information notion in the model selection analysis is described (Chapter 1). The fundamentals of differential geometry in the construction of the statistical space are introduced, illustrated also by examples of the estimation of the exponential models. At second, the notions of the relative entropy and the information channel capacity are introduced (Chapter 2). The observed and expected structural information principle (IP) and the variational IP of the modified extremal physical information (EPI) method of Frieden and Soffer are presented and discussed (Chapter 3). The derivation of the structural IP based on the analyticity of the logarithm of the likelihood function and on the metricity of the statistical space of the system is given. At third, the use of the EPI method is developed (Chapters 4-5). The information channel capacity is used for the field theory models cl...
Semiclassical decay of strings with maximum angular momentum
Iengo, R; Iengo, Roberto; Russo, Jorge G.
2003-01-01
A highly excited (closed or open) string state on the leading Regge trajectory can be represented by a rotating soliton solution. There is a semiclassical probability per unit cycle that this string can spontaneously break into two pieces. Here we find the resulting solutions for the outgoing two pieces, which describe two specific excited string states, and show that this semiclassical picture reproduces very accurately the features of the quantum calculation of decay in the large mass M limit. In particular, this picture prescribes the precise analytical relation of the masses M_1 and M_2 of the decay products, and indicates that the lifetime of these string states grows with the mass as T= const. a' M, in agreement with the quantum calculation. Thus, surprisingly, a string with maximum angular momentum becomes more stable for larger masses. We also point out some interesting features of the evolution after the splitting process.
Camarrone, Flavio; Ivanova, Anna; Decoster, Wivine; de Jong, Felix; van Hulle, Marc M
2015-01-01
To examine whether the minimum as well as the maximum voice intensity (i.e. sound pressure level, SPL) curves of a voice range profile (VRP) are required when discovering different voice groups based on a clustering analysis. In this approach, no a priori labeling of voice types is used. VRPs of 194 (84 male and 110 female) professional singers were registered and processed. Cluster analysis was performed with the use of features related to (1) both the maximum and minimum SPL curves and (2) the maximum SPL curve only. Features related to the maximum as well as the minimum SPL curves showed three clusters in both male and female voices. These clusters, or voice groups, are based on voice types with similar VRP features. However, when using features related only to the maximum SPL curve, the clusters became less obvious. Features related to the maximum and minimum SPL curves of a VRP are both needed in order to identify the three voice clusters. © 2016 S. Karger AG, Basel.
Tightness of the recentered maximum of the two-dimensional discrete Gaussian Free Field
Bramson, Maury
2010-01-01
We consider the maximum of the discrete two dimensional Gaussian free field (GFF) in a box, and prove that its maximum, centered at its mean, is tight, settling a long-standing conjecture. The proof combines a recent observation of Bolthausen, Deuschel and Zeitouni with elements from (Bramson 1978) and comparison theorems for Gaussian fields. An essential part of the argument is the precise evaluation, up to an error of order 1, of the expected value of the maximum of the GFF in a box. Related Gaussian fields, such as the GFF on a two-dimensional torus, are also discussed.
Stretching Labour Historiography: Pointers from South Asia
van Schendel, W.
2006-01-01
Studies of working people have long been framed by the concepts of "free" and "unfree" labour, a pair that distinguishes workers who are fully proletarianized from those who are not. Proletarians are working people without property, and therefore compelled to sell their capacities for money, but at
Sparse Dataflow Analysis with Pointers and Reachability
Madsen, Magnus; Møller, Anders
2014-01-01
quadtrees. The framework is presented as a systematic modification of a traditional dataflow analysis algorithm. Our experimental results demonstrate the effectiveness of the technique for a suite of JavaScript programs. By also comparing the performance with an idealized staged approach that computes...
Maximum entropy, word-frequency, Chinese characters, and multiple meanings.
Yan, Xiaoyong; Minnhagen, Petter
2015-01-01
The word-frequency distribution of a text written by an author is well accounted for by a maximum entropy distribution, the RGF (random group formation)-prediction. The RGF-distribution is completely determined by the a priori values of the total number of words in the text (M), the number of distinct words (N) and the number of repetitions of the most common word (k(max)). It is here shown that this maximum entropy prediction also describes a text written in Chinese characters. In particular it is shown that although the same Chinese text written in words and Chinese characters have quite differently shaped distributions, they are nevertheless both well predicted by their respective three a priori characteristic values. It is pointed out that this is analogous to the change in the shape of the distribution when translating a given text to another language. Another consequence of the RGF-prediction is that taking a part of a long text will change the input parameters (M, N, k(max)) and consequently also the shape of the frequency distribution. This is explicitly confirmed for texts written in Chinese characters. Since the RGF-prediction has no system-specific information beyond the three a priori values (M, N, k(max)), any specific language characteristic has to be sought in systematic deviations from the RGF-prediction and the measured frequencies. One such systematic deviation is identified and, through a statistical information theoretical argument and an extended RGF-model, it is proposed that this deviation is caused by multiple meanings of Chinese characters. The effect is stronger for Chinese characters than for Chinese words. The relation between Zipf's law, the Simon-model for texts and the present results are discussed.
On Global Magnetic ``Monopoly'' Near Solar Cycle Maximums
Kryvodubskyj, V.
During last maximums of the solar activity the both poles of the polar magnetic field had the same polarity. Since in the turbulent α Ω -dynamo model the excitation thresholds of the periodic dipole and quadrupole modes of the poloidal madnetic field (PMF) are rather close [Parker E. N.: 1971, Ap.J. V. 164, p. 491] then it is possible that the quadrupole mode may be excited due to variations of physical parameters in a some regions of the solar convection zone (SCZ). The pattern of the excited modes (dipole, quadrupole, octupole, etc.) is determined by the values of wave number of the Parker's dynamo-wave. We calculated these values for the SCZ model by Stix (1989) [Stix M.: 1989, The Sun. Berlin, p. 200] in the vicinity of solar tachocline (a region of strong shear of angular velocity at the base of the SCZ) with using our estimation of the helical turbulence parameter [Krivodubskij V. N.: 1998, Astron. Reports V. 42, No 1, p. 122] and values of the radial gradient of the angular velocity obtained from the newer helioseismic measurements (during rising phase of 23th solar cycle: 1995-1999) [Howe R.,Christensen-Dalsgaard J., Hill F. et al.: 2000, Science. V. 287, p. 2456]. It is found out that at low latitudes dynamo mechanism produces rather the dipole (wave number ≈ -7), the main antisymmetric, relatively to equatorial plane, mode of the PMF; while at the latitudes higher than 50o the conditions are more favourable for exciting of the quadrupole (wave number ≈ +8), the lowest symmetric mode. Arised north-south magnetic structure asymmetry gives an opportunity to explain the space magnetic anomaly of the PMF (``monopoly'') observed near solar cycle maximums.
Kirkegaard, Poul Henning; Nielsen, Søren R.K.; Micaletti, R. C.;
This paper considers estimation of the Maximum Damage Indicator (MSDI) by using time-frequency system identification techniques for an RC-structure subjected to earthquake excitation. The MSDI relates the global damage state of the RC-structure to the relative decrease of the fundamental eigenfre...
Upstream proton cyclotron waves at Venus near solar maximum
Delva, M.; Bertucci, C.; Volwerk, M.; Lundin, R.; Mazelle, C.; Romanelli, N.
2015-01-01
magnetometer data of Venus Express are analyzed for the occurrence of waves at the proton cyclotron frequency in the spacecraft frame in the upstream region of Venus, for conditions of rising solar activity. The data of two Venus years up to the time of highest sunspot number so far (1 Mar 2011 to 31 May 2012) are studied to reveal the properties of the waves and the interplanetary magnetic field (IMF) conditions under which they are observed. In general, waves generated by newborn protons from exospheric hydrogen are observed under quasi- (anti)parallel conditions of the IMF and the solar wind velocity, as is expected from theoretical models. The present study near solar maximum finds significantly more waves than a previous study for solar minimum, with an asymmetry in the wave occurrence, i.e., mainly under antiparallel conditions. The plasma data from the Analyzer of Space Plasmas and Energetic Atoms instrument aboard Venus Express enable analysis of the background solar wind conditions. The prevalence of waves for IMF in direction toward the Sun is related to the stronger southward tilt of the heliospheric current sheet for the rising phase of Solar Cycle 24, i.e., the "bashful ballerina" is responsible for asymmetric background solar wind conditions. The increase of the number of wave occurrences may be explained by a significant increase in the relative density of planetary protons with respect to the solar wind background. An exceptionally low solar wind proton density is observed during the rising phase of Solar Cycle 24. At the same time, higher EUV increases the ionization in the Venus exosphere, resulting in higher supply of energy from a higher number of newborn protons to the wave. We conclude that in addition to quasi- (anti)parallel conditions of the IMF and the solar wind velocity direction, the higher relative density of Venus exospheric protons with respect to the background solar wind proton density is the key parameter for the higher number of
A Note on k-Limited Maximum Base
Yang Ruishun; Yang Xiaowei
2006-01-01
The problem of k-limited maximum base was specified into two special problems of k-limited maximum base; that is, let subset D of the problem of k-limited maximum base be an independent set and a circuit of the matroid, respectively. It was proved that under this circumstance the collections of k-limited base satisfy base axioms. Then a new matroid was determined, and the problem of k-limited maximum base was transformed to the problem of maximum base of this new matroid. Aiming at the problem, two algorithms, which in essence are greedy algorithms based on former matroid, were presented for the two special problems of k-limited maximum base. They were proved to be reasonable and more efficient than the algorithm presented by Ma Zhongfan in view of the complexity of algorithm.
On the maximum-entropy/autoregressive modeling of time series
Chao, B. F.
1984-01-01
The autoregressive (AR) model of a random process is interpreted in the light of the Prony's relation which relates a complex conjugate pair of poles of the AR process in the z-plane (or the z domain) on the one hand, to the complex frequency of one complex harmonic function in the time domain on the other. Thus the AR model of a time series is one that models the time series as a linear combination of complex harmonic functions, which include pure sinusoids and real exponentials as special cases. An AR model is completely determined by its z-domain pole configuration. The maximum-entropy/autogressive (ME/AR) spectrum, defined on the unit circle of the z-plane (or the frequency domain), is nothing but a convenient, but ambiguous visual representation. It is asserted that the position and shape of a spectral peak is determined by the corresponding complex frequency, and the height of the spectral peak contains little information about the complex amplitude of the complex harmonic functions.
Maximum likelihood estimation for cytogenetic dose-response curves
Frome, E.L; DuFrain, R.J.
1983-10-01
In vitro dose-response curves are used to describe the relation between the yield of dicentric chromosome aberrations and radiation dose for human lymphocytes. The dicentric yields follow the Poisson distribution, and the expected yield depends on both the magnitude and the temporal distribution of the dose for low LET radiation. A general dose-response model that describes this relation has been obtained by Kellerer and Rossi using the theory of dual radiation action. The yield of elementary lesions is kappa(..gamma..d + g(t, tau)d/sup 2/), where t is the time and d is dose. The coefficient of the d/sup 2/ term is determined by the recovery function and the temporal mode of irradiation. Two special cases of practical interest are split-dose and continuous exposure experiments, and the resulting models are intrinsically nonlinear in the parameters. A general purpose maximum likelihood estimation procedure is described and illustrated with numerical examples from both experimental designs. Poisson regression analysis is used for estimation, hypothesis testing, and regression diagnostics. Results are discussed in the context of exposure assessment procedures for both acute and chronic human radiation exposure.
Feedback Limits to Maximum Seed Masses of Black Holes
Pacucci, Fabio; Natarajan, Priyamvada; Ferrara, Andrea
2017-02-01
The most massive black holes observed in the universe weigh up to ∼1010 M ⊙, nearly independent of redshift. Reaching these final masses likely required copious accretion and several major mergers. Employing a dynamical approach that rests on the role played by a new, relevant physical scale—the transition radius—we provide a theoretical calculation of the maximum mass achievable by a black hole seed that forms in an isolated halo, one that scarcely merged. Incorporating effects at the transition radius and their impact on the evolution of accretion in isolated halos, we are able to obtain new limits for permitted growth. We find that large black hole seeds (M • ≳ 104 M ⊙) hosted in small isolated halos (M h ≲ 109 M ⊙) accreting with relatively small radiative efficiencies (ɛ ≲ 0.1) grow optimally in these circumstances. Moreover, we show that the standard M •–σ relation observed at z ∼ 0 cannot be established in isolated halos at high-z, but requires the occurrence of mergers. Since the average limiting mass of black holes formed at z ≳ 10 is in the range 104–6 M ⊙, we expect to observe them in local galaxies as intermediate-mass black holes, when hosted in the rare halos that experienced only minor or no merging events. Such ancient black holes, formed in isolation with subsequent scant growth, could survive, almost unchanged, until present.
无
2008-01-01
Ecological systems in the headwaters of the Yellow River, characterized by hash natural environmental conditions, are very vulnerable to climatic change. In the most recent decades, this area greatly attracted the public's attention for its more and more deteriorating environmental conditions. Based on tree-ring samples from the Xiqing Mountain and A'nyêmagên Mountains at the headwaters of the Yellow River in the Northeastern Tibetan Plateau, we reconstructed the minimum temperatures in the winter half year over the last 425 years and the maximum temperatures in the summer half year over the past 700 years in this region. The variation of minimum temperature in the winter half year during the time span of 1578-1940 was a relatively stable trend, which was followed by an abrupt warming trend since 1941. However, there is no significant warming trend for the maximum temperature in the summer half year over the 20th century. The asymmetric variation patterns between the minimum and maximum temperatures were observed in this study over the past 425 years. During the past 425 years, there are similar variation patterns between the minimum and maximum temperatures; however, the minimum temperatures vary about 25 years earlier compared to the maximum temperatures. If such a trend of variation patterns between the minimum and maximum temperatures over the past 425 years continues in the future 30 years, the maximum temperature in this region will increase significantly.
Jacoby; GORDON
2008-01-01
Ecological systems in the headwaters of the Yellow River, characterized by hash natural environmental conditions, are very vulnerable to climatic change. In the most recent decades, this area greatly attracted the public’s attention for its more and more deteriorating environmental conditions. Based on tree-ring samples from the Xiqing Mountain and A’nyêmagên Mountains at the headwaters of the Yellow River in the Northeastern Tibetan Plateau, we reconstructed the minimum temperatures in the winter half year over the last 425 years and the maximum temperatures in the summer half year over the past 700 years in this region. The variation of minimum temperature in the winter half year during the time span of 1578―1940 was a relatively stable trend, which was followed by an abrupt warming trend since 1941. However, there is no significant warming trend for the maximum temperature in the summer half year over the 20th century. The asymmetric variation patterns between the minimum and maximum temperatures were observed in this study over the past 425 years. During the past 425 years, there are similar variation patterns between the minimum and maximum temperatures; however, the minimum temperatures vary about 25 years earlier compared to the maximum temperatures. If such a trend of variation patterns between the minimum and maximum temperatures over the past 425 years continues in the future 30 years, the maximum temperature in this region will increase significantly.
An Interval Maximum Entropy Method for Quadratic Programming Problem
RUI Wen-juan; CAO De-xin; SONG Xie-wu
2005-01-01
With the idea of maximum entropy function and penalty function methods, we transform the quadratic programming problem into an unconstrained differentiable optimization problem, discuss the interval extension of the maximum entropy function, provide the region deletion test rules and design an interval maximum entropy algorithm for quadratic programming problem. The convergence of the method is proved and numerical results are presented. Both theoretical and numerical results show that the method is reliable and efficient.
Hutchinson, Thomas H. [Plymouth Marine Laboratory, Prospect Place, The Hoe, Plymouth PL1 3DH (United Kingdom)], E-mail: thom1@pml.ac.uk; Boegi, Christian [BASF SE, Product Safety, GUP/PA, Z470, 67056 Ludwigshafen (Germany); Winter, Matthew J. [AstraZeneca Safety, Health and Environment, Brixham Environmental Laboratory, Devon TQ5 8BA (United Kingdom); Owens, J. Willie [The Procter and Gamble Company, Central Product Safety, 11810 East Miami River Road, Cincinnati, OH 45252 (United States)
2009-02-19
There is increasing recognition of the need to identify specific sublethal effects of chemicals, such as reproductive toxicity, and specific modes of actions of the chemicals, such as interference with the endocrine system. To achieve these aims requires criteria which provide a basis to interpret study findings so as to separate these specific toxicities and modes of action from not only acute lethality per se but also from severe inanition and malaise that non-specifically compromise reproductive capacity and the response of endocrine endpoints. Mammalian toxicologists have recognized that very high dose levels are sometimes required to elicit both specific adverse effects and present the potential of non-specific 'systemic toxicity'. Mammalian toxicologists have developed the concept of a maximum tolerated dose (MTD) beyond which a specific toxicity or action cannot be attributed to a test substance due to the compromised state of the organism. Ecotoxicologists are now confronted by a similar challenge and must develop an analogous concept of a MTD and the respective criteria. As examples of this conundrum, we note recent developments in efforts to validate protocols for fish reproductive toxicity and endocrine screens (e.g. some chemicals originally selected as 'negatives' elicited decreases in fecundity or changes in endpoints intended to be biomarkers for endocrine modes of action). Unless analogous criteria can be developed, the potentially confounding effects of systemic toxicity may then undermine the reliable assessment of specific reproductive effects or biomarkers such as vitellogenin or spiggin. The same issue confronts other areas of aquatic toxicology (e.g., genotoxicity) and the use of aquatic animals for preclinical assessments of drugs (e.g., use of zebrafish for drug safety assessment). We propose that there are benefits to adopting the concept of an MTD for toxicology and pharmacology studies using fish and other aquatic
Integer Programming Model for Maximum Clique in Graph
YUAN Xi-bo; YANG You; ZENG Xin-hai
2005-01-01
The maximum clique or maximum independent set of graph is a classical problem in graph theory. Combined with Boolean algebra and integer programming, two integer programming models for maximum clique problem,which improve the old results were designed in this paper. Then, the programming model for maximum independent set is a corollary of the main results. These two models can be easily applied to computer algorithm and software, and suitable for graphs of any scale. Finally the models are presented as Lingo algorithms, verified and compared by several examples.
Counterexamples to convergence theorem of maximum-entropy clustering algorithm
于剑; 石洪波; 黄厚宽; 孙喜晨; 程乾生
2003-01-01
In this paper, we surveyed the development of maximum-entropy clustering algorithm, pointed out that the maximum-entropy clustering algorithm is not new in essence, and constructed two examples to show that the iterative sequence given by the maximum-entropy clustering algorithm may not converge to a local minimum of its objective function, but a saddle point. Based on these results, our paper shows that the convergence theorem of maximum-entropy clustering algorithm put forward by Kenneth Rose et al. does not hold in general cases.
潘鈺筠Yu-Yun Pan
2011-03-01
Full Text Available 本研究利用眼動儀觀察學習者於觀看數位課程圖像與閱讀過程中，眼睛觀看位置與眼球運動的情形。透過不同設計的數位課程介面，探討學習者於數位課程學習時的狀況與學習成效，輔以問卷評量與訪談。研究結果顯示：數位課程使用移動指示游標，對學習者的注意力分布具有顯著影響，並且會影響學習者觀看課程時的平均掃視幅度，對學習成效則不具影響。整體而言，學習者對數位課程畫面的滿意度除「訊息配置」之外，其他各項有指標組都高於無指標組。An eye tracker was used in this study to detect the effect of instruction design of e-Learning with or without using the guided pointer on learner’s locus of attention and learning achievement. The results of the experiment found e-Learning instruction with guided pointer would significantly increase learner’s locus of attention, but have no effect on learning achievement. However, the learning attitudes of the deceptive data revealed that learners in the indexing pointer group have higher scores than that of in the group without indexing pointer.
Ashokkumar, Saranya, E-mail: saras@food.dtu.dk [Accoat A/S, Munkegardsvej 16, 3490 Kvistgard (Denmark); Food Production Engineering, DTU FOOD, Technical University of Denmark, 2800 Kgs. Lyngby (Denmark); Adler-Nissen, Jens [Food Production Engineering, DTU FOOD, Technical University of Denmark, 2800 Kgs. Lyngby (Denmark); Moller, Per [Department of Materials Science and Engineering, DTU Mechanical Engineering, Technical University of Denmark, 2800 Kgs. Lyngby (Denmark)
2012-12-15
Graphical abstract: Plot of cos {theta} versus temperature for metal and ceramic surfaces where cos {theta} rises linearly with increase in temperature. Highlights: Black-Right-Pointing-Pointer cos {theta} of olive oil on different surface materials rises linearly with increase in temperature. Black-Right-Pointing-Pointer Slopes are much higher for quasicrystalline and polymers than for ceramics. Black-Right-Pointing-Pointer Increase in surface roughness and surface flaws increases surface wettability. Black-Right-Pointing-Pointer Contact angle values gave information for grouping easy-clean polymers from other materials. Black-Right-Pointing-Pointer Contact angle measurements cannot directly estimate the cleanability of a surface. - Abstract: The main aim of the work was to investigate the wettability of different surface materials with vegetable oil (olive oil) over the temperature range of 25-200 Degree-Sign C to understand the differences in cleanability of different surfaces exposed to high temperatures in food processes. The different surface materials investigated include stainless steel (reference), PTFE (polytetrafluoroethylene), silicone, quasicrystalline (Al, Fe, Cr) and ceramic coatings: zirconium oxide (ZrO{sub 2}), zirconium nitride (ZrN) and titanium aluminum nitride (TiAlN). The ceramic coatings were deposited on stainless steel with two different levels of roughness. The cosine of the contact angle of olive oil on different surface materials rises linearly with increasing temperature. Among the materials analyzed, polymers (PTFE, silicone) gave the lowest cos {theta} values. Studies of the effect of roughness and surface flaws on wettability revealed that the cos {theta} values increases with increasing roughness and surface flaws. Correlation analysis indicates that the measured contact angle values gave useful information for grouping easy-clean polymer materials from the other materials; for the latter group, there is no direct relation between
Recent advance on the efficiency at maximum power of heat engines
Tu Zhan-Chun
2012-01-01
This review reports several key advances on the theoretical investigations of efficiency at maximum power of heat engines in the past five years.The analytical results of efficiency at maximum power for the Curzon-Ahlborn heat engine,the stochastic heat engine constructed from a Brownian particle,and Feynman's ratchet as a heat engine are presented.It is found that:the efficiency at maximum power exhibits universal behavior at small relative temperature differences; the lower and the upper bounds might exist under quite general conditions; and the problem of efficiency at maximum power comes down to seeking for the minimum irreversible entropy production in each finite-time isothermal process for a given time.
Asymptotic properties of maximum likelihood estimators in models with multiple change points
He, Heping; 10.3150/09-BEJ232
2011-01-01
Models with multiple change points are used in many fields; however, the theoretical properties of maximum likelihood estimators of such models have received relatively little attention. The goal of this paper is to establish the asymptotic properties of maximum likelihood estimators of the parameters of a multiple change-point model for a general class of models in which the form of the distribution can change from segment to segment and in which, possibly, there are parameters that are common to all segments. Consistency of the maximum likelihood estimators of the change points is established and the rate of convergence is determined; the asymptotic distribution of the maximum likelihood estimators of the parameters of the within-segment distributions is also derived. Since the approach used in single change-point models is not easily extended to multiple change-point models, these results require the introduction of those tools for analyzing the likelihood function in a multiple change-point model.
The relationship between the Guinea Highlands and the West African offshore rainfall maximum
Hamilton, H. L.; Young, G. S.; Evans, J. L.; Fuentes, J. D.; Núñez Ocasio, K. M.
2017-01-01
Satellite rainfall estimates reveal a consistent rainfall maximum off the West African coast during the monsoon season. An analysis of 16 years of rainfall in the monsoon season is conducted to explore the drivers of such copious amounts of rainfall. Composites of daily rainfall and midlevel meridional winds centered on the days with maximum rainfall show that the day with the heaviest rainfall follows the strongest midlevel northerlies but coincides with peak low-level moisture convergence. Rain type composites show that convective rain dominates the study region. The dominant contribution to the offshore rainfall maximum is convective development driven by the enhancement of upslope winds near the Guinea Highlands. The enhancement in the upslope flow is closely related to African easterly waves propagating off the continent that generate low-level cyclonic vorticity and convergence. Numerical simulations reproduce the observed rainfall maximum and indicate that it weakens if the African topography is reduced.
Night vision image fusion for target detection with improved 2D maximum entropy segmentation
Bai, Lian-fa; Liu, Ying-bin; Yue, Jiang; Zhang, Yi
2013-08-01
Infrared and LLL image are used for night vision target detection. In allusion to the characteristics of night vision imaging and lack of traditional detection algorithm for segmentation and extraction of targets, we propose a method of infrared and LLL image fusion for target detection with improved 2D maximum entropy segmentation. Firstly, two-dimensional histogram was improved by gray level and maximum gray level in weighted area, weights were selected to calculate the maximum entropy for infrared and LLL image segmentation by using the histogram. Compared with the traditional maximum entropy segmentation, the algorithm had significant effect in target detection, and the functions of background suppression and target extraction. And then, the validity of multi-dimensional characteristics AND operation on the infrared and LLL image feature level fusion for target detection is verified. Experimental results show that detection algorithm has a relatively good effect and application in target detection and multiple targets detection in complex background.
An Efficient Algorithm for Maximum-Entropy Extension of Block-Circulant Covariance Matrices
Carli, Francesca P; Pavon, Michele; Picci, Giorgio
2011-01-01
This paper deals with maximum entropy completion of partially specified block-circulant matrices. Since positive definite symmetric circulants happen to be covariance matrices of stationary periodic processes, in particular of stationary reciprocal processes, this problem has applications in signal processing, in particular to image modeling. Maximum entropy completion is strictly related to maximum likelihood estimation subject to certain conditional independence constraints. The maximum entropy completion problem for block-circulant matrices is a nonlinear problem which has recently been solved by the authors, although leaving open the problem of an efficient computation of the solution. The main contribution of this paper is to provide an efficient algorithm for computing the solution. Simulation shows that our iterative scheme outperforms various existing approaches, especially for large dimensional problems. A necessary and sufficient condition for the existence of a positive definite circulant completio...
The NFL Combine 40-Yard Dash: How Important is Maximum Velocity?
Clark, Kenneth P; Rieger, Randall H; Bruno, Richard F; Stearne, David J
2017-06-22
This investigation analyzed the sprint velocity profiles for athletes who completed the 40-yard (36.6m) dash at the 2016 NFL Combine. The purpose was to evaluate the relationship between maximum velocity and sprint performance, and to compare acceleration patterns for fast and slow athletes. Using freely available online sources, data were collected for body mass and sprint performance (36.6m time with split intervals at 9.1 and 18.3m). For each athlete, split times were utilized to generate modeled curves of distance vs. time, velocity vs. time, and velocity vs. distance using a mono-exponential equation. Model parameters were used to quantify acceleration patterns as the ratio of maximum velocity to maximum acceleration (vmax / amax, or τ). Linear regression was used to evaluate the relationship between maximum velocity and sprint performance for the entire sample. Additionally, athletes were categorized into fast and slow groups based on maximum velocity, with independent t-tests and effect size statistics used to evaluate between-group differences in sprint performance and acceleration patterns. Results indicated that maximum velocity was strongly correlated with sprint performance across 9.1m, 18.3m, and 36.6m (r of 0.72, 0.83, and 0.94, respectively). However, both fast and slow groups accelerated in a similar pattern relative to maximum velocity (τ = 0.768 ± 0.068s for the fast group and τ = 0.773 ± 0.070s for the slow group). We conclude that maximum velocity is of critical importance to 36.6m time, and inclusion of more maximum velocity training may be warranted for athletes preparing for the NFL Combine.
Combining Experiments and Simulations Using the Maximum Entropy Principle
Boomsma, Wouter; Ferkinghoff-Borg, Jesper; Lindorff-Larsen, Kresten
2014-01-01
are not in quantitative agreement with experimental data. The principle of maximum entropy is a general procedure for constructing probability distributions in the light of new data, making it a natural tool in cases when an initial model provides results that are at odds with experiments. The number of maximum entropy...
49 CFR 174.86 - Maximum allowable operating speed.
2010-10-01
... 49 Transportation 2 2010-10-01 2010-10-01 false Maximum allowable operating speed. 174.86 Section... operating speed. (a) For molten metals and molten glass shipped in packagings other than those prescribed in § 173.247 of this subchapter, the maximum allowable operating speed may not exceed 24 km/hour (15...
Parametric optimization of thermoelectric elements footprint for maximum power generation
Rezania, A.; Rosendahl, Lasse; Yin, Hao
2014-01-01
The development studies in thermoelectric generator (TEG) systems are mostly disconnected to parametric optimization of the module components. In this study, optimum footprint ratio of n- and p-type thermoelectric (TE) elements is explored to achieve maximum power generation, maximum cost-perform...
30 CFR 56.19066 - Maximum riders in a conveyance.
2010-07-01
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Maximum riders in a conveyance. 56.19066 Section 56.19066 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR METAL AND... Hoisting Hoisting Procedures § 56.19066 Maximum riders in a conveyance. In shafts inclined over 45...
30 CFR 57.19066 - Maximum riders in a conveyance.
2010-07-01
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Maximum riders in a conveyance. 57.19066 Section 57.19066 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR METAL AND... Hoisting Hoisting Procedures § 57.19066 Maximum riders in a conveyance. In shafts inclined over 45...
Maximum Atmospheric Entry Angle for Specified Retrofire Impulse
T. N. Srivastava
1969-07-01
Full Text Available Maximum atmospheric entry angles for vehicles initially moving in elliptic orbits are investigated and it is shown that tangential retrofire impulse at the apogee results in the maximum entry angle. Equivalence of maximizing the entry angle and minimizing the retrofire impulse is also established.
5 CFR 838.711 - Maximum former spouse survivor annuity.
2010-01-01
... 5 Administrative Personnel 2 2010-01-01 2010-01-01 false Maximum former spouse survivor annuity... Orders Awarding Former Spouse Survivor Annuities Limitations on Survivor Annuities § 838.711 Maximum former spouse survivor annuity. (a) Under CSRS, payments under a court order may not exceed the...
46 CFR 151.45-6 - Maximum amount of cargo.
2010-10-01
... 46 Shipping 5 2010-10-01 2010-10-01 false Maximum amount of cargo. 151.45-6 Section 151.45-6 Shipping COAST GUARD, DEPARTMENT OF HOMELAND SECURITY (CONTINUED) CERTAIN BULK DANGEROUS CARGOES BARGES CARRYING BULK LIQUID HAZARDOUS MATERIAL CARGOES Operations § 151.45-6 Maximum amount of cargo. (a)...
20 CFR 226.52 - Total annuity subject to maximum.
2010-04-01
... rate effective on the date the supplemental annuity begins, before any reduction for a private pension... 20 Employees' Benefits 1 2010-04-01 2010-04-01 false Total annuity subject to maximum. 226.52... COMPUTING EMPLOYEE, SPOUSE, AND DIVORCED SPOUSE ANNUITIES Railroad Retirement Family Maximum § 226.52...
49 CFR 195.406 - Maximum operating pressure.
2010-10-01
... 49 Transportation 3 2010-10-01 2010-10-01 false Maximum operating pressure. 195.406 Section 195... HAZARDOUS LIQUIDS BY PIPELINE Operation and Maintenance § 195.406 Maximum operating pressure. (a) Except for surge pressures and other variations from normal operations, no operator may operate a pipeline at a...
Distribution of maximum loss of fractional Brownian motion with drift
Çağlar, Mine; Vardar-Acar, Ceren
2013-01-01
In this paper, we find bounds on the distribution of the maximum loss of fractional Brownian motion with H >= 1/2 and derive estimates on its tail probability. Asymptotically, the tail of the distribution of maximum loss over [0, t] behaves like the tail of the marginal distribution at time t.
48 CFR 436.575 - Maximum workweek-construction schedule.
2010-10-01
...-construction schedule. 436.575 Section 436.575 Federal Acquisition Regulations System DEPARTMENT OF AGRICULTURE... Maximum workweek-construction schedule. The contracting officer shall insert the clause at 452.236-75, Maximum Workweek-Construction Schedule, if the clause at FAR 52.236-15 is used and the contractor's...
30 CFR 57.5039 - Maximum permissible concentration.
2010-07-01
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Maximum permissible concentration. 57.5039... Maximum permissible concentration. Except as provided by standard § 57.5005, persons shall not be exposed to air containing concentrations of radon daughters exceeding 1.0 WL in active workings. ...
5 CFR 550.105 - Biweekly maximum earnings limitation.
2010-01-01
... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false Biweekly maximum earnings limitation. 550.105 Section 550.105 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PAY ADMINISTRATION (GENERAL) Premium Pay Maximum Earnings Limitations § 550.105 Biweekly...
5 CFR 550.106 - Annual maximum earnings limitation.
2010-01-01
... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false Annual maximum earnings limitation. 550.106 Section 550.106 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PAY ADMINISTRATION (GENERAL) Premium Pay Maximum Earnings Limitations § 550.106 Annual...
32 CFR 842.35 - Depreciation and maximum allowances.
2010-07-01
... 32 National Defense 6 2010-07-01 2010-07-01 false Depreciation and maximum allowances. 842.35... LITIGATION ADMINISTRATIVE CLAIMS Personnel Claims (31 U.S.C. 3701, 3721) § 842.35 Depreciation and maximum allowances. The military services have jointly established the “Allowance List-Depreciation Guide”...
Maximum Principles for Discrete and Semidiscrete Reaction-Diffusion Equation
Petr Stehlík
2015-01-01
Full Text Available We study reaction-diffusion equations with a general reaction function f on one-dimensional lattices with continuous or discrete time ux′ (or Δtux=k(ux-1-2ux+ux+1+f(ux, x∈Z. We prove weak and strong maximum and minimum principles for corresponding initial-boundary value problems. Whereas the maximum principles in the semidiscrete case (continuous time exhibit similar features to those of fully continuous reaction-diffusion model, in the discrete case the weak maximum principle holds for a smaller class of functions and the strong maximum principle is valid in a weaker sense. We describe in detail how the validity of maximum principles depends on the nonlinearity and the time step. We illustrate our results on the Nagumo equation with the bistable nonlinearity.
Experimental study on prediction model for maximum rebound ratio
LEI Wei-dong; TENG Jun; A.HEFNY; ZHAO Jian; GUAN Jiong
2007-01-01
The proposed prediction model for estimating the maximum rebound ratio was applied to a field explosion test, Mandai test in Singapore.The estimated possible maximum Deak particle velocities(PPVs)were compared with the field records.Three of the four available field-recorded PPVs lie exactly below the estimated possible maximum values as expected.while the fourth available field-recorded PPV lies close to and a bit higher than the estimated maximum possible PPV The comparison results show that the predicted PPVs from the proposed prediction model for the maximum rebound ratio match the field.recorded PPVs better than those from two empirical formulae.The very good agreement between the estimated and field-recorded values validates the proposed prediction model for estimating PPV in a rock mass with a set of ipints due to application of a two dimensional compressional wave at the boundary of a tunnel or a borehole.
Maximum likelihood estimation for semiparametric density ratio model.
Diao, Guoqing; Ning, Jing; Qin, Jing
2012-06-27
In the statistical literature, the conditional density model specification is commonly used to study regression effects. One attractive model is the semiparametric density ratio model, under which the conditional density function is the product of an unknown baseline density function and a known parametric function containing the covariate information. This model has a natural connection with generalized linear models and is closely related to biased sampling problems. Despite the attractive features and importance of this model, most existing methods are too restrictive since they are based on multi-sample data or conditional likelihood functions. The conditional likelihood approach can eliminate the unknown baseline density but cannot estimate it. We propose efficient estimation procedures based on the nonparametric likelihood. The nonparametric likelihood approach allows for general forms of covariates and estimates the regression parameters and the baseline density simultaneously. Therefore, the nonparametric likelihood approach is more versatile than the conditional likelihood approach especially when estimation of the conditional mean or other quantities of the outcome is of interest. We show that the nonparametric maximum likelihood estimators are consistent, asymptotically normal, and asymptotically efficient. Simulation studies demonstrate that the proposed methods perform well in practical settings. A real example is used for illustration.
On a robust and efficient maximum depth estimator
ZUO YiJun; LAI ShaoYong
2009-01-01
The best breakdown point robustness is one of the most outstanding features of the univariate median. For this robustness property, the median, however, has to pay the price of a low efficiency at normal and other light-tailed models. Affine equivariant multivariate analogues of the univariate median with high breakdown points were constructed in the past two decades. For the high breakdown robustness, most of them also have to sacrifice their efficiency at normal and other models,nevertheless. The affine equivariant maximum depth estimator proposed and studied in this paper turns out to be an exception. Like the univariate median, it also possesses a highest breakdown point among all its multivariate competitors. Unlike the univariate median, it is also highly efficient relative to the sample mean at normal and various other distributions, overcoming the vital low-efficiency shortcoming of the univariate and other multivariate generalized medians. The paper also studies the asymptotics of the estimator and establishes its limit distribution without symmetry and other strong assumptions that are typically imposed on the underlying distribution.
Maximum covariance analysis to identify intraseasonal oscillations over tropical Brazil
Barreto, Naurinete J. C.; Mesquita, Michel d. S.; Mendes, David; Spyrides, Maria H. C.; Pedra, George U.; Lucio, Paulo S.
2017-09-01
A reliable prognosis of extreme precipitation events in the tropics is arguably challenging to obtain due to the interaction of meteorological systems at various time scales. A pivotal component of the global climate variability is the so-called intraseasonal oscillations, phenomena that occur between 20 and 100 days. The Madden-Julian Oscillation (MJO), which is directly related to the modulation of convective precipitation in the equatorial belt, is considered the primary oscillation in the tropical region. The aim of this study is to diagnose the connection between the MJO signal and the regional intraseasonal rainfall variability over tropical Brazil. This is achieved through the development of an index called Multivariate Intraseasonal Index for Tropical Brazil (MITB). This index is based on Maximum Covariance Analysis (MCA) applied to the filtered daily anomalies of rainfall data over tropical Brazil against a group of covariates consisting of: outgoing longwave radiation and the zonal component u of the wind at 850 and 200 hPa. The first two MCA modes, which were used to create the { MITB}_1 and { MITB}_2 indices, represent 65 and 16 % of the explained variance, respectively. The combined multivariate index was able to satisfactorily represent the pattern of intraseasonal variability over tropical Brazil, showing that there are periods of activation and inhibition of precipitation connected with the pattern of MJO propagation. The MITB index could potentially be used as a diagnostic tool for intraseasonal forecasting.
Constructing Maximum Entropy Language Models for Movie Review Subjectivity Analysis
Bo Chen; Hui He; Jun Guo
2008-01-01
Document subjectivity analysis has become an important aspect of web text content mining. This problem is similar to traditional text categorization, thus many related classification techniques can be adapted here. However, there is one significant difference that more language or semantic information is required for better estimating the subjectivity of a document. Therefore, in this paper, our focuses are mainly on two aspects. One is how to extract useful and meaningful language features, and the other is how to construct appropriate language models efficiently for this special task. For the first issue, we conduct a Global-Filtering and Local-Weighting strategy to select and evaluate language features in a series of n-grams with different orders and within various distance-windows. For the second issue, we adopt Maximum Entropy (MaxEnt) modeling methods to construct our language model framework. Besides the classical MaxEnt models, we have also constructed two kinds of improved models with Gaussian and exponential priors respectively. Detailed experiments given in this paper show that with well selected and weighted language features, MaxEnt models with exponential priors are significantly more suitable for the text subjectivity analysis task.
tmle : An R Package for Targeted Maximum Likelihood Estimation
Susan Gruber
2012-11-01
Full Text Available Targeted maximum likelihood estimation (TMLE is a general approach for constructing an efficient double-robust semi-parametric substitution estimator of a causal effect parameter or statistical association measure. tmle is a recently developed R package that implements TMLE of the effect of a binary treatment at a single point in time on an outcome of interest, controlling for user supplied covariates, including an additive treatment effect, relative risk, odds ratio, and the controlled direct effect of a binary treatment controlling for a binary intermediate variable on the pathway from treatment to the out- come. Estimation of the parameters of a marginal structural model is also available. The package allows outcome data with missingness, and experimental units that contribute repeated records of the point-treatment data structure, thereby allowing the analysis of longitudinal data structures. Relevant factors of the likelihood may be modeled or fit data-adaptively according to user specifications, or passed in from an external estimation procedure. Effect estimates, variances, p values, and 95% confidence intervals are provided by the software.
Maximum-likelihood estimation of circle parameters via convolution.
Zelniker, Emanuel E; Clarkson, I Vaughan L
2006-04-01
The accurate fitting of a circle to noisy measurements of circumferential points is a much studied problem in the literature. In this paper, we present an interpretation of the maximum-likelihood estimator (MLE) and the Delogne-Kåsa estimator (DKE) for circle-center and radius estimation in terms of convolution on an image which is ideal in a certain sense. We use our convolution-based MLE approach to find good estimates for the parameters of a circle in digital images. In digital images, it is then possible to treat these estimates as preliminary estimates into various other numerical techniques which further refine them to achieve subpixel accuracy. We also investigate the relationship between the convolution of an ideal image with a "phase-coded kernel" (PCK) and the MLE. This is related to the "phase-coded annulus" which was introduced by Atherton and Kerbyson who proposed it as one of a number of new convolution kernels for estimating circle center and radius. We show that the PCK is an approximate MLE (AMLE). We compare our AMLE method to the MLE and the DKE as well as the Cramér-Rao Lower Bound in ideal images and in both real and synthetic digital images.
2010-07-01
... as specified in 40 CFR 1065.610. This is the maximum in-use engine speed used for calculating the NOX... procedures of 40 CFR part 1065, based on the manufacturer's design and production specifications for the..., power density, and maximum in-use engine speed. 1042.140 Section 1042.140 Protection of...
Genei Antonio Dalmago
2006-06-01
Full Text Available Um experimento foi conduzido para avaliar a relação da evapotranspiração máxima (ETm da cultura de pimentão por unidade de índice de área foliar (ETmf com a radiação solar global incidente (Rg e e saldo de radiação (Rn e, externos à estufa, e com a temperatura (t am, umidade relativa (URm e déficit de saturação (D do ar no interior da estufa. A ETm foi determinada por lisimetria. Apenas a temperatura e a umidade relativa do ar foram registradas e as demais variáveis foram estimadas. O Rn e apresentou o maior efeito isolado, seguido pelo D e pela URm. A temperatura do ar às 15 horas e a temperatura máxima diária foram as variáveis pontuais de maior associação com a ETmf. A Rg e melhorou sua relação quando analisada em diferentes faixas de t am. Os resultados confirmaram o Rn e e o D como as variáveis meteorológicas de maior efeito preditivo da evapotranspiração das culturas em estufas plásticas no outono, porém com graus diferenciados de ajuste para o pimentão em relação às outras culturas.An experiment was carried out to study the relationship between sweet pepper maximum evapotranspiration (ETm per unit leaf area index (ETmf and incident solar radiation (Rg e and net radiation (Rn e outside greenhouse, and with air temperature (t am, air relative humidity (URm and water vapor pressure deficit (D inside a greenhouse. ETm was measured by lysimeters. Temperature and relative humidity were registered whereas Rg e and Rn e were estimated. The Rn e was the variable with laRg e st effect on ETmf determination, followed by D and URm. The air temperature at 15:00 and the daily maximum temperature were variables of laRg e r association with ETmf among the variables with punctual time observations. Rg e improved its relationship with ETmf when was analyzed with different t am ranges. The results confirmed Rn e and D as the meteorological variables of highest relationship with evapotranspiration in plastic greenhouse
Understanding the Role of Reservoir Size on Probable Maximum Precipitation
Woldemichael, A. T.; Hossain, F.
2011-12-01
This study addresses the question 'Does surface area of an artificial reservoir matter in the estimation of probable maximum precipitation (PMP) for an impounded basin?' The motivation of the study was based on the notion that the stationarity assumption that is implicit in the PMP for dam design can be undermined in the post-dam era due to an enhancement of extreme precipitation patterns by an artificial reservoir. In addition, the study lays the foundation for use of regional atmospheric models as one way to perform life cycle assessment for planned or existing dams to formulate best management practices. The American River Watershed (ARW) with the Folsom dam at the confluence of the American River was selected as the study region and the Dec-Jan 1996-97 storm event was selected for the study period. The numerical atmospheric model used for the study was the Regional Atmospheric Modeling System (RAMS). First, the numerical modeling system, RAMS, was calibrated and validated with selected station and spatially interpolated precipitation data. Best combinations of parameterization schemes in RAMS were accordingly selected. Second, to mimic the standard method of PMP estimation by moisture maximization technique, relative humidity terms in the model were raised to 100% from ground up to the 500mb level. The obtained model-based maximum 72-hr precipitation values were named extreme precipitation (EP) as a distinction from the PMPs obtained by the standard methods. Third, six hypothetical reservoir size scenarios ranging from no-dam (all-dry) to the reservoir submerging half of basin were established to test the influence of reservoir size variation on EP. For the case of the ARW, our study clearly demonstrated that the assumption of stationarity that is implicit the traditional estimation of PMP can be rendered invalid to a large part due to the very presence of the artificial reservoir. Cloud tracking procedures performed on the basin also give indication of the
Leaf Dynamics of Panicum maximum under Future Climatic Changes.
Carlos Henrique Britto de Assis Prado
Full Text Available Panicum maximum Jacq. 'Mombaça' (C4 was grown in field conditions with sufficient water and nutrients to examine the effects of warming and elevated CO2 concentrations during the winter. Plants were exposed to either the ambient temperature and regular atmospheric CO2 (Control; elevated CO2 (600 ppm, eC; canopy warming (+2°C above regular canopy temperature, eT; or elevated CO2 and canopy warming (eC+eT. The temperatures and CO2 in the field were controlled by temperature free-air controlled enhancement (T-FACE and mini free-air CO2 enrichment (miniFACE facilities. The most green, expanding, and expanded leaves and the highest leaf appearance rate (LAR, leaves day(-1 and leaf elongation rate (LER, cm day(-1 were observed under eT. Leaf area and leaf biomass were higher in the eT and eC+eT treatments. The higher LER and LAR without significant differences in the number of senescent leaves could explain why tillers had higher foliage area and leaf biomass in the eT treatment. The eC treatment had the lowest LER and the fewest expanded and green leaves, similar to Control. The inhibitory effect of eC on foliage development in winter was indicated by the fewer green, expanded, and expanding leaves under eC+eT than eT. The stimulatory and inhibitory effects of the eT and eC treatments, respectively, on foliage raised and lowered, respectively, the foliar nitrogen concentration. The inhibition of foliage by eC was confirmed by the eC treatment having the lowest leaf/stem biomass ratio and by the change in leaf biomass-area relationships from linear or exponential growth to rectangular hyperbolic growth under eC. Besides, eC+eT had a synergist effect, speeding up leaf maturation. Therefore, with sufficient water and nutrients in winter, the inhibitory effect of elevated CO2 on foliage could be partially offset by elevated temperatures and relatively high P. maximum foliage production could be achieved under future climatic change.
Leaf Dynamics of Panicum maximum under Future Climatic Changes.
Britto de Assis Prado, Carlos Henrique; Haik Guedes de Camargo-Bortolin, Lívia; Castro, Érique; Martinez, Carlos Alberto
2016-01-01
Panicum maximum Jacq. 'Mombaça' (C4) was grown in field conditions with sufficient water and nutrients to examine the effects of warming and elevated CO2 concentrations during the winter. Plants were exposed to either the ambient temperature and regular atmospheric CO2 (Control); elevated CO2 (600 ppm, eC); canopy warming (+2°C above regular canopy temperature, eT); or elevated CO2 and canopy warming (eC+eT). The temperatures and CO2 in the field were controlled by temperature free-air controlled enhancement (T-FACE) and mini free-air CO2 enrichment (miniFACE) facilities. The most green, expanding, and expanded leaves and the highest leaf appearance rate (LAR, leaves day(-1)) and leaf elongation rate (LER, cm day(-1)) were observed under eT. Leaf area and leaf biomass were higher in the eT and eC+eT treatments. The higher LER and LAR without significant differences in the number of senescent leaves could explain why tillers had higher foliage area and leaf biomass in the eT treatment. The eC treatment had the lowest LER and the fewest expanded and green leaves, similar to Control. The inhibitory effect of eC on foliage development in winter was indicated by the fewer green, expanded, and expanding leaves under eC+eT than eT. The stimulatory and inhibitory effects of the eT and eC treatments, respectively, on foliage raised and lowered, respectively, the foliar nitrogen concentration. The inhibition of foliage by eC was confirmed by the eC treatment having the lowest leaf/stem biomass ratio and by the change in leaf biomass-area relationships from linear or exponential growth to rectangular hyperbolic growth under eC. Besides, eC+eT had a synergist effect, speeding up leaf maturation. Therefore, with sufficient water and nutrients in winter, the inhibitory effect of elevated CO2 on foliage could be partially offset by elevated temperatures and relatively high P. maximum foliage production could be achieved under future climatic change.
Yeast hnRNP-related proteins contribute to the maintenance of telomeres
Lee-Soety, Julia Y., E-mail: jlee04@sju.edu [Department of Biology, Saint Joseph' s University, PA 19131 (United States); Jones, Jennifer; MacGibeny, Margaret A.; Remaly, Erin C.; Daniels, Lynsey; Ito, Andrea; Jean, Jessica; Radecki, Hannah; Spencer, Shannon [Department of Biology, Saint Joseph' s University, PA 19131 (United States)
2012-09-14
Highlights: Black-Right-Pointing-Pointer Yeast hnRNP-related proteins are able to prevent faster senescence in telomerase-null cells. Black-Right-Pointing-Pointer The conserved RRMs in Npl3 are important for telomere maintenance. Black-Right-Pointing-Pointer Human hnRNP A1 is unable to complement the lack of NPL3 in yeast. Black-Right-Pointing-Pointer Npl3 and Cbc2 may work as telomere capping proteins. -- Abstract: Telomeres protect the ends of linear chromosomes, which if eroded to a critical length can become uncapped and lead to replicative senescence. Telomerase maintains telomere length in some cells, but inappropriate expression facilitates the immortality of cancer cells. Recently, proteins involved in RNA processing and ribosome assembly, such as hnRNP (heterogeneous nuclear ribonucleoprotein) A1, have been found to participate in telomere maintenance in mammals. The Saccharomyces cerevisiae protein Npl3 shares significant amino acid sequence similarities with hnRNP A1. We found that deleting NPL3 accelerated the senescence of telomerase null cells. The highly conserved RNA recognition motifs (RRM) in Npl3 appear to be important for preventing faster senescence. Npl3 preferentially binds telomere sequences in vitro, suggesting that Npl3 may affect telomeres directly. Despite similarities between the two proteins, human hnRNP A1 is unable to complement the lack of Npl3 to rescue accelerated senescence in tlc1 npl3 cells. Deletion of CBC2, which encodes another hnRNP-related protein that associates with Npl3, also accelerates senescence. Potential mechanisms by which hnRNP-related proteins maintain telomeres are discussed.
Potential role of motion for enhancing maximum output energy of triboelectric nanogenerator
Byun, Kyung-Eun; Lee, Min-Hyun; Cho, Yeonchoo; Nam, Seung-Geol; Shin, Hyeon-Jin; Park, Seongjun
2017-07-01
Although triboelectric nanogenerator (TENG) has been explored as one of the possible candidates for the auxiliary power source of portable and wearable devices, the output energy of a TENG is still insufficient to charge the devices with daily motion. Moreover, the fundamental aspects of the maximum possible energy of a TENG related with human motion are not understood systematically. Here, we confirmed the possibility of charging commercialized portable and wearable devices such as smart phones and smart watches by utilizing the mechanical energy generated by human motion. We confirmed by theoretical extraction that the maximum possible energy is related with specific form factors of a TENG. Furthermore, we experimentally demonstrated the effect of human motion in an aspect of the kinetic energy and impulse using varying velocity and elasticity, and clarified how to improve the maximum possible energy of a TENG. This study gives insight into design of a TENG to obtain a large amount of energy in a limited space.
Stone, Wesley W.; Gilliom, Robert J.; Crawford, Charles G.
2008-01-01
Regression models were developed for predicting annual maximum and selected annual maximum moving-average concentrations of atrazine in streams using the Watershed Regressions for Pesticides (WARP) methodology developed by the National Water-Quality Assessment Program (NAWQA) of the U.S. Geological Survey (USGS). The current effort builds on the original WARP models, which were based on the annual mean and selected percentiles of the annual frequency distribution of atrazine concentrations. Estimates of annual maximum and annual maximum moving-average concentrations for selected durations are needed to characterize the levels of atrazine and other pesticides for comparison to specific water-quality benchmarks for evaluation of potential concerns regarding human health or aquatic life. Separate regression models were derived for the annual maximum and annual maximum 21-day, 60-day, and 90-day moving-average concentrations. Development of the regression models used the same explanatory variables, transformations, model development data, model validation data, and regression methods as those used in the original development of WARP. The models accounted for 72 to 75 percent of the variability in the concentration statistics among the 112 sampling sites used for model development. Predicted concentration statistics from the four models were within a factor of 10 of the observed concentration statistics for most of the model development and validation sites. Overall, performance of the models for the development and validation sites supports the application of the WARP models for predicting annual maximum and selected annual maximum moving-average atrazine concentration in streams and provides a framework to interpret the predictions in terms of uncertainty. For streams with inadequate direct measurements of atrazine concentrations, the WARP model predictions for the annual maximum and the annual maximum moving-average atrazine concentrations can be used to characterize
Site Specific Probable Maximum Precipitation Estimates and Professional Judgement
Hayes, B. D.; Kao, S. C.; Kanney, J. F.; Quinlan, K. R.; DeNeale, S. T.
2015-12-01
State and federal regulatory authorities currently rely upon the US National Weather Service Hydrometeorological Reports (HMRs) to determine probable maximum precipitation (PMP) estimates (i.e., rainfall depths and durations) for estimating flooding hazards for relatively broad regions in the US. PMP estimates for the contributing watersheds upstream of vulnerable facilities are used to estimate riverine flooding hazards while site-specific estimates for small water sheds are appropriate for individual facilities such as nuclear power plants. The HMRs are often criticized due to their limitations on basin size, questionable applicability in regions affected by orographic effects, their lack of consist methods, and generally by their age. HMR-51 for generalized PMP estimates for the United States east of the 105th meridian, was published in 1978 and is sometimes perceived as overly conservative. The US Nuclear Regulatory Commission (NRC), is currently reviewing several flood hazard evaluation reports that rely on site specific PMP estimates that have been commercially developed. As such, NRC has recently investigated key areas of expert judgement via a generic audit and one in-depth site specific review as they relate to identifying and quantifying actual and potential storm moisture sources, determining storm transposition limits, and adjusting available moisture during storm transposition. Though much of the approach reviewed was considered a logical extension of HMRs, two key points of expert judgement stood out for further in-depth review. The first relates primarily to small storms and the use of a heuristic for storm representative dew point adjustment developed for the Electric Power Research Institute by North American Weather Consultants in 1993 in order to harmonize historic storms for which only 12 hour dew point data was available with more recent storms in a single database. The second issue relates to the use of climatological averages for spatially
Maximum Likelihood Estimation of the Identification Parameters and Its Correction
无
2002-01-01
By taking the subsequence out of the input-output sequence of a system polluted by white noise, anindependent observation sequence and its probability density are obtained and then a maximum likelihood estimation of theidentification parameters is given. In order to decrease the asymptotic error, a corrector of maximum likelihood (CML)estimation with its recursive algorithm is given. It has been proved that the corrector has smaller asymptotic error thanthe least square methods. A simulation example shows that the corrector of maximum likelihood estimation is of higherapproximating precision to the true parameters than the least square methods.
Maximum frequency of the decametric radiation from Jupiter
Barrow, C. H.; Alexander, J. K.
1980-01-01
The upper frequency limits of Jupiter's decametric radio emission are found to be essentially the same when observed from the earth or, with considerably higher sensitivity, from the Voyager spacecraft close to Jupiter. This suggests that the maximum frequency is a real cut-off corresponding to a maximum gyrofrequency of about 38-40 MHz at Jupiter. It no longer appears to be necessary to specify different cut-off frequencies for the Io and non-Io emission as the maximum frequencies are roughly the same in each case.
李化; 杨新春; 李剑; 陈娇; 程昌奎
2012-01-01
抑制干扰是GIS局部放电在线监测的关键技术之一。尽管局部放电超高频检测方法能够有效避开低频干扰，但来自测量系统的白噪声仍然为准确测量局部放电带来困难。为有效抑制白噪声，提高局部放电超高频法的测量精度，本文提出一种用于GIS局部放电超高频信号的自适应小波分解去噪算法，该算法基于每层小波分解尺度系数能量最大的原则，逐层自适应选取最优的小波进行分解，并结合Donoho提出的软阈值法进行去噪。对人工绝缘缺陷产生的四种GIS超高频信号的去噪结果证明了该算法较其他小波算法能更好地去除白噪声且去噪后信号波形畸变较小，具有很好的应用前景。%Interference suppression was one of the key technologies in on-line partial discharge（PD） monitoring of gas insulated switchgear（GIS）. Although ultra-high-frequency （UHF） is qualified to avoid low-frequency noises, the system white noise from the high voltage transmission line still make it difficult to accurately measure the level of PD. For active inhibition of the white noise interference and improving the precision of the UHF detection methods, this paper presents a adaptive de-noising scheme, which is suitable for de-noising UHF signal detected by the UHF detection system of PD in GIS. The method utilizes various basic wavelet to decompose a signal, and calculate and compare the signal energies caused by decomposition using different wavelets in each scale. The basic wavelet corresponding to the maximum signal energy is considered as the optimum wavelet in the current scale, thus the optimum wavelets family of all the scales is obtained, and the soft threshold function presented by Donoho is used to de-nosing. The result of de-noising a UHF signal generated by an artificial insulation defect convinces that the adaptive wavelet de-noising method is more effective to suppress the white noise mixed in
The Application of Maximum Principle in Supply Chain Cost Optimization
Zhou Ling; Wang Jun
2013-01-01
In this paper, using the maximum principle for analyzing dynamic cost, we propose a new two-stage supply chain model of the manufacturing-assembly mode for high-tech perishable products supply chain...
Maximum Principle for Nonlinear Cooperative Elliptic Systems on IR N
LEADI Liamidi; MARCOS Aboubacar
2011-01-01
We investigate in this work necessary and sufficient conditions for having a Maximum Principle for a cooperative elliptic system on the whole (IR)N.Moreover,we prove the existence of solutions by an approximation method for the considered system.
Maximum Likelihood Factor Structure of the Family Environment Scale.
Fowler, Patrick C.
1981-01-01
Presents the maximum likelihood factor structure of the Family Environment Scale. The first bipolar dimension, "cohesion v conflict," measures relationship-centered concerns, while the second unipolar dimension is an index of "organizational and control" activities. (Author)
Multiresolution maximum intensity volume rendering by morphological adjunction pyramids
Roerdink, Jos B.T.M.
We describe a multiresolution extension to maximum intensity projection (MIP) volume rendering, allowing progressive refinement and perfect reconstruction. The method makes use of morphological adjunction pyramids. The pyramidal analysis and synthesis operators are composed of morphological 3-D
Multiresolution Maximum Intensity Volume Rendering by Morphological Adjunction Pyramids
Roerdink, Jos B.T.M.
2001-01-01
We describe a multiresolution extension to maximum intensity projection (MIP) volume rendering, allowing progressive refinement and perfect reconstruction. The method makes use of morphological adjunction pyramids. The pyramidal analysis and synthesis operators are composed of morphological 3-D
Changes in context and perception of maximum reaching height.
Wagman, Jeffrey B; Day, Brian M
2014-01-01
Successfully performing a given behavior requires flexibility in both perception and behavior. In particular, doing so requires perceiving whether that behavior is possible across the variety of contexts in which it might be performed. Three experiments investigated how (changes in) context (ie point of observation and intended reaching task) influenced perception of maximum reaching height. The results of experiment 1 showed that perceived maximum reaching height more closely reflected actual reaching ability when perceivers occupied a point of observation that was compatible with that required for the reaching task. The results of experiments 2 and 3 showed that practice perceiving maximum reaching height from a given point of observation improved perception of maximum reaching height from a different point of observation, regardless of whether such practice occurred at a compatible or incompatible point of observation. In general, such findings show bounded flexibility in perception of affordances and are thus consistent with a description of perceptual systems as smart perceptual devices.
Water Quality Assessment and Total Maximum Daily Loads Information (ATTAINS)
U.S. Environmental Protection Agency — The Water Quality Assessment TMDL Tracking And Implementation System (ATTAINS) stores and tracks state water quality assessment decisions, Total Maximum Daily Loads...
Combining Experiments and Simulations Using the Maximum Entropy Principle
Boomsma, Wouter; Ferkinghoff-Borg, Jesper; Lindorff-Larsen, Kresten
2014-01-01
are not in quantitative agreement with experimental data. The principle of maximum entropy is a general procedure for constructing probability distributions in the light of new data, making it a natural tool in cases when an initial model provides results that are at odds with experiments. The number of maximum entropy...... in the context of a simple example, after which we proceed with a real-world application in the field of molecular simulations, where the maximum entropy procedure has recently provided new insight. Given the limited accuracy of force fields, macromolecular simulations sometimes produce results....... Three very recent papers have explored this problem using the maximum entropy approach, providing both new theoretical and practical insights to the problem. We highlight each of these contributions in turn and conclude with a discussion on remaining challenges....
On the sufficiency of the linear maximum principle
Vidal, Rene Victor Valqui
1987-01-01
Presents a family of linear maximum principles for the discrete-time optimal control problem, derived from the saddle-point theorem of mathematical programming. Some simple examples illustrate the applicability of the main theoretical results...
Maximum Photovoltaic Penetration Levels on Typical Distribution Feeders: Preprint
Hoke, A.; Butler, R.; Hambrick, J.; Kroposki, B.
2012-07-01
This paper presents simulation results for a taxonomy of typical distribution feeders with various levels of photovoltaic (PV) penetration. For each of the 16 feeders simulated, the maximum PV penetration that did not result in steady-state voltage or current violation is presented for several PV location scenarios: clustered near the feeder source, clustered near the midpoint of the feeder, clustered near the end of the feeder, randomly located, and evenly distributed. In addition, the maximum level of PV is presented for single, large PV systems at each location. Maximum PV penetration was determined by requiring that feeder voltages stay within ANSI Range A and that feeder currents stay within the ranges determined by overcurrent protection devices. Simulations were run in GridLAB-D using hourly time steps over a year with randomized load profiles based on utility data and typical meteorological year weather data. For 86% of the cases simulated, maximum PV penetration was at least 30% of peak load.
16 CFR 1505.8 - Maximum acceptable material temperatures.
2010-01-01
... Association, 155 East 44th Street, New York, NY 10017. Material Degrees C. Degrees F. Capacitors (1) (1) Class... capacitor has no marked temperature limit, the maximum acceptable temperature will be assumed to be 65...
Environmental Monitoring, Water Quality - Total Maximum Daily Load (TMDL)
NSGIC GIS Inventory (aka Ramona) — The Clean Water Act Section 303(d) establishes the Total Maximum Daily Load (TMDL) program. The purpose of the TMDL program is to identify sources of pollution and...
Environmental Monitoring, Water Quality - Total Maximum Daily Load (TMDL)
NSGIC Education | GIS Inventory — The Clean Water Act Section 303(d) establishes the Total Maximum Daily Load (TMDL) program. The purpose of the TMDL program is to identify sources of pollution and...
Solar Panel Maximum Power Point Tracker for Power Utilities
Sandeep Banik,
2014-01-01
Full Text Available ―Solar Panel Maximum Power Point Tracker For power utilities‖ As the name implied, it is a photovoltaic system that uses the photovoltaic array as a source of electrical power supply and since every photovoltaic (PV array has an optimum operating point, called the maximum power point, which varies depending on the insolation level and array voltage. A maximum power point tracker (MPPT is needed to operate the PV array at its maximum power point. The objective of this thesis project is to build a photovoltaic (PV array Of 121.6V DC Voltage(6 cell each 20V, 100watt And convert the DC voltage to Single phase 120v,50Hz AC voltage by switch mode power converter‘s and inverter‘s.
A Family of Maximum SNR Filters for Noise Reduction
Huang, Gongping; Benesty, Jacob; Long, Tao;
2014-01-01
This paper is devoted to the study and analysis of the maximum signal-to-noise ratio (SNR) filters for noise reduction both in the time and short-time Fourier transform (STFT) domains with one single microphone and multiple microphones. In the time domain, we show that the maximum SNR filters can...... significantly increase the SNR but at the expense of tremendous speech distortion. As a consequence, the speech quality improvement, measured by the perceptual evaluation of speech quality (PESQ) algorithm, is marginal if any, regardless of the number of microphones used. In the STFT domain, the maximum SNR....... This demonstrates that the maximum SNR filters, particularly the multichannel ones, in the STFT domain may be of great practical value....
Maximum likelihood estimation of finite mixture model for economic data
Phoong, Seuk-Yen; Ismail, Mohd Tahir
2014-06-01
Finite mixture model is a mixture model with finite-dimension. This models are provides a natural representation of heterogeneity in a finite number of latent classes. In addition, finite mixture models also known as latent class models or unsupervised learning models. Recently, maximum likelihood estimation fitted finite mixture models has greatly drawn statistician's attention. The main reason is because maximum likelihood estimation is a powerful statistical method which provides consistent findings as the sample sizes increases to infinity. Thus, the application of maximum likelihood estimation is used to fit finite mixture model in the present paper in order to explore the relationship between nonlinear economic data. In this paper, a two-component normal mixture model is fitted by maximum likelihood estimation in order to investigate the relationship among stock market price and rubber price for sampled countries. Results described that there is a negative effect among rubber price and stock market price for Malaysia, Thailand, Philippines and Indonesia.
Malone, Stephen M.; McGue, Matt; Iacono, William G.
2010-01-01
Background: The maximum number of alcoholic drinks consumed in a single 24-hr period is an alcoholism-related phenotype with both face and empirical validity. It has been associated with severity of withdrawal symptoms and sensitivity to alcohol, genes implicated in alcohol metabolism, and amplitude of a measure of brain activity associated with…
MAXIMUM PRINCIPLES OF NONHOMOGENEOUS SUBELLIPTIC P-LAPLACE EQUATIONS AND APPLICATIONS
Liu Haifeng; Niu Pengcheng
2006-01-01
Maximum principles for weak solutions of nonhomogeneous subelliptic p-Laplace equations related to smooth vector fields {Xj} satisfying the H(o)rmander condition are proved by the choice of suitable test functions and the adaption of the classical Moser iteration method. Some applications are given in this paper.
Lagrange Multipliers, Adjoint Equations, the Pontryagin Maximum Principle and Heuristic Proofs
Ollerton, Richard L.
2013-01-01
Deeper understanding of important mathematical concepts by students may be promoted through the (initial) use of heuristic proofs, especially when the concepts are also related back to previously encountered mathematical ideas or tools. The approach is illustrated by use of the Pontryagin maximum principle which is then illuminated by reference to…
Jingtao Shi
2013-01-01
Full Text Available This paper is concerned with the relationship between maximum principle and dynamic programming for stochastic recursive optimal control problems. Under certain differentiability conditions, relations among the adjoint processes, the generalized Hamiltonian function, and the value function are given. A linear quadratic recursive utility portfolio optimization problem in the financial engineering is discussed as an explicitly illustrated example of the main result.
24 CFR 235.331 - Increased maximum mortgage amount for physically handicapped persons.
2010-04-01
... for physically handicapped persons. 235.331 Section 235.331 Housing and Urban Development Regulations... maximum mortgage amount for physically handicapped persons. If the mortgage relates to a dwelling unit to be occupied by a handicapped person as defined in § 235.5(c)(2), the otherwise applicable...
Maximum Likelihood Analysis of a Two-Level Nonlinear Structural Equation Model with Fixed Covariates
Lee, Sik-Yum; Song, Xin-Yuan
2005-01-01
In this article, a maximum likelihood (ML) approach for analyzing a rather general two-level structural equation model is developed for hierarchically structured data that are very common in educational and/or behavioral research. The proposed two-level model can accommodate nonlinear causal relations among latent variables as well as effects…
Constrained Maximum Likelihood Estimation for Two-Level Mean and Covariance Structure Models
Bentler, Peter M.; Liang, Jiajuan; Tang, Man-Lai; Yuan, Ke-Hai
2011-01-01
Maximum likelihood is commonly used for the estimation of model parameters in the analysis of two-level structural equation models. Constraints on model parameters could be encountered in some situations such as equal factor loadings for different factors. Linear constraints are the most common ones and they are relatively easy to handle in…
On the maximum sufficient range of interstellar vessels
Cartin, Daniel
2011-01-01
This paper considers the likely maximum range of space vessels providing the basis of a mature interstellar transportation network. Using the principle of sufficiency, it is argued that this range will be less than three parsecs for the average interstellar vessel. This maximum range provides access from the Solar System to a large majority of nearby stellar systems, with total travel distances within the network not excessively greater than actual physical distance.
Efficiency at Maximum Power of Interacting Molecular Machines
Golubeva, Natalia; Imparato, Alberto
2012-01-01
We investigate the efficiency of systems of molecular motors operating at maximum power. We consider two models of kinesin motors on a microtubule: for both the simplified and the detailed model, we find that the many-body exclusion effect enhances the efficiency at maximum power of the many- motor...... system, with respect to the single motor case. Remarkably, we find that this effect occurs in a limited region of the system parameters, compatible with the biologically relevant range....
Filtering Additive Measurement Noise with Maximum Entropy in the Mean
Gzyl, Henryk
2007-01-01
The purpose of this note is to show how the method of maximum entropy in the mean (MEM) may be used to improve parametric estimation when the measurements are corrupted by large level of noise. The method is developed in the context on a concrete example: that of estimation of the parameter in an exponential distribution. We compare the performance of our method with the bayesian and maximum likelihood approaches.
The maximum entropy production principle: two basic questions.
Martyushev, Leonid M
2010-05-12
The overwhelming majority of maximum entropy production applications to ecological and environmental systems are based on thermodynamics and statistical physics. Here, we discuss briefly maximum entropy production principle and raises two questions: (i) can this principle be used as the basis for non-equilibrium thermodynamics and statistical mechanics and (ii) is it possible to 'prove' the principle? We adduce one more proof which is most concise today.
A tropospheric ozone maximum over the equatorial Southern Indian Ocean
L. Zhang
2012-05-01
Full Text Available We examine the distribution of tropical tropospheric ozone (O_{3} from the Microwave Limb Sounder (MLS and the Tropospheric Emission Spectrometer (TES by using a global three-dimensional model of tropospheric chemistry (GEOS-Chem. MLS and TES observations of tropospheric O_{3} during 2005 to 2009 reveal a distinct, persistent O_{3} maximum, both in mixing ratio and tropospheric column, in May over the Equatorial Southern Indian Ocean (ESIO. The maximum is most pronounced in 2006 and 2008 and less evident in the other three years. This feature is also consistent with the total column O_{3} observations from the Ozone Mapping Instrument (OMI and the Atmospheric Infrared Sounder (AIRS. Model results reproduce the observed May O_{3} maximum and the associated interannual variability. The origin of the maximum reflects a complex interplay of chemical and dynamic factors. The O_{3} maximum is dominated by the O_{3} production driven by lightning nitrogen oxides (NO_{x} emissions, which accounts for 62% of the tropospheric column O_{3} in May 2006. We find the contribution from biomass burning, soil, anthropogenic and biogenic sources to the O_{3} maximum are rather small. The O_{3} productions in the lightning outflow from Central Africa and South America both peak in May and are directly responsible for the O_{3} maximum over the western ESIO. The lightning outflow from Equatorial Asia dominates over the eastern ESIO. The interannual variability of the O_{3} maximum is driven largely by the anomalous anti-cyclones over the southern Indian Ocean in May 2006 and 2008. The lightning outflow from Central Africa and South America is effectively entrained by the anti-cyclones followed by northward transport to the ESIO.
On the sufficiency of the linear maximum principle
Vidal, Rene Victor Valqui
1987-01-01
Presents a family of linear maximum principles for the discrete-time optimal control problem, derived from the saddle-point theorem of mathematical programming. Some simple examples illustrate the applicability of the main theoretical results......Presents a family of linear maximum principles for the discrete-time optimal control problem, derived from the saddle-point theorem of mathematical programming. Some simple examples illustrate the applicability of the main theoretical results...
Semidefinite Programming for Approximate Maximum Likelihood Sinusoidal Parameter Estimation
2009-01-01
We study the convex optimization approach for parameter estimation of several sinusoidal models, namely, single complex/real tone, multiple complex sinusoids, and single two-dimensional complex tone, in the presence of additive Gaussian noise. The major difficulty for optimally determining the parameters is that the corresponding maximum likelihood (ML) estimators involve finding the global minimum or maximum of multimodal cost functions because the frequencies are nonlinear in the observed s...
Hybrid TOA/AOA Approximate Maximum Likelihood Mobile Localization
Mohamed Zhaounia; Mohamed Adnan Landolsi; Ridha Bouallegue
2010-01-01
This letter deals with a hybrid time-of-arrival/angle-of-arrival (TOA/AOA) approximate maximum likelihood (AML) wireless location algorithm. Thanks to the use of both TOA/AOA measurements, the proposed technique can rely on two base stations (BS) only and achieves better performance compared to the original approximate maximum likelihood (AML) method. The use of two BSs is an important advantage in wireless cellular communication systems because it avoids hearability problems and reduces netw...
[Study on the maximum entropy principle and population genetic equilibrium].
Zhang, Hong-Li; Zhang, Hong-Yan
2006-03-01
A general mathematic model of population genetic equilibrium about one locus was constructed based on the maximum entropy principle by WANG Xiao-Long et al. They proved that the maximum solve of the model was just the frequency distribution that a population reached Hardy-Weinberg genetic equilibrium. It can suggest that a population reached Hardy-Weinberg genetic equilibrium when the genotype entropy of the population reached the maximal possible value, and that the frequency distribution of the maximum entropy was equivalent to the distribution of Hardy-Weinberg equilibrium law about one locus. They further assumed that the frequency distribution of the maximum entropy was equivalent to all genetic equilibrium distributions. This is incorrect, however. The frequency distribution of the maximum entropy was only equivalent to the distribution of Hardy-Weinberg equilibrium with respect to one locus or several limited loci. The case with regard to limited loci was proved in this paper. Finally we also discussed an example where the maximum entropy principle was not the equivalent of other genetic equilibria.
Maximum-likelihood estimation of haplotype frequencies in nuclear families.
Becker, Tim; Knapp, Michael
2004-07-01
The importance of haplotype analysis in the context of association fine mapping of disease genes has grown steadily over the last years. Since experimental methods to determine haplotypes on a large scale are not available, phase has to be inferred statistically. For individual genotype data, several reconstruction techniques and many implementations of the expectation-maximization (EM) algorithm for haplotype frequency estimation exist. Recent research work has shown that incorporating available genotype information of related individuals largely increases the precision of haplotype frequency estimates. We, therefore, implemented a highly flexible program written in C, called FAMHAP, which calculates maximum likelihood estimates (MLEs) of haplotype frequencies from general nuclear families with an arbitrary number of children via the EM-algorithm for up to 20 SNPs. For more loci, we have implemented a locus-iterative mode of the EM-algorithm, which gives reliable approximations of the MLEs for up to 63 SNP loci, or less when multi-allelic markers are incorporated into the analysis. Missing genotypes can be handled as well. The program is able to distinguish cases (haplotypes transmitted to the first affected child of a family) from pseudo-controls (non-transmitted haplotypes with respect to the child). We tested the performance of FAMHAP and the accuracy of the obtained haplotype frequencies on a variety of simulated data sets. The implementation proved to work well when many markers were considered and no significant differences between the estimates obtained with the usual EM-algorithm and those obtained in its locus-iterative mode were observed. We conclude from the simulations that the accuracy of haplotype frequency estimation and reconstruction in nuclear families is very reliable in general and robust against missing genotypes.
A Clustering Method Based on the Maximum Entropy Principle
Edwin Aldana-Bobadilla
2015-01-01
Full Text Available Clustering is an unsupervised process to determine which unlabeled objects in a set share interesting properties. The objects are grouped into k subsets (clusters whose elements optimize a proximity measure. Methods based on information theory have proven to be feasible alternatives. They are based on the assumption that a cluster is one subset with the minimal possible degree of “disorder”. They attempt to minimize the entropy of each cluster. We propose a clustering method based on the maximum entropy principle. Such a method explores the space of all possible probability distributions of the data to find one that maximizes the entropy subject to extra conditions based on prior information about the clusters. The prior information is based on the assumption that the elements of a cluster are “similar” to each other in accordance with some statistical measure. As a consequence of such a principle, those distributions of high entropy that satisfy the conditions are favored over others. Searching the space to find the optimal distribution of object in the clusters represents a hard combinatorial problem, which disallows the use of traditional optimization techniques. Genetic algorithms are a good alternative to solve this problem. We benchmark our method relative to the best theoretical performance, which is given by the Bayes classifier when data are normally distributed, and a multilayer perceptron network, which offers the best practical performance when data are not normal. In general, a supervised classification method will outperform a non-supervised one, since, in the first case, the elements of the classes are known a priori. In what follows, we show that our method’s effectiveness is comparable to a supervised one. This clearly exhibits the superiority of our method.
Estimate of the maximum induced magnetic field in relativistic shocks
Ghorbanalilu, M.; Sadegzadeh, S.
2017-01-01
The proton-driven Weibel instability is a crucial process for amplifying the generated magnetic fields in gamma-ray bursts. An expression for the saturation level of magnetic fields is estimated in a relativistic shock consisting of electron-proton plasmas. Within the shock transition layer, the plasma is modelled with the waterbag and Maxwell-Jüttner distribution functions for asymmetric counter-propagating proton beams and isotropic background electrons, respectively. The proton-driven Weibel-type instability in the linear phase is investigated thoroughly and then the instability conditions and the stabilization mechanisms are considered in details just after the shutdown of the electron Weibel instability. The growth rate of the instability and the saturated magnetic field strength are obtained in terms of the effective proton beam Mach number, asymmetry parameter, and the background electron temperature. In this paper, fully relativistic kinetic treatment is used to formulate the dispersion relation for the proton Weibel-type instability. Then, by using the magnetic trapping criteria, the saturated magnetic field strength is computed. In the present scenario, the instability includes two stages: in the first stage the electron Weibel instability evolves very rapidly, but in the second one because of the free energy stored in the slow counter-propagating proton beams, the instability is further amplified in the context of electrons with an isotropic distribution function. Increment of the growth rate and saturated magnetic field by increasing (decreasing) the effective proton beam Mach number (the asymmetry parameter) is deduced from the results. It is shown that at the temperatures around 108 K a maximum magnetic field up to around 56 G can be detected by this mechanism after the saturation time.
Mechanisms of maximum information preservation in the Drosophila antennal lobe.
Ryota Satoh
Full Text Available We examined the presence of maximum information preservation, which may be a fundamental principle of information transmission in all sensory modalities, in the Drosophila antennal lobe using an experimentally grounded network model and physiological data. Recent studies have shown a nonlinear firing rate transformation between olfactory receptor neurons (ORNs and second-order projection neurons (PNs. As a result, PNs can use their dynamic range more uniformly than ORNs in response to a diverse set of odors. Although this firing rate transformation is thought to assist the decoder in discriminating between odors, there are no comprehensive, quantitatively supported studies examining this notion. Therefore, we quantitatively investigated the efficiency of this firing rate transformation from the viewpoint of information preservation by computing the mutual information between odor stimuli and PN responses in our network model. In the Drosophila olfactory system, all ORNs and PNs are divided into unique functional processing units called glomeruli. The nonlinear transformation between ORNs and PNs is formed by intraglomerular transformation and interglomerular interaction through local neurons (LNs. By exploring possible nonlinear transformations produced by these two factors in our network model, we found that mutual information is maximized when a weak ORN input is preferentially amplified within a glomerulus and the net LN input to each glomerulus is inhibitory. It is noteworthy that this is the very combination observed experimentally. Furthermore, the shape of the resultant nonlinear transformation is similar to that observed experimentally. These results imply that information related to odor stimuli is almost maximally preserved in the Drosophila olfactory circuit. We also discuss how intraglomerular transformation and interglomerular inhibition combine to maximize mutual information.
THE STRENGTH OF THE DISSOLVED OXYGEN MAXIMUM IN THE VERTICAL DISTRIBUTION OF N ANSHA ISLANDS WATERS
林洪瑛; 程赛伟; 韩舞鹰
2002-01-01
Observation data from a cruise in the Nansha Islands, in May to June 1990, Decem ber, 1993, September to October 1994, and July, 1999, respectively, were used to develop the method presented here to indicate the existing strength of the diss ol ved oxygen maximum in the vertical distribution of Nansha Islands waters. Its se asonal variation and regional distribution are discussed in this paper. Analysis results showed that the distribution of the strength of dissolved oxygen maximum (Domax-Dosur) was closely related to the upper layer circulation and the bioactivity of Nansha Islands seawater.
Kuracina Richard
2015-06-01
Full Text Available The article deals with the measurement of maximum explosion pressure and the maximum rate of exposure pressure rise of wood dust cloud. The measurements were carried out according to STN EN 14034-1+A1:2011 Determination of explosion characteristics of dust clouds. Part 1: Determination of the maximum explosion pressure pmax of dust clouds and the maximum rate of explosion pressure rise according to STN EN 14034-2+A1:2012 Determination of explosion characteristics of dust clouds - Part 2: Determination of the maximum rate of explosion pressure rise (dp/dtmax of dust clouds. The wood dust cloud in the chamber is achieved mechanically. The testing of explosions of wood dust clouds showed that the maximum value of the pressure was reached at the concentrations of 450 g / m3 and its value is 7.95 bar. The fastest increase of pressure was observed at the concentrations of 450 g / m3 and its value was 68 bar / s.
Parameter Estimation for an Electric Arc Furnace Model Using Maximum Likelihood
Jesser J. Marulanda-Durango
2012-12-01
Full Text Available In this paper, we present a methodology for estimating the parameters of a model for an electrical arc furnace, by using maximum likelihood estimation. Maximum likelihood estimation is one of the most employed methods for parameter estimation in practical settings. The model for the electrical arc furnace that we consider, takes into account the non-periodic and non-linear variations in the voltage-current characteristic. We use NETLAB, an open source MATLAB® toolbox, for solving a set of non-linear algebraic equations that relate all the parameters to be estimated. Results obtained through simulation of the model in PSCADTM, are contrasted against real measurements taken during the furnance's most critical operating point. We show how the model for the electrical arc furnace, with appropriate parameter tuning, captures with great detail the real voltage and current waveforms generated by the system. Results obtained show a maximum error of 5% for the current's root mean square error.
Maximum Matchings of a Digraph Based on the Largest Geometric Multiplicity
Yunyun Yang
2016-01-01
Full Text Available Matching theory is one of the most forefront issues of graph theory. Based on the largest geometric multiplicity, we develop an efficient approach to identify maximum matchings in a digraph. For a given digraph, it has been proved that the number of maximum matched nodes has close relationship with the largest geometric multiplicity of the transpose of the adjacency matrix. Moreover, through fundamental column transformations, we can obtain the matched nodes and related matching edges. In particular, when a digraph contains a cycle factor, the largest geometric multiplicity is equal to one. In this case, the maximum matching is a perfect matching and each node in the digraph is a matched node. The method is validated by an example.
Size dependence of efficiency at maximum power of heat engine
Izumida, Y.
2013-10-01
We perform a molecular dynamics computer simulation of a heat engine model to study how the engine size difference affects its performance. Upon tactically increasing the size of the model anisotropically, we determine that there exists an optimum size at which the model attains the maximum power for the shortest working period. This optimum size locates between the ballistic heat transport region and the diffusive heat transport one. We also study the size dependence of the efficiency at the maximum power. Interestingly, we find that the efficiency at the maximum power around the optimum size attains a value that has been proposed as a universal upper bound, and it even begins to exceed the bound as the size further increases. We explain this behavior of the efficiency at maximum power by using a linear response theory for the heat engine operating under a finite working period, which naturally extends the low-dissipation Carnot cycle model [M. Esposito, R. Kawai, K. Lindenberg, C. Van den Broeck, Phys. Rev. Lett. 105, 150603 (2010)]. The theory also shows that the efficiency at the maximum power under an extreme condition may reach the Carnot efficiency in principle.© EDP Sciences Società Italiana di Fisica Springer-Verlag 2013.
How long do centenarians survive? Life expectancy and maximum lifespan.
Modig, K; Andersson, T; Vaupel, J; Rau, R; Ahlbom, A
2017-08-01
The purpose of this study was to explore the pattern of mortality above the age of 100 years. In particular, we aimed to examine whether Scandinavian data support the theory that mortality reaches a plateau at particularly old ages. Whether the maximum length of life increases with time was also investigated. The analyses were based on individual level data on all Swedish and Danish centenarians born from 1870 to 1901; in total 3006 men and 10 963 women were included. Birth cohort-specific probabilities of dying were calculated. Exact ages were used for calculations of maximum length of life. Whether maximum age changed over time was analysed taking into account increases in cohort size. The results confirm that there has not been any improvement in mortality amongst centenarians in the past 30 years and that the current rise in life expectancy is driven by reductions in mortality below the age of 100 years. The death risks seem to reach a plateau of around 50% at the age 103 years for men and 107 years for women. Despite the rising life expectancy, the maximum age does not appear to increase, in particular after accounting for the increasing number of individuals of advanced age. Mortality amongst centenarians is not changing despite improvements at younger ages. An extension of the maximum lifespan and a sizeable extension of life expectancy both require reductions in mortality above the age of 100 years. © 2017 The Association for the Publication of the Journal of Internal Medicine.
Prediction of three dimensional maximum isometric neck strength.
Fice, Jason B; Siegmund, Gunter P; Blouin, Jean-Sébastien
2014-09-01
We measured maximum isometric neck strength under combinations of flexion/extension, lateral bending and axial rotation to determine whether neck strength in three dimensions (3D) can be predicted from principal axes strength. This would allow biomechanical modelers to validate their neck models across many directions using only principal axis strength data. Maximum isometric neck moments were measured in 9 male volunteers (29±9 years) for 17 directions. The 3D moments were normalized by the principal axis moments, and compared to unity for all directions tested. Finally, each subject's maximum principal axis moments were used to predict their resultant moment in the off-axis directions. Maximum moments were 30±6 N m in flexion, 32±9 N m in lateral bending, 51±11 N m in extension, and 13±5 N m in axial rotation. The normalized 3D moments were not significantly different from unity (95% confidence interval contained one), except for three directions that combined ipsilateral axial rotation and lateral bending; in these directions the normalized moments exceeded one. Predicted resultant moments compared well to the actual measured values (r2=0.88). Despite exceeding unity, the normalized moments were consistent across subjects to allow prediction of maximum 3D neck strength using principal axes neck strength.
Evaluation of Maximum O2 Consumption: Using Ergo-Spirometry in Severe Heart Failure
Majid Malekmohammad
2012-09-01
Full Text Available Although sport-physiologists have repeatedly analyzed respiratory gases through exercise, it is relatively new in the cardiovascular field and is obviously more acceptable than standard exercise test, which gives only information about the existence or absence of cardiovascular diseases (CVDs. Through the new method of exercise test, parameters including aerobic and anaerobic are checked and monitored. 22 severe cases of heart failure, who were candidates of heart transplantation, referring to Massih Daneshvari Hospital in Tehran from Nov. 2007 to Nov. 2008 enrolled this study. The study was designed as a cross-sectional performance and evaluated only patients with ejection fraction less than 30%. O2 mean consumption was 6.27±4.9 ml/kg/min at rest and 9.48±3.38 at anaerobic threshold (AT exceeding 13 ml/kg/min in maximum which was significantly more than the expected levels. Respiratory exchange ratio (RER was over 1 for all patients. This study could not find any statistical correlations between VO2 max and participants' ergonomic factors such as age, height, weight, BMI, as well as EF. This study showed no significant correlation between VO2 max and maximum heart rate (HR max, although O2 maximum consumption was rationally correlated with expiratory ventilation. This means that the patients achieved maximum ventilation through exercise in this study, but failed to have their maximum heart rate being led probably by HF-induced brady-arrhythmia or deconditioning of skeletal muscles.
Assurance of the Maximum Destruction in Battlefield using Cost-Effective Approximation Techniques
Fariha Tasmin Jaigirdar
2012-12-01
Full Text Available Military Applications of Wireless Sensor Network in domains of maximizing security and gaining maximum benefits while attacking the opponent is a challenging and prominent area of research now-a-days. A commander’s goal in a battle field is not limited by securing his troops and the country but also to deliver proper commands to assault the enemies using the minimum number of resources. In this paper, we propose two efficient and low cost approximation algorithms—the maximum clique analysis and the maximum degree analysis techniques. Both of the techniques find the strategies of maximizing the destruction in a battlefield to defeat the opponent by utilizing limited resources. Experimental results show the effectiveness of the proposed algorithms in the prescribed areas of applications. Gaining the cost-effectiveness of the algorithms are also major concerns of this research. A comparative study explaining the number of resources required for commencing required level of destruction made to the opponents has been provided in this paper. The studies show that the maximum degree analysis technique is able to perform more destruction than the maximum clique analysis technique using same number of resources and requires relatively less computational complexity as well.