Directory of Open Access Journals (Sweden)
Elie Nadal
2018-02-01
Full Text Available In this study we fabricate gold nanocomposites and model their optical properties. The nanocomposites are either homogeneous films or gratings containing gold nanoparticles embedded in a polymer matrix. The samples are fabricated using a recently developed technique making use of laser interferometry. The gratings present original plasmon-enhanced diffraction properties. In this work, we develop a new approach to model the optical properties of our composites. We combine the extended Maxwell–Garnett model of effective media with the Rigorous Coupled Wave Analysis (RCWA method and compute both the absorption spectra and the diffraction efficiency spectra of the gratings. We show that such a semi-analytical approach allows us to reproduce the original plasmonic features of the composites and can provide us with details about their inner structure. Such an approach, considering reasonably high particle concentrations, could be a simple and efficient tool to study complex micro-structured system based on plasmonic components, such as metamaterials.
Estimation of the convergence order of rigorous coupled-wave analysis for OCD metrology
Ma, Yuan; Liu, Shiyuan; Chen, Xiuguo; Zhang, Chuanwei
2011-12-01
In most cases of optical critical dimension (OCD) metrology, when applying rigorous coupled-wave analysis (RCWA) to optical modeling, a high order of Fourier harmonics is usually set up to guarantee the convergence of the final results. However, the total number of floating point operations grows dramatically as the truncation order increases. Therefore, it is critical to choose an appropriate order to obtain high computational efficiency without losing much accuracy in the meantime. In this paper, the convergence order associated with the structural and optical parameters has been estimated through simulation. The results indicate that the convergence order is linear with the period of the sample when fixing the other parameters, both for planar diffraction and conical diffraction. The illuminated wavelength also affects the convergence of a final result. With further investigations concentrated on the ratio of illuminated wavelength to period, it is discovered that the convergence order decreases with the growth of the ratio, and when the ratio is fixed, convergence order jumps slightly, especially in a specific range of wavelength. This characteristic could be applied to estimate the optimum convergence order of given samples to obtain high computational efficiency.
Using grounded theory as a method for rigorously reviewing literature
Wolfswinkel, J.; Furtmueller-Ettinger, Elfriede; Wilderom, Celeste P.M.
2013-01-01
This paper offers guidance to conducting a rigorous literature review. We present this in the form of a five-stage process in which we use Grounded Theory as a method. We first probe the guidelines explicated by Webster and Watson, and then we show the added value of Grounded Theory for rigorously
Recent Development in Rigorous Computational Methods in Dynamical Systems
Arai, Zin; Kokubu, Hiroshi; Pilarczyk, Paweł
2009-01-01
We highlight selected results of recent development in the area of rigorous computations which use interval arithmetic to analyse dynamical systems. We describe general ideas and selected details of different ways of approach and we provide specific sample applications to illustrate the effectiveness of these methods. The emphasis is put on a topological approach, which combined with rigorous calculations provides a broad range of new methods that yield mathematically rel...
Coupled wave sensor technology
International Nuclear Information System (INIS)
Maki, M.C.
1988-01-01
Buried line guided radar sensors have been used successfully for a number of years to provide perimeter security for high value resources. This paper introduces a new complementary sensor advancement at Computing Devices termed 'coupled wave device technology' (CWD). It provides many of the inherent advantages of leakey cable sensors, such as terrain-following and the ability to discriminate between humans and small animals. It also is able to provide a high or wide detection zone, and allows the sensor to be mounted aerially and adjacent to a wall or fence. Several alternative sensors have been developed which include a single-line sensor, a dual-line hybrid sensor that combines the elements of ported coax and CWD technology, and a rapid-deployment portable sensor for temporary or mobile applications. A description of the technology, the sensors, and their characteristics is provided
The MIXED framework: A novel approach to evaluating mixed-methods rigor.
Eckhardt, Ann L; DeVon, Holli A
2017-10-01
Evaluation of rigor in mixed-methods (MM) research is a persistent challenge due to the combination of inconsistent philosophical paradigms, the use of multiple research methods which require different skill sets, and the need to combine research at different points in the research process. Researchers have proposed a variety of ways to thoroughly evaluate MM research, but each method fails to provide a framework that is useful for the consumer of research. In contrast, the MIXED framework is meant to bridge the gap between an academic exercise and practical assessment of a published work. The MIXED framework (methods, inference, expertise, evaluation, and design) borrows from previously published frameworks to create a useful tool for the evaluation of a published study. The MIXED framework uses an experimental eight-item scale that allows for comprehensive integrated assessment of MM rigor in published manuscripts. Mixed methods are becoming increasingly prevalent in nursing and healthcare research requiring researchers and consumers to address issues unique to MM such as evaluation of rigor. © 2017 John Wiley & Sons Ltd.
Optimal correction and design parameter search by modern methods of rigorous global optimization
International Nuclear Information System (INIS)
Makino, K.; Berz, M.
2011-01-01
Frequently the design of schemes for correction of aberrations or the determination of possible operating ranges for beamlines and cells in synchrotrons exhibit multitudes of possibilities for their correction, usually appearing in disconnected regions of parameter space which cannot be directly qualified by analytical means. In such cases, frequently an abundance of optimization runs are carried out, each of which determines a local minimum depending on the specific chosen initial conditions. Practical solutions are then obtained through an often extended interplay of experienced manual adjustment of certain suitable parameters and local searches by varying other parameters. However, in a formal sense this problem can be viewed as a global optimization problem, i.e. the determination of all solutions within a certain range of parameters that lead to a specific optimum. For example, it may be of interest to find all possible settings of multiple quadrupoles that can achieve imaging; or to find ahead of time all possible settings that achieve a particular tune; or to find all possible manners to adjust nonlinear parameters to achieve correction of high order aberrations. These tasks can easily be phrased in terms of such an optimization problem; but while mathematically this formulation is often straightforward, it has been common belief that it is of limited practical value since the resulting optimization problem cannot usually be solved. However, recent significant advances in modern methods of rigorous global optimization make these methods feasible for optics design for the first time. The key ideas of the method lie in an interplay of rigorous local underestimators of the objective functions, and by using the underestimators to rigorously iteratively eliminate regions that lie above already known upper bounds of the minima, in what is commonly known as a branch-and-bound approach. Recent enhancements of the Differential Algebraic methods used in particle
A study into first-year engineering education success using a rigorous mixed methods approach
DEFF Research Database (Denmark)
van den Bogaard, M.E.D.; de Graaff, Erik; Verbraek, Alexander
2015-01-01
The aim of this paper is to combine qualitative and quantitative research methods into rigorous research into student success. Research methods have weaknesses that can be overcome by clever combinations. In this paper we use a situated study into student success as an example of how methods...... using statistical techniques. The main elements of the model were student behaviour and student disposition, which were influenced by the students’ perceptions of the education environment. The outcomes of the qualitative studies were useful in interpreting the outcomes of the structural equation...
Application of the rigorous method to x-ray and neutron beam scattering on rough surfaces
International Nuclear Information System (INIS)
Goray, Leonid I.
2010-01-01
The paper presents a comprehensive numerical analysis of x-ray and neutron scattering from finite-conducting rough surfaces which is performed in the frame of the boundary integral equation method in a rigorous formulation for high ratios of characteristic dimension to wavelength. The single integral equation obtained involves boundary integrals of the single and double layer potentials. A more general treatment of the energy conservation law applicable to absorption gratings and rough mirrors is considered. In order to compute the scattering intensity of rough surfaces using the forward electromagnetic solver, Monte Carlo simulation is employed to average the deterministic diffraction grating efficiency due to individual surfaces over an ensemble of realizations. Some rules appropriate for numerical implementation of the theory at small wavelength-to-period ratios are presented. The difference between the rigorous approach and approximations can be clearly seen in specular reflectances of Au mirrors with different roughness parameters at wavelengths where grazing incidence occurs at close to or larger than the critical angle. This difference may give rise to wrong estimates of rms roughness and correlation length if they are obtained by comparing experimental data with calculations. Besides, the rigorous approach permits taking into account any known roughness statistics and allows exact computation of diffuse scattering.
Dimitrakopoulos, Panagiotis
2018-03-01
The calculation of polytropic efficiencies is a very important task, especially during the development of new compression units, like compressor impellers, stages and stage groups. Such calculations are also crucial for the determination of the performance of a whole compressor. As processors and computational capacities have substantially been improved in the last years, the need for a new, rigorous, robust, accurate and at the same time standardized method merged, regarding the computation of the polytropic efficiencies, especially based on thermodynamics of real gases. The proposed method is based on the rigorous definition of the polytropic efficiency. The input consists of pressure and temperature values at the end points of the compression path (suction and discharge), for a given working fluid. The average relative error for the studied cases was 0.536 %. Thus, this high-accuracy method is proposed for efficiency calculations related with turbocompressors and their compression units, especially when they are operating at high power levels, for example in jet engines and high-power plants.
How to Map Theory: Reliable Methods Are Fruitless Without Rigorous Theory.
Gray, Kurt
2017-09-01
Good science requires both reliable methods and rigorous theory. Theory allows us to build a unified structure of knowledge, to connect the dots of individual studies and reveal the bigger picture. Some have criticized the proliferation of pet "Theories," but generic "theory" is essential to healthy science, because questions of theory are ultimately those of validity. Although reliable methods and rigorous theory are synergistic, Action Identification suggests psychological tension between them: The more we focus on methodological details, the less we notice the broader connections. Therefore, psychology needs to supplement training in methods (how to design studies and analyze data) with training in theory (how to connect studies and synthesize ideas). This article provides a technique for visually outlining theory: theory mapping. Theory mapping contains five elements, which are illustrated with moral judgment and with cars. Also included are 15 additional theory maps provided by experts in emotion, culture, priming, power, stress, ideology, morality, marketing, decision-making, and more (see all at theorymaps.org ). Theory mapping provides both precision and synthesis, which helps to resolve arguments, prevent redundancies, assess the theoretical contribution of papers, and evaluate the likelihood of surprising effects.
Manuel, Sharrón L; Johnson, Brian W; Frevert, Charles W; Duncan, Francesca E
2018-04-21
Immunohistochemistry (IHC) is a robust scientific tool whereby cellular components are visualized within a tissue, and this method has been and continues to be a mainstay for many reproductive biologists. IHC is highly informative if performed and interpreted correctly, but studies have shown that the general use and reporting of appropriate controls in IHC experiments is low. This omission of the scientific method can result in data that lacks rigor and reproducibility. In this editorial, we highlight key concepts in IHC controls and describe an opportunity for our field to partner with the Histochemical Society to adopt their IHC guidelines broadly as researchers, authors, ad hoc reviewers, editorial board members, and editors-in-chief. Such cross-professional society interactions will ensure that we produce the highest quality data as new technologies emerge that still rely upon the foundations of classic histological and immunohistochemical principles.
A new method for deriving rigorous results on ππ scattering
International Nuclear Information System (INIS)
Caprini, I.; Dita, P.
1979-06-01
We develop a new approach to the problem of constraining the ππ scattering amplitudes by means of the axiomatically proved properties of unitarity, analyticity and crossing symmetry. The method is based on the solution of an extremal problem on a convex set of analytic functions and provides a global description of the domain of values taken by any finite number of partial waves at an arbitrary set of unphysical energies, compatible with unitarity, the bounds at complex energies derived from generalized dispersion relations and the crossing integral relations. From this doma domain we obtain new absolute bounds for the amplitudes as well as rigorous correlations between the values of various partial waves. (author)
Energy Technology Data Exchange (ETDEWEB)
Zhaoyuan Liu; Kord Smith; Benoit Forget; Javier Ortensi
2016-05-01
A new method for computing homogenized assembly neutron transport cross sections and dif- fusion coefficients that is both rigorous and computationally efficient is proposed in this paper. In the limit of a homogeneous hydrogen slab, the new method is equivalent to the long-used, and only-recently-published CASMO transport method. The rigorous method is used to demonstrate the sources of inaccuracy in the commonly applied “out-scatter” transport correction. It is also demonstrated that the newly developed method is directly applicable to lattice calculations per- formed by Monte Carlo and is capable of computing rigorous homogenized transport cross sections for arbitrarily heterogeneous lattices. Comparisons of several common transport cross section ap- proximations are presented for a simple problem of infinite medium hydrogen. The new method has also been applied in computing 2-group diffusion data for an actual PWR lattice from BEAVRS benchmark.
Directory of Open Access Journals (Sweden)
Henrik von Wehrden
2017-02-01
Full Text Available Sustainability science encompasses a unique field that is defined through its purpose, the problem it addresses, and its solution-oriented agenda. However, this orientation creates significant methodological challenges. In this discussion paper, we conceptualize sustainability problems as wicked problems to tease out the key challenges that sustainability science is facing if scientists intend to deliver on its solution-oriented agenda. Building on the available literature, we discuss three aspects that demand increased attention for advancing sustainability science: 1 methods with higher diversity and complementarity are needed to increase the chance of deriving solutions to the unique aspects of wicked problems; for instance, mixed methods approaches are potentially better suited to allow for an approximation of solutions, since they cover wider arrays of knowledge; 2 methodologies capable of dealing with wicked problems demand strict procedural and ethical guidelines, in order to ensure their integration potential; for example, learning from solution implementation in different contexts requires increased comparability between research approaches while carefully addressing issues of legitimacy and credibility; and 3 approaches are needed that allow for longitudinal research, since wicked problems are continuous and solutions can only be diagnosed in retrospect; for example, complex dynamics of wicked problems play out across temporal patterns that are not necessarily aligned with the common timeframe of participatory sustainability research. Taken together, we call for plurality in methodologies, emphasizing procedural rigor and the necessity of continuous research to effectively addressing wicked problems as well as methodological challenges in sustainability science.
A plea for rigorous conceptual analysis as central method in transnational law design
Rijgersberg, R.; van der Kaaij, H.
2013-01-01
Although shared problems are generally easily identified in transnational law design, it is considerably more difficult to design frameworks that transcend the peculiarities of local law in a univocal fashion. The following exposition is a plea for giving more prominence to rigorous conceptual
Diffraction-based overlay measurement on dedicated mark using rigorous modeling method
Lu, Hailiang; Wang, Fan; Zhang, Qingyun; Chen, Yonghui; Zhou, Chang
2012-03-01
Diffraction Based Overlay (DBO) is widely evaluated by numerous authors, results show DBO can provide better performance than Imaging Based Overlay (IBO). However, DBO has its own problems. As well known, Modeling based DBO (mDBO) faces challenges of low measurement sensitivity and crosstalk between various structure parameters, which may result in poor accuracy and precision. Meanwhile, main obstacle encountered by empirical DBO (eDBO) is that a few pads must be employed to gain sufficient information on overlay-induced diffraction signature variations, which consumes more wafer space and costs more measuring time. Also, eDBO may suffer from mark profile asymmetry caused by processes. In this paper, we propose an alternative DBO technology that employs a dedicated overlay mark and takes a rigorous modeling approach. This technology needs only two or three pads for each direction, which is economic and time saving. While overlay measurement error induced by mark profile asymmetry being reduced, this technology is expected to be as accurate and precise as scatterometry technologies.
Simulation methods for multiperiodic and aperiodic nanostructured dielectric waveguides
DEFF Research Database (Denmark)
Paulsen, Moritz; Neustock, Lars Thorben; Jahns, Sabrina
2017-01-01
on Rudin–Shapiro, Fibonacci, and Thue–Morse binary sequences. The near-field and far-field properties are computed employing the finite-element method (FEM), the finite-difference time-domain (FDTD) method as well as a rigorous coupled wave algorithm (RCWA). The results show that all three methods...
Bamberger, Michael; Tarsilla, Michele; Hesse-Biber, Sharlene
2016-04-01
Many widely-used impact evaluation designs, including randomized control trials (RCTs) and quasi-experimental designs (QEDs), frequently fail to detect what are often quite serious unintended consequences of development programs. This seems surprising as experienced planners and evaluators are well aware that unintended consequences frequently occur. Most evaluation designs are intended to determine whether there is credible evidence (statistical, theory-based or narrative) that programs have achieved their intended objectives and the logic of many evaluation designs, even those that are considered the most "rigorous," does not permit the identification of outcomes that were not specified in the program design. We take the example of RCTs as they are considered by many to be the most rigorous evaluation designs. We present a numbers of cases to illustrate how infusing RCTs with a mixed-methods approach (sometimes called an "RCT+" design) can strengthen the credibility of these designs and can also capture important unintended consequences. We provide a Mixed Methods Evaluation Framework that identifies 9 ways in which UCs can occur, and we apply this framework to two of the case studies. Copyright © 2016 Elsevier Ltd. All rights reserved.
Scientific rigor through videogames.
Treuille, Adrien; Das, Rhiju
2014-11-01
Hypothesis-driven experimentation - the scientific method - can be subverted by fraud, irreproducibility, and lack of rigorous predictive tests. A robust solution to these problems may be the 'massive open laboratory' model, recently embodied in the internet-scale videogame EteRNA. Deploying similar platforms throughout biology could enforce the scientific method more broadly. Copyright © 2014 Elsevier Ltd. All rights reserved.
Simulation Methods for Multiperiodic and Aperiodic Nanostructured Dielectric Waveguides
DEFF Research Database (Denmark)
Paulsen, Moritz; Neustock, Lars Thorben; Jahns, Sabrina
on Rudin-Shapiro, Fibonacci, and Thue-Morse binary sequences. The near-field and far-field properties are calculated employing the finite-element method (FEM), the finite- difference time-domain (FDTD) method as well as a rigorous coupled wave algorithm (RCWA). References [1] S. V. Boriskina, A. Gopinath...
Shinogle-Decker, Heather; Martinez-Rivera, Noraida; O'Brien, John; Powell, Richard D.; Joshi, Vishwas N.; Connell, Samuel; Rosa-Molinar, Eduardo
2018-02-01
A new correlative Förster Resonance Energy Transfer (FRET) microscopy method using FluoroNanogold™, a fluorescent immunoprobe with a covalently attached Nanogold® particle (1.4nm Au), overcomes resolution limitations in determining distances within synaptic nanoscale architecture. FRET by acceptor photobleaching has long been used as a method to increase fluorescence resolution. The transfer of energy from a donor to an acceptor generally occurs between 10-100Å, which is the relative distance between the donor molecule and the acceptor molecule. For the correlative FRET microscopy method using FluoroNanogold™, we immuno-labeled GFP-tagged-HeLa-expressing Connexin 35 (Cx35) with anti-GFP and with anti-Cx35/36 antibodies, and then photo-bleached the Cx before processing the sample for electron microscopic imaging. Preliminary studies reveal the use of Alexa Fluor® 594 FluoroNanogold™ slightly increases FRET distance to 70Å, in contrast to the 62.5Å using AlexaFluor 594®. Preliminary studies also show that using a FluoroNanogold™ probe inhibits photobleaching. After one photobleaching session, Alexa Fluor 594® fluorescence dropped to 19% of its original fluorescence; in contrast, after one photobleaching session, Alexa Fluor 594® FluoroNanogold™ fluorescence dropped to 53% of its original intensity. This result confirms that Alexa Fluor 594® FluoroNanogold™ is a much better donor probe than is Alexa Fluor 594®. The new method (a) creates a double confirmation method in determining structure and orientation of synaptic architecture, (b) allows development of a two-dimensional in vitro model to be used for precise testing of multiple parameters, and (c) increases throughput. Future work will include development of FluoroNanogold™ probes with different sizes of gold for additional correlative microscopy studies.
International Nuclear Information System (INIS)
Hiratsuka, Y.; Oryu, S.; Gojuki, S.
2011-01-01
Reliability of the screened Coulomb renormalization method, which was proposed in an elegant way by Alt-Sandhas-Zankel-Ziegelmann (ASZZ), is discussed on the basis of 'two-potential theory' for the three-body AGS equations with the Coulomb potential. In order to obtain ASZZ's formula, we define the on-shell Moller function, and calculate it by using the Haeringen criterion, i. e. 'the half-shell Coulomb amplitude is zero'. By these two steps, we can finally obtain the ASZZ formula for a small Coulomb phase shift. Furthermore, the reliability of the Haeringen criterion is thoroughly checked by a numerically rigorous calculation for the Coulomb LS-type equation. We find that the Haeringen criterion can be satisfied only in the higher energy region. We conclude that the ASZZ method can be verified in the case that the on-shell approximation to the Moller function is reasonable, and the Haeringen criterion is reliable. (author)
Directory of Open Access Journals (Sweden)
Feng Xiao-Jiang
2008-10-01
Full Text Available Abstract Background The analysis of large-scale data sets via clustering techniques is utilized in a number of applications. Biclustering in particular has emerged as an important problem in the analysis of gene expression data since genes may only jointly respond over a subset of conditions. Biclustering algorithms also have important applications in sample classification where, for instance, tissue samples can be classified as cancerous or normal. Many of the methods for biclustering, and clustering algorithms in general, utilize simplified models or heuristic strategies for identifying the "best" grouping of elements according to some metric and cluster definition and thus result in suboptimal clusters. Results In this article, we present a rigorous approach to biclustering, OREO, which is based on the Optimal RE-Ordering of the rows and columns of a data matrix so as to globally minimize the dissimilarity metric. The physical permutations of the rows and columns of the data matrix can be modeled as either a network flow problem or a traveling salesman problem. Cluster boundaries in one dimension are used to partition and re-order the other dimensions of the corresponding submatrices to generate biclusters. The performance of OREO is tested on (a metabolite concentration data, (b an image reconstruction matrix, (c synthetic data with implanted biclusters, and gene expression data for (d colon cancer data, (e breast cancer data, as well as (f yeast segregant data to validate the ability of the proposed method and compare it to existing biclustering and clustering methods. Conclusion We demonstrate that this rigorous global optimization method for biclustering produces clusters with more insightful groupings of similar entities, such as genes or metabolites sharing common functions, than other clustering and biclustering algorithms and can reconstruct underlying fundamental patterns in the data for several distinct sets of data matrices arising
On error estimation in the fourier modal method for diffractive gratings
Hlod, A.; Maubach, J.M.L.
2010-01-01
The Fourier Modal Method (FMM, also called the Rigorous Coupled Wave Analysis, RCWA) is a numerical discretization method which is often used to calculate a scattered field from a periodic diffraction grating. For 1D periodic gratings in FMM the electromagnetic field is presented by a truncated
Roberts, Brenden; Vidick, Thomas; Motrunich, Olexei I.
2017-12-01
The success of polynomial-time tensor network methods for computing ground states of certain quantum local Hamiltonians has recently been given a sound theoretical basis by Arad et al. [Math. Phys. 356, 65 (2017), 10.1007/s00220-017-2973-z]. The convergence proof, however, relies on "rigorous renormalization group" (RRG) techniques which differ fundamentally from existing algorithms. We introduce a practical adaptation of the RRG procedure which, while no longer theoretically guaranteed to converge, finds matrix product state ansatz approximations to the ground spaces and low-lying excited spectra of local Hamiltonians in realistic situations. In contrast to other schemes, RRG does not utilize variational methods on tensor networks. Rather, it operates on subsets of the system Hilbert space by constructing approximations to the global ground space in a treelike manner. We evaluate the algorithm numerically, finding similar performance to density matrix renormalization group (DMRG) in the case of a gapped nondegenerate Hamiltonian. Even in challenging situations of criticality, large ground-state degeneracy, or long-range entanglement, RRG remains able to identify candidate states having large overlap with ground and low-energy eigenstates, outperforming DMRG in some cases.
Development of interface between MCNP-FISPACT-MCNP (IPR-MFM) based on rigorous two step method
International Nuclear Information System (INIS)
Shaw, A.K.; Swami, H.L.; Danani, C.
2015-01-01
In this work we present the development of interface tool between MCNP-FISPACT-MCNP (MFM) based on Rigorous Two Step method for the shutdown dose rate (SDDR) calculation. The MFM links MCNP radiation transport and the FISPACT inventory code through a suitable coupling scheme. MFM coupling scheme has three steps. In first step it picks neutron spectrum and total flux from MCNP output file to use as input parameter for FISPACT. It prepares the FISPACT input files by using irradiation history, neutron flux and neutron spectrum and then execute the FISPACT input file in the second step. Third step of MFM coupling scheme extracts the decay gammas from the FISPACT output file and prepares MCNP input file for decay gamma transport followed by execution of MCNP input file and estimation of SDDR. Here detailing of MFM methodology and flow scheme has been described. The programming language PYTHON has been chosen for this development of the coupling scheme. A complete loop of MCNP-FISPACT-MCNP has been developed to handle the simplified geometrical problems. For validation of MFM interface a manual cross-check has been performed which shows good agreements. The MFM interface also has been validated with exiting MCNP-D1S method for a simple geometry with 14 MeV cylindrical neutron source. (author)
International Nuclear Information System (INIS)
Espinoza-Ojeda, O M; Santoyo, E; Andaverde, J
2011-01-01
Approximate and rigorous solutions of seven heat transfer models were statistically examined, for the first time, to estimate stabilized formation temperatures (SFT) of geothermal and petroleum boreholes. Constant linear and cylindrical heat source models were used to describe the heat flow (either conductive or conductive/convective) involved during a borehole drilling. A comprehensive statistical assessment of the major error sources associated with the use of these models was carried out. The mathematical methods (based on approximate and rigorous solutions of heat transfer models) were thoroughly examined by using four statistical analyses: (i) the use of linear and quadratic regression models to infer the SFT; (ii) the application of statistical tests of linearity to evaluate the actual relationship between bottom-hole temperatures and time function data for each selected method; (iii) the comparative analysis of SFT estimates between the approximate and rigorous predictions of each analytical method using a β ratio parameter to evaluate the similarity of both solutions, and (iv) the evaluation of accuracy in each method using statistical tests of significance, and deviation percentages between 'true' formation temperatures and SFT estimates (predicted from approximate and rigorous solutions). The present study also enabled us to determine the sensitivity parameters that should be considered for a reliable calculation of SFT, as well as to define the main physical and mathematical constraints where the approximate and rigorous methods could provide consistent SFT estimates
Directory of Open Access Journals (Sweden)
Christian Kobbernagel
2016-06-01
Full Text Available In the last couple of decades there has been an unprecedented explosion of news media platforms and formats, as a succession of digital and social media have joined the ranks of legacy media. We live in a ‘hybrid media system’ (Chadwick, 2013, in which people build their cross-media news repertoires from the ensemble of old and new media available. This article presents an innovative mixed-method approach with considerable explanatory power to the exploration of patterns of news media consumption. This approach tailors Q-methodology in the direction of a qualitative study of news consumption, in which a card sorting exercise serves to translate the participants’ news media preferences into a form that enables the researcher to undertake a rigorous factor-analytical construction of their news consumption repertoires. This interpretive, factor-analytical procedure, which results in the building of six audience news repertoires in Denmark, also preserves the qualitative thickness of the participants’ verbal accounts of the communicative figurations of their day-in-the-life with the news media.
International Nuclear Information System (INIS)
Goodman, F.O.
1977-01-01
A rigorous treatment of the scattering of atoms by a stationary sinusoidal hard wall in (n+1) dimensions is presented, a previous treatment by Masel, Merrill, and Miller for n=1 being contained as a special case. Numerical comparisons are made with the GR method of Garcia, which incorporates the Rayleigh hypothesis. Advantages and disadvantages of both methods are discussed, and it is concluded that the Rayleigh GR method, if handled properly, will probably work satisfactorily in physically realistic cases
International Nuclear Information System (INIS)
Takahashi, A.; Rusch, D.
1979-07-01
Some recent neutronics experiments for fusion reactor blankets show that the precise treatment of anisotropic secondary emissions for all types of neutron scattering is needed for neutron transport calculations. In the present work new rigorous methods, i.e. based on non-approximative microscopic neutron balance equations, are applied to treat the anisotropic collision source term in transport equations. The collision source calculation is free from approximations except for the discretization of energy, angle and space variables and includes the rigorous treatment of nonelastic collisions, as far as nuclear data are given. Two methods are presented: first the Ii-method, which relies on existing nuclear data files and then, as an ultimate goal, the I*-method, which aims at the use of future double-differential cross section data, but which is also applicable to the present single-differential data basis to allow a smooth transition to the new data type. An application of the Ii-method is given in the code system NITRAN which employs the Ssub(N)-method to solve the transport equations. Both rigorous methods, the Ii- and the I*-method, are applicable to all radiation transport problems and they can be used also in the Monte-Carlo-method to solve the transport problem. (orig./RW) [de
Volume Holograms in Photopolymers: Comparison between Analytical and Rigorous Theories
Gallego, Sergi; Neipp, Cristian; Estepa, Luis A.; Ortuño, Manuel; Márquez, Andrés; Francés, Jorge; Pascual, Inmaculada; Beléndez, Augusto
2012-01-01
There is no doubt that the concept of volume holography has led to an incredibly great amount of scientific research and technological applications. One of these applications is the use of volume holograms as optical memories, and in particular, the use of a photosensitive medium like a photopolymeric material to record information in all its volume. In this work we analyze the applicability of Kogelnik’s Coupled Wave theory to the study of volume holograms recorded in photopolymers. Some of the theoretical models in the literature describing the mechanism of hologram formation in photopolymer materials use Kogelnik’s theory to analyze the gratings recorded in photopolymeric materials. If Kogelnik’s theory cannot be applied is necessary to use a more general Coupled Wave theory (CW) or the Rigorous Coupled Wave theory (RCW). The RCW does not incorporate any approximation and thus, since it is rigorous, permits judging the accurateness of the approximations included in Kogelnik’s and CW theories. In this article, a comparison between the predictions of the three theories for phase transmission diffraction gratings is carried out. We have demonstrated the agreement in the prediction of CW and RCW and the validity of Kogelnik’s theory only for gratings with spatial frequencies higher than 500 lines/mm for the usual values of the refractive index modulations obtained in photopolymers.
Volume Holograms in Photopolymers: Comparison between Analytical and Rigorous Theories
Directory of Open Access Journals (Sweden)
Augusto Beléndez
2012-08-01
Full Text Available There is no doubt that the concept of volume holography has led to an incredibly great amount of scientific research and technological applications. One of these applications is the use of volume holograms as optical memories, and in particular, the use of a photosensitive medium like a photopolymeric material to record information in all its volume. In this work we analyze the applicability of Kogelnik’s Coupled Wave theory to the study of volume holograms recorded in photopolymers. Some of the theoretical models in the literature describing the mechanism of hologram formation in photopolymer materials use Kogelnik’s theory to analyze the gratings recorded in photopolymeric materials. If Kogelnik’s theory cannot be applied is necessary to use a more general Coupled Wave theory (CW or the Rigorous Coupled Wave theory (RCW. The RCW does not incorporate any approximation and thus, since it is rigorous, permits judging the accurateness of the approximations included in Kogelnik’s and CW theories. In this article, a comparison between the predictions of the three theories for phase transmission diffraction gratings is carried out. We have demonstrated the agreement in the prediction of CW and RCW and the validity of Kogelnik’s theory only for gratings with spatial frequencies higher than 500 lines/mm for the usual values of the refractive index modulations obtained in photopolymers.
Coupled wave equations theory of surface-enhanced femtosecond stimulated Raman scattering.
McAnally, Michael O; McMahon, Jeffrey M; Van Duyne, Richard P; Schatz, George C
2016-09-07
We present a coupled wave semiclassical theory to describe plasmonic enhancement effects in surface-enhanced femtosecond stimulated Raman scattering (SE-FSRS). A key result is that the plasmon enhanced fields which drive the vibrational equation of motion for each normal mode results in dispersive lineshapes in the SE-FSRS spectrum. This result, which reproduces experimental lineshapes, demonstrates that plasmon-enhanced stimulated Raman methods provide unique sensitivity to a plasmonic response. Our derived SE-FSRS theory shows a plasmonic enhancement of |gpu|(2)ImχR(ω)gst (2)/ImχR(ω), where |gpu|(2) is the absolute square of the plasmonic enhancement from the Raman pump, χR(ω) is the Raman susceptibility, and gst is the plasmonic enhancement of the Stokes field in SE-FSRS. We conclude with a discussion on potential future experimental and theoretical directions for the field of plasmonically enhanced coherent Raman scattering.
Ehret, Gerd; Bodermann, Bernd; Woehler, Martin
2007-06-01
The optical microscopy is an important instrument for dimensional characterisation or calibration of micro- and nanostructures, e.g. chrome structures on photomasks. In comparison to scanning electron microscopy (possible contamination of the sample) and atomic force microscopy (slow, risk of damage) optical microscopy is a fast and non destructive metrology method. The precise quantitative determination of the linewidth from the microscope image is, however, only possible by knowledge of the geometry of the structures and their consideration in the optical modelling. We compared two different rigorous model approaches, the Rigorous Coupled Wave Analysis (RCWA) and the Finite Elements Method (FEM) for modelling of structures with different edge angles, linewidths, line to space ratios and polarisations. The RCWA method can adapt inclined edges profiles only by a staircase approximation leading to increased modelling errors of the RCWA method. Even today's sophisticated rigorous methods still show problems with TM-polarisation. Therefore both rigorous methods are compared in terms of their convergence for TE and TM- polarisation. Beyond that also the influence of typical illumination wavelengths (365 nm, 248 nm and 193 nm) on the microscope images and their contribution to the measuring uncertainty budget will be discussed.
International Nuclear Information System (INIS)
Pierru, A.; Babusiaux, D.
2009-02-01
The problem studied is that of valuing investment projects of an international oil company subject to tax schemes that vary from one country to another. The existing disparities in the tax treatment of interest paid can lead the firm to seek an optimal allocation of its debt capacity among the various projects. In this context, the generalized ATWACC (After-Tax Weighted Average Cost of Capital) method presents numerous advantages over standard methods and is particularly well suited to the valuation of oil-field development projects where debt financing differs from the amount that would correspond to the debt ratio targeted by the firm at the corporate scale. In this paper, we discuss adapting the generalized ATWACC method to the specificities of the oil industry and offer new proof of its validity, based on a model that maximizes, under constraints, the firm's equity value. (authors)
DiMaggio, PA; McAllister, SR; Floudas, CA; Feng, X-J; Rabinowitz, JD; Rabitz, HA
2008-01-01
Abstract Background The analysis of large-scale data sets via clustering techniques is utilized in a number of applications. Biclustering in particular has emerged as an important problem in the analysis of gene expression data since genes may only jointly respond over a subset of conditions. Biclustering algorithms also have important applications in sample classification where, for instance, tissue samples can be classified as cancerous or normal. Many of the methods for biclustering, and c...
Statistical mechanics rigorous results
Ruelle, David
1999-01-01
This classic book marks the beginning of an era of vigorous mathematical progress in equilibrium statistical mechanics. Its treatment of the infinite system limit has not been superseded, and the discussion of thermodynamic functions and states remains basic for more recent work. The conceptual foundation provided by the Rigorous Results remains invaluable for the study of the spectacular developments of statistical mechanics in the second half of the 20th century.
Application of Coupled-Wave Wentzel-Kramers-Brillouin Approximation to Ground Penetrating Radar
Directory of Open Access Journals (Sweden)
Igor Prokopovich
2017-12-01
Full Text Available This paper deals with bistatic subsurface probing of a horizontally layered dielectric half-space by means of ultra-wideband electromagnetic waves. In particular, the main objective of this work is to present a new method for the solution of the two-dimensional back-scattering problem arising when a pulsed electromagnetic signal impinges on a non-uniform dielectric half-space; this scenario is of interest for ground penetrating radar (GPR applications. For the analytical description of the signal generated by the interaction of the emitted pulse with the environment, we developed and implemented a novel time-domain version of the coupled-wave Wentzel-Kramers-Brillouin approximation. We compared our solution with finite-difference time-domain (FDTD results, achieving a very good agreement. We then applied the proposed technique to two case studies: in particular, our method was employed for the post-processing of experimental radargrams collected on Lake Chebarkul, in Russia, and for the simulation of GPR probing of the Moon surface, to detect smooth gradients of the dielectric permittivity in lunar regolith. The main conclusions resulting from our study are that our semi-analytical method is accurate, radically accelerates calculations compared to simpler mathematical formulations with a mostly numerical nature (such as the FDTD technique, and can be effectively used to aid the interpretation of GPR data. The method is capable to correctly predict the protracted return signals originated by smooth transition layers of the subsurface dielectric medium. The accuracy and numerical efficiency of our computational approach make promising its further development.
Putrefactive rigor: apparent rigor mortis due to gas distension.
Gill, James R; Landi, Kristen
2011-09-01
Artifacts due to decomposition may cause confusion for the initial death investigator, leading to an incorrect suspicion of foul play. Putrefaction is a microorganism-driven process that results in foul odor, skin discoloration, purge, and bloating. Various decompositional gases including methane, hydrogen sulfide, carbon dioxide, and hydrogen will cause the body to bloat. We describe 3 instances of putrefactive gas distension (bloating) that produced the appearance of inappropriate rigor, so-called putrefactive rigor. These gases may distend the body to an extent that the extremities extend and lose contact with their underlying support surface. The medicolegal investigator must recognize that this is not true rigor mortis and the body was not necessarily moved after death for this gravity-defying position to occur.
Simple design of slanted grating with simplified modal method.
Li, Shubin; Zhou, Changhe; Cao, Hongchao; Wu, Jun
2014-02-15
A simplified modal method (SMM) is presented that offers a clear physical image for subwavelength slanted grating. The diffraction characteristic of the slanted grating under Littrow configuration is revealed by the SMM as an equivalent rectangular grating, which is in good agreement with rigorous coupled-wave analysis. Based on the equivalence, we obtained an effective analytic solution for simplifying the design and optimization of a slanted grating. It offers a new approach for design of the slanted grating, e.g., a 1×2 beam splitter can be easily designed. This method should be helpful for designing various new slanted grating devices.
Mathematical Rigor in Introductory Physics
Vandyke, Michael; Bassichis, William
2011-10-01
Calculus-based introductory physics courses intended for future engineers and physicists are often designed and taught in the same fashion as those intended for students of other disciplines. A more mathematically rigorous curriculum should be more appropriate and, ultimately, more beneficial for the student in his or her future coursework. This work investigates the effects of mathematical rigor on student understanding of introductory mechanics. Using a series of diagnostic tools in conjunction with individual student course performance, a statistical analysis will be performed to examine student learning of introductory mechanics and its relation to student understanding of the underlying calculus.
Rigorous solution to Bargmann-Wigner equation for integer spin
Huang Shi Zhong; Wu Ning; Zheng Zhi Peng
2002-01-01
A rigorous method is developed to solve the Bargamann-Wigner equation for arbitrary integer spin in coordinate representation in a step by step way. The Bargmann-Wigner equation is first transformed to a form easier to solve, the new equations are then solved rigorously in coordinate representation, and the wave functions in a closed form are thus derived
A case of instantaneous rigor?
Pirch, J; Schulz, Y; Klintschar, M
2013-09-01
The question of whether instantaneous rigor mortis (IR), the hypothetic sudden occurrence of stiffening of the muscles upon death, actually exists has been controversially debated over the last 150 years. While modern German forensic literature rejects this concept, the contemporary British literature is more willing to embrace it. We present the case of a young woman who suffered from diabetes and who was found dead in an upright standing position with back and shoulders leaned against a punchbag and a cupboard. Rigor mortis was fully established, livor mortis was strong and according to the position the body was found in. After autopsy and toxicological analysis, it was stated that death most probably occurred due to a ketoacidotic coma with markedly increased values of glucose and lactate in the cerebrospinal fluid as well as acetone in blood and urine. Whereas the position of the body is most unusual, a detailed analysis revealed that it is a stable position even without rigor mortis. Therefore, this case does not further support the controversial concept of IR.
International Nuclear Information System (INIS)
Takahashi, A.; Rusch, D.
1979-10-01
The I*-method, which is a non-approximative treatment of the neutron balance equations by the use of double-differential cross sections and a generalized angular transfer probability, is realized within the NITRAN system. It is shown, by means of test calculations for assemblies related to fusion reactor neutronics that double-differential cross section data provide substantial progress in transport problems with kinematically complicated reaction channels like (n,2n), (n,n'γ), and (n,n'α), because the I*-method is free from kinematic assumptions. The properties of the exponential method to generate the supplementary equations to the SN equations are investigated. (orig.) [de
DEFF Research Database (Denmark)
Knudsen, Søren Valgreen; Laursen, Henrik Vitus Bering; Bartels, Paul Daniel
2018-01-01
healthcare systems, where PDSA cycles are becoming central in national QI strategies. Before the health systems start to enroll these vast strategies, it is important to document whether the PDSA method provide an effect in terms of better clinical practices and outcomes. The scientific literature indicates...... that the PDSA method have not been used properly. Improper use of the method is a challenge for the internal and external validity of the method and makes it difficult to establish a relation between the use of PDSA and the effects on QI projects. However, in the recent years there has been an increased focus...... against the key features: use of iterative cycles, prediction-based tests of change, testing from small to large scale and use of data over time. The assessment was performed by two independent reviewers. Results: 106 of 176 individual studies identified met the inclusion criteria. 3/5 of these documented...
Directory of Open Access Journals (Sweden)
Charles Onyutha
2017-10-01
Full Text Available Some of the problems in drought assessments are that: analyses tend to focus on coarse temporal scales, many of the methods yield skewed indices, a few terminologies are ambiguously used, and analyses comprise an implicit assumption that the observations come from a stationary process. To solve these problems, this paper introduces non-stationary frequency analyses of quantiles. How to use non-parametric rescaling to obtain robust indices that are not (or minimally skewed is also introduced. To avoid ambiguity, some concepts on, e.g., incidence, extremity, etc., were revisited through shift from monthly to daily time scale. Demonstrations on the introduced methods were made using daily flow and precipitation insufficiency (precipitation minus potential evapotranspiration from the Blue Nile basin in Africa. Results show that, when a significant trend exists in extreme events, stationarity-based quantiles can be far different from those when non-stationarity is considered. The introduced non-parametric indices were found to closely agree with the well-known standardized precipitation evapotranspiration indices in many aspects but skewness. Apart from revisiting some concepts, the advantages of the use of fine instead of coarse time scales in drought assessment were given. The links for obtaining freely downloadable tools on how to implement the introduced methods were provided.
Krompecher, T; Bergerioux, C; Brandt-Casadevall, C; Gujer, H R
1983-07-01
The evolution of rigor mortis was studied in cases of nitrogen asphyxia, drowning and strangulation, as well as in fatal intoxications due to strychnine, carbon monoxide and curariform drugs, using a modified method of measurement. Our experiments demonstrated that: (1) Strychnine intoxication hastens the onset and passing of rigor mortis. (2) CO intoxication delays the resolution of rigor mortis. (3) The intensity of rigor may vary depending upon the cause of death. (4) If the stage of rigidity is to be used to estimate the time of death, it is necessary: (a) to perform a succession of objective measurements of rigor mortis intensity; and (b) to verify the eventual presence of factors that could play a role in the modification of its development.
Rigor in Qualitative Supply Chain Management Research
DEFF Research Database (Denmark)
Goffin, Keith; Raja, Jawwad; Claes, Björn
2012-01-01
, reliability, and theoretical saturation. Originality/value – It is the authors' contention that the addition of the repertory grid technique to the toolset of methods used by logistics and supply chain management researchers can only enhance insights and the building of robust theories. Qualitative studies......Purpose – The purpose of this paper is to share the authors' experiences of using the repertory grid technique in two supply chain management studies. The paper aims to demonstrate how the two studies provided insights into how qualitative techniques such as the repertory grid can be made more...... rigorous than in the past, and how results can be generated that are inaccessible using quantitative methods. Design/methodology/approach – This paper presents two studies undertaken using the repertory grid technique to illustrate its application in supply chain management research. Findings – The paper...
Tenderness of pre- and post rigor lamb longissimus muscle.
Geesink, Geert; Sujang, Sadi; Koohmaraie, Mohammad
2011-08-01
Lamb longissimus muscle (n=6) sections were cooked at different times post mortem (prerigor, at rigor, 1dayp.m., and 7 days p.m.) using two cooking methods. Using a boiling waterbath, samples were either cooked to a core temperature of 70 °C or boiled for 3h. The latter method was meant to reflect the traditional cooking method employed in countries where preparation of prerigor meat is practiced. The time postmortem at which the meat was prepared had a large effect on the tenderness (shear force) of the meat (PCooking prerigor and at rigor meat to 70 °C resulted in higher shear force values than their post rigor counterparts at 1 and 7 days p.m. (9.4 and 9.6 vs. 7.2 and 3.7 kg, respectively). The differences in tenderness between the treatment groups could be largely explained by a difference in contraction status of the meat after cooking and the effect of ageing on tenderness. Cooking pre and at rigor meat resulted in severe muscle contraction as evidenced by the differences in sarcomere length of the cooked samples. Mean sarcomere lengths in the pre and at rigor samples ranged from 1.05 to 1.20 μm. The mean sarcomere length in the post rigor samples was 1.44 μm. Cooking for 3 h at 100 °C did improve the tenderness of pre and at rigor prepared meat as compared to cooking to 70 °C, but not to the extent that ageing did. It is concluded that additional intervention methods are needed to improve the tenderness of prerigor cooked meat. Copyright © 2011 Elsevier B.V. All rights reserved.
Realizing rigor in the mathematics classroom
Hull, Ted H (Henry); Balka, Don S
2014-01-01
Rigor put within reach! Rigor: The Common Core has made it policy-and this first-of-its-kind guide takes math teachers and leaders through the process of making it reality. Using the Proficiency Matrix as a framework, the authors offer proven strategies and practical tools for successful implementation of the CCSS mathematical practices-with rigor as a central objective. You'll learn how to Define rigor in the context of each mathematical practice Identify and overcome potential issues, including differentiating instruction and using data
Statistics for mathematicians a rigorous first course
Panaretos, Victor M
2016-01-01
This textbook provides a coherent introduction to the main concepts and methods of one-parameter statistical inference. Intended for students of Mathematics taking their first course in Statistics, the focus is on Statistics for Mathematicians rather than on Mathematical Statistics. The goal is not to focus on the mathematical/theoretical aspects of the subject, but rather to provide an introduction to the subject tailored to the mindset and tastes of Mathematics students, who are sometimes turned off by the informal nature of Statistics courses. This book can be used as the basis for an elementary semester-long first course on Statistics with a firm sense of direction that does not sacrifice rigor. The deeper goal of the text is to attract the attention of promising Mathematics students.
Classroom Talk for Rigorous Reading Comprehension Instruction
Wolf, Mikyung Kim; Crosson, Amy C.; Resnick, Lauren B.
2004-01-01
This study examined the quality of classroom talk and its relation to academic rigor in reading-comprehension lessons. Additionally, the study aimed to characterize effective questions to support rigorous reading comprehension lessons. The data for this study included 21 reading-comprehension lessons in several elementary and middle schools from…
DEFF Research Database (Denmark)
Cappeln, Gertrud; Jessen, Flemming
2002-01-01
Variation in glycogen, ATP, and IMP contents within individual cod muscles were studied in ice stored fish during the progress of rigor mortis. Rigor index was determined before muscle samples for chemical analyzes were taken at 16 different positions on the fish. During development of rigor......, the contents of glycogen and ATP decreased differently in relation to rigor index depending on sampling location. Although fish were considered to be in strong rigor according to the rigor index method, parts of the muscle were not in rigor as high ATP concentrations were found in dorsal and tall muscle....
Rigorous Science: a How-To Guide
Directory of Open Access Journals (Sweden)
Arturo Casadevall
2016-11-01
Full Text Available Proposals to improve the reproducibility of biomedical research have emphasized scientific rigor. Although the word “rigor” is widely used, there has been little specific discussion as to what it means and how it can be achieved. We suggest that scientific rigor combines elements of mathematics, logic, philosophy, and ethics. We propose a framework for rigor that includes redundant experimental design, sound statistical analysis, recognition of error, avoidance of logical fallacies, and intellectual honesty. These elements lead to five actionable recommendations for research education.
Estimation of the breaking of rigor mortis by myotonometry.
Vain, A; Kauppila, R; Vuori, E
1996-05-31
Myotonometry was used to detect breaking of rigor mortis. The myotonometer is a new instrument which measures the decaying oscillations of a muscle after a brief mechanical impact. The method gives two numerical parameters for rigor mortis, namely the period and decrement of the oscillations, both of which depend on the time period elapsed after death. In the case of breaking the rigor mortis by muscle lengthening, both the oscillation period and decrement decreased, whereas, shortening the muscle caused the opposite changes. Fourteen h after breaking the stiffness characteristics of the right and left m. biceps brachii, or oscillation periods, were assimilated. However, the values for decrement of the muscle, reflecting the dissipation of mechanical energy, maintained their differences.
Krompecher, T
1981-01-01
Objective measurements were carried out to study the evolution of rigor mortis on rats at various temperatures. Our experiments showed that: (1) at 6 degrees C rigor mortis reaches full development between 48 and 60 hours post mortem, and is resolved at 168 hours post mortem; (2) at 24 degrees C rigor mortis reaches full development at 5 hours post mortem, and is resolved at 16 hours post mortem; (3) at 37 degrees C rigor mortis reaches full development at 3 hours post mortem, and is resolved at 6 hours post mortem; (4) the intensity of rigor mortis grows with increase in temperature (difference between values obtained at 24 degrees C and 37 degrees C); and (5) and 6 degrees C a "cold rigidity" was found, in addition to and independent of rigor mortis.
Rigor, vigor, and the study of health disparities.
Adler, Nancy; Bush, Nicole R; Pantell, Matthew S
2012-10-16
Health disparities research spans multiple fields and methods and documents strong links between social disadvantage and poor health. Associations between socioeconomic status (SES) and health are often taken as evidence for the causal impact of SES on health, but alternative explanations, including the impact of health on SES, are plausible. Studies showing the influence of parents' SES on their children's health provide evidence for a causal pathway from SES to health, but have limitations. Health disparities researchers face tradeoffs between "rigor" and "vigor" in designing studies that demonstrate how social disadvantage becomes biologically embedded and results in poorer health. Rigorous designs aim to maximize precision in the measurement of SES and health outcomes through methods that provide the greatest control over temporal ordering and causal direction. To achieve precision, many studies use a single SES predictor and single disease. However, doing so oversimplifies the multifaceted, entwined nature of social disadvantage and may overestimate the impact of that one variable and underestimate the true impact of social disadvantage on health. In addition, SES effects on overall health and functioning are likely to be greater than effects on any one disease. Vigorous designs aim to capture this complexity and maximize ecological validity through more complete assessment of social disadvantage and health status, but may provide less-compelling evidence of causality. Newer approaches to both measurement and analysis may enable enhanced vigor as well as rigor. Incorporating both rigor and vigor into studies will provide a fuller understanding of the causes of health disparities.
"Rigor mortis" in a live patient.
Chakravarthy, Murali
2010-03-01
Rigor mortis is conventionally a postmortem change. Its occurrence suggests that death has occurred at least a few hours ago. The authors report a case of "Rigor Mortis" in a live patient after cardiac surgery. The likely factors that may have predisposed such premortem muscle stiffening in the reported patient are, intense low cardiac output status, use of unusually high dose of inotropic and vasopressor agents and likely sepsis. Such an event may be of importance while determining the time of death in individuals such as described in the report. It may also suggest requirement of careful examination of patients with muscle stiffening prior to declaration of death. This report is being published to point out the likely controversies that might arise out of muscle stiffening, which should not always be termed rigor mortis and/ or postmortem.
[Rigor mortis -- a definite sign of death?].
Heller, A R; Müller, M P; Frank, M D; Dressler, J
2005-04-01
In the past years an ongoing controversial debate exists in Germany, regarding quality of the coroner's inquest and declaration of death by physicians. We report the case of a 90-year old female, who was found after an unknown time following a suicide attempt with benzodiazepine. The examination of the patient showed livores (mortis?) on the left forearm and left lower leg. Moreover, rigor (mortis?) of the left arm was apparent which prevented arm flexion and extension. The hypothermic patient with insufficient respiration was intubated and mechanically ventilated. Chest compressions were not performed, because central pulses were (hardly) palpable and a sinus bradycardia 45/min (AV-block 2 degrees and sole premature ventricular complexes) was present. After placement of an intravenous line (17 G, external jugular vein) the hemodynamic situation was stabilized with intermittent boli of epinephrine and with sodium bicarbonate. With improved circulation livores and rigor disappeared. In the present case a minimal central circulation was noted, which could be stabilized, despite the presence of certain signs of death ( livores and rigor mortis). Considering the finding of an abrogated peripheral perfusion (livores), we postulate a centripetal collapse of glycogen and ATP supply in the patients left arm (rigor), which was restored after resuscitation and reperfusion. Thus, it appears that livores and rigor are not sensitive enough to exclude a vita minima, in particular in hypothermic patients with intoxications. Consequently a careful ABC-check should be performed even in the presence of apparently certain signs of death, to avoid underdiagnosing a vita minima. Additional ECG- monitoring is required to reduce the rate of false positive declarations of death. To what extent basic life support by paramedics should commence when rigor and livores are present until physician DNR order, deserves further discussion.
An ultramicroscopic study on rigor mortis.
Suzuki, T
1976-01-01
Gastrocnemius muscles taken from decapitated mice at various intervals after death and from mice killed by 2,4-dinitrophenol or mono-iodoacetic acid injection to induce rigor mortis soon after death, were observed by electron microscopy. The prominent appearance of many fine cross striations in the myofibrils (occurring about every 400 A) was considered to be characteristic of rigor mortis. These striations were caused by minute granules studded along the surfaces of both thick and thin filaments and appeared to be the bridges connecting the 2 kinds of filaments and accounted for the hardness and rigidity of the muscle.
The Rigor Mortis of Education: Rigor Is Required in a Dying Educational System
Mixon, Jason; Stuart, Jerry
2009-01-01
In an effort to answer the "Educational Call to Arms", our national public schools have turned to Advanced Placement (AP) courses as the predominate vehicle used to address the lack of academic rigor in our public high schools. Advanced Placement is believed by many to provide students with the rigor and work ethic necessary to…
Trends: Rigor Mortis in the Arts.
Blodget, Alden S.
1991-01-01
Outlines how past art education provided a refuge for students from the rigors of other academic subjects. Observes that in recent years art education has become "discipline based." Argues that art educators need to reaffirm their commitment to a humanistic way of knowing. (KM)
Photoconductivity of amorphous silicon-rigorous modelling
International Nuclear Information System (INIS)
Brada, P.; Schauer, F.
1991-01-01
It is our great pleasure to express our gratitude to Prof. Grigorovici, the pioneer of the exciting field of amorphous state by our modest contribution to this area. In this paper are presented the outline of the rigorous modelling program of the steady-state photoconductivity in amorphous silicon and related materials. (Author)
A Rigorous Methodology for Analyzing and Designing Plug-Ins
DEFF Research Database (Denmark)
Fasie, Marieta V.; Haxthausen, Anne Elisabeth; Kiniry, Joseph
2013-01-01
. This paper addresses these problems by describing a rigorous methodology for analyzing and designing plug-ins. The methodology is grounded in the Extended Business Object Notation (EBON) and covers informal analysis and design of features, GUI, actions, and scenarios, formal architecture design, including...... behavioral semantics, and validation. The methodology is illustrated via a case study whose focus is an Eclipse environment for the RAISE formal method's tool suite....
Application of Coupled-Wave Wentzel-Kramers-Brillouin Approximation to Ground Penetrating Radar
Igor Prokopovich; Alexei Popov; Lara Pajewski; Marian Marciniak
2017-01-01
This paper deals with bistatic subsurface probing of a horizontally layered dielectric half-space by means of ultra-wideband electromagnetic waves. In particular, the main objective of this work is to present a new method for the solution of the two-dimensional back-scattering problem arising when a pulsed electromagnetic signal impinges on a non-uniform dielectric half-space; this scenario is of interest for ground penetrating radar (GPR) applications. For the analytical description of the s...
Accelerating Biomedical Discoveries through Rigor and Transparency.
Hewitt, Judith A; Brown, Liliana L; Murphy, Stephanie J; Grieder, Franziska; Silberberg, Shai D
2017-07-01
Difficulties in reproducing published research findings have garnered a lot of press in recent years. As a funder of biomedical research, the National Institutes of Health (NIH) has taken measures to address underlying causes of low reproducibility. Extensive deliberations resulted in a policy, released in 2015, to enhance reproducibility through rigor and transparency. We briefly explain what led to the policy, describe its elements, provide examples and resources for the biomedical research community, and discuss the potential impact of the policy on translatability with a focus on research using animal models. Importantly, while increased attention to rigor and transparency may lead to an increase in the number of laboratory animals used in the near term, it will lead to more efficient and productive use of such resources in the long run. The translational value of animal studies will be improved through more rigorous assessment of experimental variables and data, leading to better assessments of the translational potential of animal models, for the benefit of the research community and society. Published by Oxford University Press on behalf of the Institute for Laboratory Animal Research 2017. This work is written by (a) US Government employee(s) and is in the public domain in the US.
Software metrics a rigorous and practical approach
Fenton, Norman
2014-01-01
A Framework for Managing, Measuring, and Predicting Attributes of Software Development Products and ProcessesReflecting the immense progress in the development and use of software metrics in the past decades, Software Metrics: A Rigorous and Practical Approach, Third Edition provides an up-to-date, accessible, and comprehensive introduction to software metrics. Like its popular predecessors, this third edition discusses important issues, explains essential concepts, and offers new approaches for tackling long-standing problems.New to the Third EditionThis edition contains new material relevant
Development of rigor mortis is not affected by muscle volume.
Kobayashi, M; Ikegaya, H; Takase, I; Hatanaka, K; Sakurada, K; Iwase, H
2001-04-01
There is a hypothesis suggesting that rigor mortis progresses more rapidly in small muscles than in large muscles. We measured rigor mortis as tension determined isometrically in rat musculus erector spinae that had been cut into muscle bundles of various volumes. The muscle volume did not influence either the progress or the resolution of rigor mortis, which contradicts the hypothesis. Differences in pre-rigor load on the muscles influenced the onset and resolution of rigor mortis in a few pairs of samples, but did not influence the time taken for rigor mortis to reach its full extent after death. Moreover, the progress of rigor mortis in this muscle was biphasic; this may reflect the early rigor of red muscle fibres and the late rigor of white muscle fibres.
Krompecher, T; Fryc, O
1978-01-01
The use of new methods and an appropriate apparatus has allowed us to make successive measurements of rigor mortis and a study of its evolution in the rat. By a comparative examination on the front and hind limbs, we have determined the following: (1) The muscular mass of the hind limbs is 2.89 times greater than that of the front limbs. (2) In the initial phase rigor mortis is more pronounced in the front limbs. (3) The front and hind limbs reach maximum rigor mortis at the same time and this state is maintained for 2 hours. (4) Resolution of rigor mortis is accelerated in the front limbs during the initial phase, but both front and hind limbs reach complete resolution at the same time.
Rigorous derivation from Landau-de Gennes theory to Ericksen-Leslie theory
Wang, Wei; Zhang, Pingwen; Zhang, Zhifei
2013-01-01
Starting from Beris-Edwards system for the liquid crystal, we present a rigorous derivation of Ericksen-Leslie system with general Ericksen stress and Leslie stress by using the Hilbert expansion method.
Zhao, Huajun; Yuan, Dairong
2010-02-10
Examples of optimal designs for a fused-silica transmitted grating with high-intensity tolerance are discussed. It has the potential of placing up to 99% incident polarized light in a single diffraction order. The modal method has been used to analyze the effective indices for TE and TM polarization propagating through the grating region, and the eigenvalue equation of the modal method is transformed to a new form. It is shown that the effective indices of the first two modes depend on the value of the period under Littrow mounting with filling factor f=0.5. The polarization properties of the polarizing beam splitter are analyzed by rigorous coupled-wave analysis (RCWA) at the wavelength of 1.064 microm. The optimal design perfectly matches the RCWA simulation result.
Studies on the estimation of the postmortem interval. 3. Rigor mortis (author's transl).
Suzutani, T; Ishibashi, H; Takatori, T
1978-11-01
The authors have devised a method for classifying rigor mortis into 10 types based on its appearance and strength in various parts of a cadaver. By applying the method to the findings of 436 cadavers which were subjected to medico-legal autopsies in our laboratory during the last 10 years, it has been demonstrated that the classifying method is effective for analyzing the phenomenon of onset, persistence and disappearance of rigor mortis statistically. The investigation of the relationship between each type of rigor mortis and the postmortem interval has demonstrated that rigor mortis may be utilized as a basis for estimating the postmortem interval but the values have greater deviation than those described in current textbooks.
Sonoelasticity to monitor mechanical changes during rigor and ageing.
Ayadi, A; Culioli, J; Abouelkaram, S
2007-06-01
We propose the use of sonoelasticity as a non-destructive method to monitor changes in the resistance of muscle fibres, unaffected by connective tissue. Vibrations were applied at low frequency to induce oscillations in soft tissues and an ultrasound transducer was used to detect the motions. The experiments were carried out on the M. biceps femoris muscles of three beef cattle. In addition to the sonoelasticity measurements, the changes in meat during rigor and ageing were followed by measurements of both the mechanical resistance of myofibres and pH. The variations of mechanical resistance and pH were compared to those of the sonoelastic variables (velocity and attenuation) at two frequencies. The relationships between pH and velocity or attenuation and between the velocity or attenuation and the stress at 20% deformation were highly correlated. We concluded that sonoelasticity is a non-destructive method that can be used to monitor mechanical changes in muscle fibers during rigor-mortis and ageing.
A methodology for the rigorous verification of plasma simulation codes
Riva, Fabio
2016-10-01
The methodology used to assess the reliability of numerical simulation codes constitutes the Verification and Validation (V&V) procedure. V&V is composed by two separate tasks: the verification, which is a mathematical issue targeted to assess that the physical model is correctly solved, and the validation, which determines the consistency of the code results, and therefore of the physical model, with experimental data. In the present talk we focus our attention on the verification, which in turn is composed by the code verification, targeted to assess that a physical model is correctly implemented in a simulation code, and the solution verification, that quantifies the numerical error affecting a simulation. Bridging the gap between plasma physics and other scientific domains, we introduced for the first time in our domain a rigorous methodology for the code verification, based on the method of manufactured solutions, as well as a solution verification based on the Richardson extrapolation. This methodology was applied to GBS, a three-dimensional fluid code based on a finite difference scheme, used to investigate the plasma turbulence in basic plasma physics experiments and in the tokamak scrape-off layer. Overcoming the difficulty of dealing with a numerical method intrinsically affected by statistical noise, we have now generalized the rigorous verification methodology to simulation codes based on the particle-in-cell algorithm, which are employed to solve Vlasov equation in the investigation of a number of plasma physics phenomena.
Forster, B; Ropohl, D; Raule, P
1977-07-05
The manual examination of rigor mortis as currently used and its often subjective evaluation frequently produced highly incorrect deductions. It is therefore desirable that such inaccuracies should be replaced by the objective measuring of rigor mortis at the extremities. To that purpose a method is described which can also be applied in on-the-spot investigations and a new formula for the determination of rigor mortis--indices (FRR) is introduced.
Rigorous theory of molecular orientational nonlinear optics
International Nuclear Information System (INIS)
Kwak, Chong Hoon; Kim, Gun Yeup
2015-01-01
Classical statistical mechanics of the molecular optics theory proposed by Buckingham [A. D. Buckingham and J. A. Pople, Proc. Phys. Soc. A 68, 905 (1955)] has been extended to describe the field induced molecular orientational polarization effects on nonlinear optics. In this paper, we present the generalized molecular orientational nonlinear optical processes (MONLO) through the calculation of the classical orientational averaging using the Boltzmann type time-averaged orientational interaction energy in the randomly oriented molecular system under the influence of applied electric fields. The focal points of the calculation are (1) the derivation of rigorous tensorial components of the effective molecular hyperpolarizabilities, (2) the molecular orientational polarizations and the electronic polarizations including the well-known third-order dc polarization, dc electric field induced Kerr effect (dc Kerr effect), optical Kerr effect (OKE), dc electric field induced second harmonic generation (EFISH), degenerate four wave mixing (DFWM) and third harmonic generation (THG). We also present some of the new predictive MONLO processes. For second-order MONLO, second-order optical rectification (SOR), Pockels effect and difference frequency generation (DFG) are described in terms of the anisotropic coefficients of first hyperpolarizability. And, for third-order MONLO, third-order optical rectification (TOR), dc electric field induced difference frequency generation (EFIDFG) and pump-probe transmission are presented
Rigor or mortis: best practices for preclinical research in neuroscience.
Steward, Oswald; Balice-Gordon, Rita
2014-11-05
Numerous recent reports document a lack of reproducibility of preclinical studies, raising concerns about potential lack of rigor. Examples of lack of rigor have been extensively documented and proposals for practices to improve rigor are appearing. Here, we discuss some of the details and implications of previously proposed best practices and consider some new ones, focusing on preclinical studies relevant to human neurological and psychiatric disorders. Copyright © 2014 Elsevier Inc. All rights reserved.
[Experimental study of restiffening of the rigor mortis].
Wang, X; Li, M; Liao, Z G; Yi, X F; Peng, X M
2001-11-01
To observe changes of the length of sarcomere of rat when restiffening. We measured the length of sarcomere of quadriceps in 40 rats in different condition by scanning electron microscope. The length of sarcomere of rigor mortis without destroy is obviously shorter than that of restiffening. The length of sarcomere is negatively correlative to the intensity of rigor mortis. Measuring the length of sarcomere can determine the intensity of rigor mortis and provide evidence for estimation of time since death.
Rigorous force field optimization principles based on statistical distance minimization
Energy Technology Data Exchange (ETDEWEB)
Vlcek, Lukas, E-mail: vlcekl1@ornl.gov [Chemical Sciences Division, Geochemistry & Interfacial Sciences Group, Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831-6110 (United States); Joint Institute for Computational Sciences, University of Tennessee, Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831-6173 (United States); Chialvo, Ariel A. [Chemical Sciences Division, Geochemistry & Interfacial Sciences Group, Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831-6110 (United States)
2015-10-14
We use the concept of statistical distance to define a measure of distinguishability between a pair of statistical mechanical systems, i.e., a model and its target, and show that its minimization leads to general convergence of the model’s static measurable properties to those of the target. We exploit this feature to define a rigorous basis for the development of accurate and robust effective molecular force fields that are inherently compatible with coarse-grained experimental data. The new model optimization principles and their efficient implementation are illustrated through selected examples, whose outcome demonstrates the higher robustness and predictive accuracy of the approach compared to other currently used methods, such as force matching and relative entropy minimization. We also discuss relations between the newly developed principles and established thermodynamic concepts, which include the Gibbs-Bogoliubov inequality and the thermodynamic length.
From everyday communicative figurations to rigorous audience news repertoires
DEFF Research Database (Denmark)
Kobbernagel, Christian; Schrøder, Kim Christian
2016-01-01
In the last couple of decades there has been an unprecedented explosion of news media platforms and formats, as a succession of digital and social media have joined the ranks of legacy media. We live in a ‘hybrid media system’ (Chadwick, 2013), in which people build their cross-media news...... repertoires from the ensemble of old and new media available. This article presents an innovative mixed-method approach with considerable explanatory power to the exploration of patterns of news media consumption. This approach tailors Q-methodology in the direction of a qualitative study of news consumption......, in which a card sorting exercise serves to translate the participants’ news media preferences into a form that enables the researcher to undertake a rigorous factor-analytical construction of their news consumption repertoires. This interpretive, factor-analytical procedure, which results in the building...
Rigorous Quantum Field Theory A Festschrift for Jacques Bros
Monvel, Anne Boutet; Iagolnitzer, Daniel; Moschella, Ugo
2007-01-01
Jacques Bros has greatly advanced our present understanding of rigorous quantum field theory through numerous fundamental contributions. This book arose from an international symposium held in honour of Jacques Bros on the occasion of his 70th birthday, at the Department of Theoretical Physics of the CEA in Saclay, France. The impact of the work of Jacques Bros is evident in several articles in this book. Quantum fields are regarded as genuine mathematical objects, whose various properties and relevant physical interpretations must be studied in a well-defined mathematical framework. The key topics in this volume include analytic structures of Quantum Field Theory (QFT), renormalization group methods, gauge QFT, stability properties and extension of the axiomatic framework, QFT on models of curved spacetimes, QFT on noncommutative Minkowski spacetime. Contributors: D. Bahns, M. Bertola, R. Brunetti, D. Buchholz, A. Connes, F. Corbetta, S. Doplicher, M. Dubois-Violette, M. Dütsch, H. Epstein, C.J. Fewster, K....
Long persistence of rigor mortis at constant low temperature.
Varetto, Lorenzo; Curto, Ombretta
2005-01-06
We studied the persistence of rigor mortis by using physical manipulation. We tested the mobility of the knee on 146 corpses kept under refrigeration at Torino's city mortuary at a constant temperature of +4 degrees C. We found a persistence of complete rigor lasting for 10 days in all the cadavers we kept under observation; and in one case, rigor lasted for 16 days. Between the 11th and the 17th days, a progressively increasing number of corpses showed a change from complete into partial rigor (characterized by partial bending of the articulation). After the 17th day, all the remaining corpses showed partial rigor and in the two cadavers that were kept under observation "à outrance" we found the absolute resolution of rigor mortis occurred on the 28th day. Our results prove that it is possible to find a persistence of rigor mortis that is much longer than the expected when environmental conditions resemble average outdoor winter temperatures in temperate zones. Therefore, this datum must be considered when a corpse is found in those environmental conditions so that when estimating the time of death, we are not misled by the long persistence of rigor mortis.
Evaluating Rigor in Qualitative Methodology and Research Dissemination
Trainor, Audrey A.; Graue, Elizabeth
2014-01-01
Despite previous and successful attempts to outline general criteria for rigor, researchers in special education have debated the application of rigor criteria, the significance or importance of small n research, the purpose of interpretivist approaches, and the generalizability of qualitative empirical results. Adding to these complications, the…
Rigorous time slicing approach to Feynman path integrals
Fujiwara, Daisuke
2017-01-01
This book proves that Feynman's original definition of the path integral actually converges to the fundamental solution of the Schrödinger equation at least in the short term if the potential is differentiable sufficiently many times and its derivatives of order equal to or higher than two are bounded. The semi-classical asymptotic formula up to the second term of the fundamental solution is also proved by a method different from that of Birkhoff. A bound of the remainder term is also proved. The Feynman path integral is a method of quantization using the Lagrangian function, whereas Schrödinger's quantization uses the Hamiltonian function. These two methods are believed to be equivalent. But equivalence is not fully proved mathematically, because, compared with Schrödinger's method, there is still much to be done concerning rigorous mathematical treatment of Feynman's method. Feynman himself defined a path integral as the limit of a sequence of integrals over finite-dimensional spaces which is obtained by...
Disciplining Bioethics: Towards a Standard of Methodological Rigor in Bioethics Research
Adler, Daniel; Shaul, Randi Zlotnik
2012-01-01
Contemporary bioethics research is often described as multi- or interdisciplinary. Disciplines are characterized, in part, by their methods. Thus, when bioethics research draws on a variety of methods, it crosses disciplinary boundaries. Yet each discipline has its own standard of rigor—so when multiple disciplinary perspectives are considered, what constitutes rigor? This question has received inadequate attention, as there is considerable disagreement regarding the disciplinary status of bioethics. This disagreement has presented five challenges to bioethics research. Addressing them requires consideration of the main types of cross-disciplinary research, and consideration of proposals aiming to ensure rigor in bioethics research. PMID:22686634
Concrete ensemble Kalman filters with rigorous catastrophic filter divergence.
Kelly, David; Majda, Andrew J; Tong, Xin T
2015-08-25
The ensemble Kalman filter and ensemble square root filters are data assimilation methods used to combine high-dimensional, nonlinear dynamical models with observed data. Ensemble methods are indispensable tools in science and engineering and have enjoyed great success in geophysical sciences, because they allow for computationally cheap low-ensemble-state approximation for extremely high-dimensional turbulent forecast models. From a theoretical perspective, the dynamical properties of these methods are poorly understood. One of the central mysteries is the numerical phenomenon known as catastrophic filter divergence, whereby ensemble-state estimates explode to machine infinity, despite the true state remaining in a bounded region. In this article we provide a breakthrough insight into the phenomenon, by introducing a simple and natural forecast model that transparently exhibits catastrophic filter divergence under all ensemble methods and a large set of initializations. For this model, catastrophic filter divergence is not an artifact of numerical instability, but rather a true dynamical property of the filter. The divergence is not only validated numerically but also proven rigorously. The model cleanly illustrates mechanisms that give rise to catastrophic divergence and confirms intuitive accounts of the phenomena given in past literature.
PRO development: rigorous qualitative research as the crucial foundation.
Lasch, Kathryn Eilene; Marquis, Patrick; Vigneux, Marc; Abetz, Linda; Arnould, Benoit; Bayliss, Martha; Crawford, Bruce; Rosa, Kathleen
2010-10-01
Recently published articles have described criteria to assess qualitative research in the health field in general, but very few articles have delineated qualitative methods to be used in the development of Patient-Reported Outcomes (PROs). In fact, how PROs are developed with subject input through focus groups and interviews has been given relatively short shrift in the PRO literature when compared to the plethora of quantitative articles on the psychometric properties of PROs. If documented at all, most PRO validation articles give little for the reader to evaluate the content validity of the measures and the credibility and trustworthiness of the methods used to develop them. Increasingly, however, scientists and authorities want to be assured that PRO items and scales have meaning and relevance to subjects. This article was developed by an international, interdisciplinary group of psychologists, psychometricians, regulatory experts, a physician, and a sociologist. It presents rigorous and appropriate qualitative research methods for developing PROs with content validity. The approach described combines an overarching phenomenological theoretical framework with grounded theory data collection and analysis methods to yield PRO items and scales that have content validity.
Rigorous Electromagnetic Analysis of Uncooled Microbolometer
National Research Council Canada - National Science Library
Mirotznik, Mark
2000-01-01
.... The optical absorption of the thin-film thermal infrared detector was calculated as a function of wavelength, pixel size, incident angle and area fill factor using the finite-difference-time-domain (FDTD) method...
Krompecher, T; Bergerioux, C
1988-01-01
The influence of electrocution on the evolution of rigor mortis was studied on rats. Our experiments showed that: (1) Electrocution hastens the onset of rigor mortis. After an electrocution of 90 s, a complete rigor develops already 1 h post-mortem (p.m.) compared to 5 h p.m. for the controls. (2) Electrocution hastens the passing of rigor mortis. After an electrocution of 90 s, the first significant decrease occurs at 3 h p.m. (8 h p.m. in the controls). (3) These modifications in rigor mortis evolution are less pronounced in the limbs not directly touched by the electric current. (4) In case of post-mortem electrocution, the changes are slightly less pronounced, the resistance is higher and the absorbed energy is lower as compared with the ante-mortem electrocution cases. The results are completed by two practical observations on human electrocution cases.
Memory sparing, fast scattering formalism for rigorous diffraction modeling
Iff, W.; Kämpfe, T.; Jourlin, Y.; Tishchenko, A. V.
2017-07-01
The basics and algorithmic steps of a novel scattering formalism suited for memory sparing and fast electromagnetic calculations are presented. The formalism, called ‘S-vector algorithm’ (by analogy with the known scattering-matrix algorithm), allows the calculation of the collective scattering spectra of individual layered micro-structured scattering objects. A rigorous method of linear complexity is applied to model the scattering at individual layers; here the generalized source method (GSM) resorting to Fourier harmonics as basis functions is used as one possible method of linear complexity. The concatenation of the individual scattering events can be achieved sequentially or in parallel, both having pros and cons. The present development will largely concentrate on a consecutive approach based on the multiple reflection series. The latter will be reformulated into an implicit formalism which will be associated with an iterative solver, resulting in improved convergence. The examples will first refer to 1D grating diffraction for the sake of simplicity and intelligibility, with a final 2D application example.
Monitoring muscle optical scattering properties during rigor mortis
Xia, J.; Ranasinghesagara, J.; Ku, C. W.; Yao, G.
2007-09-01
Sarcomere is the fundamental functional unit in skeletal muscle for force generation. In addition, sarcomere structure is also an important factor that affects the eating quality of muscle food, the meat. The sarcomere structure is altered significantly during rigor mortis, which is the critical stage involved in transforming muscle to meat. In this paper, we investigated optical scattering changes during the rigor process in Sternomandibularis muscles. The measured optical scattering parameters were analyzed along with the simultaneously measured passive tension, pH value, and histology analysis. We found that the temporal changes of optical scattering, passive tension, pH value and fiber microstructures were closely correlated during the rigor process. These results suggested that sarcomere structure changes during rigor mortis can be monitored and characterized by optical scattering, which may find practical applications in predicting meat quality.
Increased scientific rigor will improve reliability of research and effectiveness of management
Sells, Sarah N.; Bassing, Sarah B.; Barker, Kristin J.; Forshee, Shannon C.; Keever, Allison; Goerz, James W.; Mitchell, Michael S.
2018-01-01
Rigorous science that produces reliable knowledge is critical to wildlife management because it increases accurate understanding of the natural world and informs management decisions effectively. Application of a rigorous scientific method based on hypothesis testing minimizes unreliable knowledge produced by research. To evaluate the prevalence of scientific rigor in wildlife research, we examined 24 issues of the Journal of Wildlife Management from August 2013 through July 2016. We found 43.9% of studies did not state or imply a priori hypotheses, which are necessary to produce reliable knowledge. We posit that this is due, at least in part, to a lack of common understanding of what rigorous science entails, how it produces more reliable knowledge than other forms of interpreting observations, and how research should be designed to maximize inferential strength and usefulness of application. Current primary literature does not provide succinct explanations of the logic behind a rigorous scientific method or readily applicable guidance for employing it, particularly in wildlife biology; we therefore synthesized an overview of the history, philosophy, and logic that define scientific rigor for biological studies. A rigorous scientific method includes 1) generating a research question from theory and prior observations, 2) developing hypotheses (i.e., plausible biological answers to the question), 3) formulating predictions (i.e., facts that must be true if the hypothesis is true), 4) designing and implementing research to collect data potentially consistent with predictions, 5) evaluating whether predictions are consistent with collected data, and 6) drawing inferences based on the evaluation. Explicitly testing a priori hypotheses reduces overall uncertainty by reducing the number of plausible biological explanations to only those that are logically well supported. Such research also draws inferences that are robust to idiosyncratic observations and
Physiological studies of muscle rigor mortis in the fowl
International Nuclear Information System (INIS)
Nakahira, S.; Kaneko, K.; Tanaka, K.
1990-01-01
A simple system was developed for continuous measurement of muscle contraction during nor mortis. Longitudinal muscle strips dissected from the Peroneus Longus were suspended in a plastic tube containing liquid paraffin. Mechanical activity was transmitted to a strain-gauge transducer which is connected to a potentiometric pen-recorder. At the onset of measurement 1.2g was loaded on the muscle strip. This model was used to study the muscle response to various treatments during nor mortis. All measurements were carried out under the anaerobic condition at 17°C, except otherwise stated. 1. The present system was found to be quite useful for continuous measurement of muscle rigor course. 2. Muscle contraction under the anaerobic condition at 17°C reached a peak about 2 hours after the onset of measurement and thereafter it relaxed at a slow rate. In contrast, the aerobic condition under a high humidity resulted in a strong rigor, about three times stronger than that in the anaerobic condition. 3. Ultrasonic treatment (37, 000-47, 000Hz) at 25°C for 10 minutes resulted in a moderate muscle rigor. 4. Treatment of muscle strip with 2mM EGTA at 30°C for 30 minutes led to a relaxation of the muscle. 5. The muscle from the birds killed during anesthesia with pentobarbital sodium resulted in a slow rate of rigor, whereas the birds killed one day after hypophysectomy led to a quick muscle rigor as seen in intact controls. 6. A slight muscle rigor was observed when muscle strip was placed in a refrigerator at 0°C for 18.5 hours and thereafter temperature was kept at 17°C. (author)
A rigorous pole representation of multilevel cross sections and its practical applications
International Nuclear Information System (INIS)
Hwang, R.N.
1987-01-01
In this article a rigorous method for representing the multilevel cross sections and its practical applications are described. It is a generalization of the rationale suggested by de Saussure and Perez for the s-wave resonances. A computer code WHOPPER has been developed to convert the Reich-Moore parameters into the pole and residue parameters in momentum space. Sample calculations have been carried out to illustrate that the proposed method preserves the rigor of the Reich-Moore cross sections exactly. An analytical method has been developed to evaluate the pertinent Doppler-broadened line shape functions. A discussion is presented on how to minimize the number of pole parameters so that the existing reactor codes can be best utilized
Rigorous simulations of a helical core fiber by the use of transformation optics formalism.
Napiorkowski, Maciej; Urbanczyk, Waclaw
2014-09-22
We report for the first time on rigorous numerical simulations of a helical-core fiber by using a full vectorial method based on the transformation optics formalism. We modeled the dependence of circular birefringence of the fundamental mode on the helix pitch and analyzed the effect of a birefringence increase caused by the mode displacement induced by a core twist. Furthermore, we analyzed the complex field evolution versus the helix pitch in the first order modes, including polarization and intensity distribution. Finally, we show that the use of the rigorous vectorial method allows to better predict the confinement loss of the guided modes compared to approximate methods based on equivalent in-plane bending models.
Validation of a coupled wave-flow model in a high-energy setting: the mouth of the Columbia River
Elias, Edwin P.L.; Gelfenbaum, Guy R.; van der Westhuysen, André J.
2012-01-01
A monthlong time series of wave, current, salinity, and suspended-sediment measurements was made at five sites on a transect across the Mouth of Columbia River (MCR). These data were used to calibrate and evaluate the performance of a coupled hydrodynamic and wave model for the MCR based on the Delft3D modeling system. The MCR is a dynamic estuary inlet in which tidal currents, river discharge, and wave-driven currents are all important. Model tuning consisted primarily of spatial adjustments to bottom drag coefficients. In combination with (near-) default parameter settings, the MCR model application is able to simulate the dominant features in the tidal flow, salinity and wavefields observed in field measurements. The wave-orbital averaged method for representing the current velocity profile in the wave model is considered the most realistic for the MCR. The hydrodynamic model is particularly effective in reproducing the observed vertical residual and temporal variations in current structure. Density gradients introduce the observed and modeled reversal of the mean flow at the bed and augment mean and peak flow in the upper half of the water column. This implies that sediment transport during calmer summer conditions is controlled by density stratification and is likely net landward due to the reversal of flow near the bed. The correspondence between observed and modeled hydrodynamics makes this application a tool to investigate hydrodynamics and associated sediment transport.
Reconciling the Rigor-Relevance Dilemma in Intellectual Capital Research
Andriessen, Daniel
2004-01-01
This paper raises the issue of research methodology for intellectual capital and other types of management research by focusing on the dilemma of rigour versus relevance. The more traditional explanatory approach to research often leads to rigorous results that are not of much help to solve practical problems. This paper describes an alternative…
Paper 3: Content and Rigor of Algebra Credit Recovery Courses
Walters, Kirk; Stachel, Suzanne
2014-01-01
This paper describes the content, organization and rigor of the f2f and online summer algebra courses that were delivered in summers 2011 and 2012. Examining the content of both types of courses is important because research suggests that algebra courses with certain features may be better than others in promoting success for struggling students.…
A rigorous treatment of uncertainty quantification for Silicon damage metrics
International Nuclear Information System (INIS)
Griffin, P.
2016-01-01
These report summaries the contributions made by Sandia National Laboratories in support of the International Atomic Energy Agency (IAEA) Nuclear Data Section (NDS) Technical Meeting (TM) on Nuclear Reaction Data and Uncertainties for Radiation Damage. This work focused on a rigorous treatment of the uncertainties affecting the characterization of the displacement damage seen in silicon semiconductors. (author)
Effects of post mortem temperature on rigor tension, shortening and ...
African Journals Online (AJOL)
Fully developed rigor mortis in muscle is characterised by maximum loss of extensibility. The course of post mortem changes in ostrich muscle was studied by following isometric tension, shortening and change in pH during the first 24 h post mortem within muscle strips from the muscularis gastrocnemius, pars interna at ...
Characterization of rigor mortis of longissimus dorsi and triceps ...
African Journals Online (AJOL)
24 h) of the longissimus dorsi (LD) and triceps brachii (TB) muscles as well as the shear force (meat tenderness) and colour were evaluated, aiming at characterizing the rigor mortis in the meat during industrial processing. Data statistic treatment demonstrated that carcass temperature and pH decreased gradually during ...
A rigorous proof for the Landauer-Büttiker formula
DEFF Research Database (Denmark)
Cornean, Horia Decebal; Jensen, Arne; Moldoveanu, V.
Recently, Avron et al. shed new light on the question of quantum transport in mesoscopic samples coupled to particle reservoirs by semi-infinite leads. They rigorously treat the case when the sample undergoes an adiabatic evolution thus generating a current through th leads, and prove the so call...
Cavitt, L C; Sams, A R
2003-07-01
Studies were conducted to develop a non-destructive method for monitoring the rate of rigor mortis development in poultry and to evaluate the effectiveness of electrical stimulation (ES). In the first study, 36 male broilers in each of two trials were processed at 7 wk of age. After being bled, half of the birds received electrical stimulation (400 to 450 V, 400 to 450 mA, for seven pulses of 2 s on and 1 s off), and the other half were designated as controls. At 0.25 and 1.5 h postmortem (PM), carcasses were evaluated for the angles of the shoulder, elbow, and wing tip and the distance between the elbows. Breast fillets were harvested at 1.5 h PM (after chilling) from all carcasses. Fillet samples were excised and frozen for later measurement of pH and R-value, and the remainder of each fillet was held on ice until 24 h postmortem. Shear value and pH means were significantly lower, but R-value means were higher (P rigor mortis by ES. The physical dimensions of the shoulder and elbow changed (P rigor mortis development and with ES. These results indicate that physical measurements of the wings maybe useful as a nondestructive indicator of rigor development and for monitoring the effectiveness of ES. In the second study, 60 male broilers in each of two trials were processed at 7 wk of age. At 0.25, 1.5, 3.0, and 6.0 h PM, carcasses were evaluated for the distance between the elbows. At each time point, breast fillets were harvested from each carcass. Fillet samples were excised and frozen for later measurement of pH and sacromere length, whereas the remainder of each fillet was held on ice until 24 h PM. Shear value and pH means (P rigor mortis development. Elbow distance decreased (P rigor development and was correlated (P rigor mortis development in broiler carcasses.
Rigorous simulation: a tool to enhance decision making
Energy Technology Data Exchange (ETDEWEB)
Neiva, Raquel; Larson, Mel; Baks, Arjan [KBC Advanced Technologies plc, Surrey (United Kingdom)
2012-07-01
The world refining industries continue to be challenged by population growth (increased demand), regional market changes and the pressure of regulatory requirements to operate a 'green' refinery. Environmental regulations are reducing the value and use of heavy fuel oils, and leading to convert more of the heavier products or even heavier crude into lighter products while meeting increasingly stringent transportation fuel specifications. As a result actions are required for establishing a sustainable advantage for future success. Rigorous simulation provides a key advantage improving the time and efficient use of capital investment and maximizing profitability. Sustainably maximizing profit through rigorous modeling is achieved through enhanced performance monitoring and improved Linear Programme (LP) model accuracy. This paper contains examples on these two items. The combination of both increases overall rates of return. As refiners consider optimizing existing assets and expanding projects, the process agreed to achieve these goals is key for a successful profit improvement. The benefit of rigorous kinetic simulation with detailed fractionation allows for optimizing existing asset utilization while focusing the capital investment in the new unit(s), and therefore optimizing the overall strategic plan and return on investment. Individual process unit's monitoring works as a mechanism for validating and optimizing the plant performance. Unit monitoring is important to rectify poor performance and increase profitability. The key to a good LP relies upon the accuracy of the data used to generate the LP sub-model data. The value of rigorous unit monitoring are that the results are heat and mass balanced consistently, and are unique for a refiners unit / refinery. With the improved match of the refinery operation, the rigorous simulation models will allow capturing more accurately the non linearity of those process units and therefore provide correct
Estimation of the time since death--reconsidering the re-establishment of rigor mortis.
Anders, Sven; Kunz, Michaela; Gehl, Axel; Sehner, Susanne; Raupach, Tobias; Beck-Bornholdt, Hans-Peter
2013-01-01
In forensic medicine, there is an undefined data background for the phenomenon of re-establishment of rigor mortis after mechanical loosening, a method used in establishing time since death in forensic casework that is thought to occur up to 8 h post-mortem. Nevertheless, the method is widely described in textbooks on forensic medicine. We examined 314 joints (elbow and knee) of 79 deceased at defined time points up to 21 h post-mortem (hpm). Data were analysed using a random intercept model. Here, we show that re-establishment occurred in 38.5% of joints at 7.5 to 19 hpm. Therefore, the maximum time span for the re-establishment of rigor mortis appears to be 2.5-fold longer than thought so far. These findings have major impact on the estimation of time since death in forensic casework.
Modulational instability of coupled waves
International Nuclear Information System (INIS)
McKinstrie, C.J.; Bingham, R.
1989-01-01
The collinear propagation of an arbitrary number of finite-amplitude waves is modeled by a system of coupled nonlinear Schroedinger equations; one equation for each complex wave amplitude. In general, the waves are modulationally unstable with a maximal growth rate larger than the modulational growth rate of any wave alone. Moreover, waves that are modulationally stable by themselves can be driven unstable by the nonlinear coupling. The general theory is then applied to the relativistic modulational instability of two laser beams in a beat-wave accelerator. For parameters typical of a proposed beat-wave accelerator, this instability can seriously distort the incident laser pulse shapes on the particle-acceleration time scale, with detrimental consequences for particle acceleration
Institute of Scientific and Technical Information of China (English)
DENG Zeng'an; XIE Li'an; HAN Guijun; ZHANG Xuefeng; WU Kejian
2012-01-01
We investigated the Stokes drift-driven ocean currents and Stokes drift-induced wind energy input into the upper ocean using a two-way coupled wave-current modeling system that consists of the Princeton Ocean Model generalized coordinate system (POMgcs),Simulating WAves Nearshore (SWAN) wave model,and the Model Coupling Toolkit (MCT).The Coriolis-Stokes forcing (CSF) computed using the wave parameters from SWAN was incorporated with the momentum equation of POMgcs as the core coupling process.Experimental results in an idealized setting show that under the steady state,the scale of the speed of CSF-driven current was 0.001 m/s and the maximum reached 0.02 rn/s.The Stokes drift-induced energy rate input into the model ocean was estimated to be 28.5 GW,taking 14％ of the direct wind energy rate input.Considering the Stokes drift effects,the total mechanical energy rate input was increased by approximately 14％,which highlights the importance of CSF in modulating the upper ocean circulation.The actual run conducted in Taiwan Adjacent Sea (TAS) shows that:1) CSF-based wave-current coupling has an impact on ocean surface currents,which is related to the activities of monsoon winds; 2) wave-current coupling plays a significant role in a place where strong eddies present and tends to intensify the eddy's vorticity; 3) wave-current coupling affects the volume transport of the Taiwan Strait (TS) throughflow in a nontrivial degree,3.75％ on average.
Einstein's Theory A Rigorous Introduction for the Mathematically Untrained
Grøn, Øyvind
2011-01-01
This book provides an introduction to the theory of relativity and the mathematics used in its processes. Three elements of the book make it stand apart from previously published books on the theory of relativity. First, the book starts at a lower mathematical level than standard books with tensor calculus of sufficient maturity to make it possible to give detailed calculations of relativistic predictions of practical experiments. Self-contained introductions are given, for example vector calculus, differential calculus and integrations. Second, in-between calculations have been included, making it possible for the non-technical reader to follow step-by-step calculations. Thirdly, the conceptual development is gradual and rigorous in order to provide the inexperienced reader with a philosophically satisfying understanding of the theory. Einstein's Theory: A Rigorous Introduction for the Mathematically Untrained aims to provide the reader with a sound conceptual understanding of both the special and genera...
Rigor mortis in an unusual position: Forensic considerations.
D'Souza, Deepak H; Harish, S; Rajesh, M; Kiran, J
2011-07-01
We report a case in which the dead body was found with rigor mortis in an unusual position. The dead body was lying on its back with limbs raised, defying gravity. Direction of the salivary stains on the face was also defying the gravity. We opined that the scene of occurrence of crime is unlikely to be the final place where the dead body was found. The clues were revealing a homicidal offence and an attempt to destroy the evidence. The forensic use of 'rigor mortis in an unusual position' is in furthering the investigations, and the scientific confirmation of two facts - the scene of death (occurrence) is different from the scene of disposal of dead body, and time gap between the two places.
Some rigorous results concerning spectral theory for ideal MHD
International Nuclear Information System (INIS)
Laurence, P.
1986-01-01
Spectral theory for linear ideal MHD is laid on a firm foundation by defining appropriate function spaces for the operators associated with both the first- and second-order (in time and space) partial differential operators. Thus, it is rigorously established that a self-adjoint extension of F(xi) exists. It is shown that the operator L associated with the first-order formulation satisfies the conditions of the Hille--Yosida theorem. A foundation is laid thereby within which the domains associated with the first- and second-order formulations can be compared. This allows future work in a rigorous setting that will clarify the differences (in the two formulations) between the structure of the generalized eigenspaces corresponding to the marginal point of the spectrum ω = 0
Some rigorous results concerning spectral theory for ideal MHD
International Nuclear Information System (INIS)
Laurence, P.
1985-05-01
Spectral theory for linear ideal MHD is laid on a firm foundation by defining appropriate function spaces for the operators associated with both the first and second order (in time and space) partial differential operators. Thus, it is rigorously established that a self-adjoint extension of F(xi) exists. It is shown that the operator L associated with the first order formulation satisfies the conditions of the Hille-Yosida theorem. A foundation is laid thereby within which the domains associated with the first and second order formulations can be compared. This allows future work in a rigorous setting that will clarify the differences (in the two formulations) between the structure of the generalized eigenspaces corresponding to the marginal point of the spectrum ω = 0
Rigorous results on measuring the quark charge below color threshold
International Nuclear Information System (INIS)
Lipkin, H.J.
1979-01-01
Rigorous theorems are presented showing that contributions from a color nonsinglet component of the current to matrix elements of a second order electromagnetic transition are suppressed by factors inversely proportional to the energy of the color threshold. Parton models which obtain matrix elements proportional to the color average of the square of the quark charge are shown to neglect terms of the same order of magnitude as terms kept. (author)
Striation Patterns of Ox Muscle in Rigor Mortis
Locker, Ronald H.
1959-01-01
Ox muscle in rigor mortis offers a selection of myofibrils fixed at varying degrees of contraction from sarcomere lengths of 3.7 to 0.7 µ. A study of this material by phase contrast and electron microscopy has revealed four distinct successive patterns of contraction, including besides the familiar relaxed and contracture patterns, two intermediate types (2.4 to 1.9 µ, 1.8 to 1.5 µ) not previously well described. PMID:14417790
Rigorous Analysis of a Randomised Number Field Sieve
Lee, Jonathan; Venkatesan, Ramarathnam
2018-01-01
Factorisation of integers $n$ is of number theoretic and cryptographic significance. The Number Field Sieve (NFS) introduced circa 1990, is still the state of the art algorithm, but no rigorous proof that it halts or generates relationships is known. We propose and analyse an explicitly randomised variant. For each $n$, we show that these randomised variants of the NFS and Coppersmith's multiple polynomial sieve find congruences of squares in expected times matching the best-known heuristic e...
Reciprocity relations in transmission electron microscopy: A rigorous derivation.
Krause, Florian F; Rosenauer, Andreas
2017-01-01
A concise derivation of the principle of reciprocity applied to realistic transmission electron microscopy setups is presented making use of the multislice formalism. The equivalence of images acquired in conventional and scanning mode is thereby rigorously shown. The conditions for the applicability of the found reciprocity relations is discussed. Furthermore the positions of apertures in relation to the corresponding lenses are considered, a subject which scarcely has been addressed in previous publications. Copyright Â© 2016 Elsevier Ltd. All rights reserved.
Critical Analysis of Strategies for Determining Rigor in Qualitative Inquiry.
Morse, Janice M
2015-09-01
Criteria for determining the trustworthiness of qualitative research were introduced by Guba and Lincoln in the 1980s when they replaced terminology for achieving rigor, reliability, validity, and generalizability with dependability, credibility, and transferability. Strategies for achieving trustworthiness were also introduced. This landmark contribution to qualitative research remains in use today, with only minor modifications in format. Despite the significance of this contribution over the past four decades, the strategies recommended to achieve trustworthiness have not been critically examined. Recommendations for where, why, and how to use these strategies have not been developed, and how well they achieve their intended goal has not been examined. We do not know, for example, what impact these strategies have on the completed research. In this article, I critique these strategies. I recommend that qualitative researchers return to the terminology of social sciences, using rigor, reliability, validity, and generalizability. I then make recommendations for the appropriate use of the strategies recommended to achieve rigor: prolonged engagement, persistent observation, and thick, rich description; inter-rater reliability, negative case analysis; peer review or debriefing; clarifying researcher bias; member checking; external audits; and triangulation. © The Author(s) 2015.
Krompecher, T
1994-10-21
The development of the intensity of rigor mortis was monitored in nine groups of rats. The measurements were initiated after 2, 4, 5, 6, 8, 12, 15, 24, and 48 h post mortem (p.m.) and lasted 5-9 h, which ideally should correspond to the usual procedure after the discovery of a corpse. The experiments were carried out at an ambient temperature of 24 degrees C. Measurements initiated early after death resulted in curves with a rising portion, a plateau, and a descending slope. Delaying the initial measurement translated into shorter rising portions, and curves initiated 8 h p.m. or later are comprised of a plateau and/or a downward slope only. Three different phases were observed suggesting simple rules that can help estimate the time since death: (1) if an increase in intensity was found, the initial measurements were conducted not later than 5 h p.m.; (2) if only a decrease in intensity was observed, the initial measurements were conducted not earlier than 7 h p.m.; and (3) at 24 h p.m., the resolution is complete, and no further changes in intensity should occur. Our results clearly demonstrate that repeated measurements of the intensity of rigor mortis allow a more accurate estimation of the time since death of the experimental animals than the single measurement method used earlier. A critical review of the literature on the estimation of time since death on the basis of objective measurements of the intensity of rigor mortis is also presented.
Re-establishment of rigor mortis: evidence for a considerably longer post-mortem time span.
Crostack, Chiara; Sehner, Susanne; Raupach, Tobias; Anders, Sven
2017-07-01
Re-establishment of rigor mortis following mechanical loosening is used as part of the complex method for the forensic estimation of the time since death in human bodies and has formerly been reported to occur up to 8-12 h post-mortem (hpm). We recently described our observation of the phenomenon in up to 19 hpm in cases with in-hospital death. Due to the case selection (preceding illness, immobilisation), transfer of these results to forensic cases might be limited. We therefore examined 67 out-of-hospital cases of sudden death with known time points of death. Re-establishment of rigor mortis was positive in 52.2% of cases and was observed up to 20 hpm. In contrast to the current doctrine that a recurrence of rigor mortis is always of a lesser degree than its first manifestation in a given patient, muscular rigidity at re-establishment equalled or even exceeded the degree observed before dissolving in 21 joints. Furthermore, this is the first study to describe that the phenomenon appears to be independent of body or ambient temperature.
New rigorous asymptotic theorems for inverse scattering amplitudes
International Nuclear Information System (INIS)
Lomsadze, Sh.Yu.; Lomsadze, Yu.M.
1984-01-01
The rigorous asymptotic theorems both of integral and local types obtained earlier and establishing logarithmic and in some cases even power correlations aetdeen the real and imaginary parts of scattering amplitudes Fsub(+-) are extended to the inverse amplitudes 1/Fsub(+-). One also succeeds in establishing power correlations of a new type between the real and imaginary parts, both for the amplitudes themselves and for the inverse ones. All the obtained assertions are convenient to be tested in high energy experiments when the amplitudes show asymptotic behaviour
Rigorous quantum limits on monitoring free masses and harmonic oscillators
Roy, S. M.
2018-03-01
There are heuristic arguments proposing that the accuracy of monitoring position of a free mass m is limited by the standard quantum limit (SQL): σ2( X (t ) ) ≥σ2( X (0 ) ) +(t2/m2) σ2( P (0 ) ) ≥ℏ t /m , where σ2( X (t ) ) and σ2( P (t ) ) denote variances of the Heisenberg representation position and momentum operators. Yuen [Phys. Rev. Lett. 51, 719 (1983), 10.1103/PhysRevLett.51.719] discovered that there are contractive states for which this result is incorrect. Here I prove universally valid rigorous quantum limits (RQL), viz. rigorous upper and lower bounds on σ2( X (t ) ) in terms of σ2( X (0 ) ) and σ2( P (0 ) ) , given by Eq. (12) for a free mass and by Eq. (36) for an oscillator. I also obtain the maximally contractive and maximally expanding states which saturate the RQL, and use the contractive states to set up an Ozawa-type measurement theory with accuracies respecting the RQL but beating the standard quantum limit. The contractive states for oscillators improve on the Schrödinger coherent states of constant variance and may be useful for gravitational wave detection and optical communication.
A rigorous test for a new conceptual model for collisions
International Nuclear Information System (INIS)
Peixoto, E.M.A.; Mu-Tao, L.
1979-01-01
A rigorous theoretical foundation for the previously proposed model is formulated and applied to electron scattering by H 2 in the gas phase. An rigorous treatment of the interaction potential between the incident electron and the Hydrogen molecule is carried out to calculate Differential Cross Sections for 1 KeV electrons, using Glauber's approximation Wang's molecular wave function for the ground electronic state of H 2 . Moreover, it is shown for the first time that, when adequately done, the omission of two center terms does not adversely influence the results of molecular calculations. It is shown that the new model is far superior to the Independent Atom Model (or Independent Particle Model). The accuracy and simplicity of the new model suggest that it may be fruitfully applied to the description of other collision phenomena (e.g., in molecular beam experiments and nuclear physics). A new techniques is presented for calculations involving two center integrals within the frame work of the Glauber's approximation for scattering. (Author) [pt
Krompecher, Thomas; Gilles, André; Brandt-Casadevall, Conception; Mangin, Patrice
2008-04-07
Objective measurements were carried out to study the possible re-establishment of rigor mortis on rats after "breaking" (mechanical solution). Our experiments showed that: *Cadaveric rigidity can re-establish after breaking. *A significant rigidity can reappear if the breaking occurs before the process is complete. *Rigidity will be considerably weaker after the breaking. *The time course of the intensity does not change in comparison to the controls: --the re-establishment begins immediately after the breaking; --maximal values are reached at the same time as in the controls; --the course of the resolution is the same as in the controls.
Direct integration of the S-matrix applied to rigorous diffraction
International Nuclear Information System (INIS)
Iff, W; Lindlein, N; Tishchenko, A V
2014-01-01
A novel Fourier method for rigorous diffraction computation at periodic structures is presented. The procedure is based on a differential equation for the S-matrix, which allows direct integration of the S-matrix blocks. This results in a new method in Fourier space, which can be considered as a numerically stable and well-parallelizable alternative to the conventional differential method based on T-matrix integration and subsequent conversions from the T-matrices to S-matrix blocks. Integration of the novel differential equation in implicit manner is expounded. The applicability of the new method is shown on the basis of 1D periodic structures. It is clear however, that the new technique can also be applied to arbitrary 2D periodic or periodized structures. The complexity of the new method is O(N 3 ) similar to the conventional differential method with N being the number of diffraction orders. (fast track communication)
Rigorous spin-spin correlation function of Ising model on a special kind of Sierpinski Carpets
International Nuclear Information System (INIS)
Yang, Z.R.
1993-10-01
We have exactly calculated the rigorous spin-spin correlation function of Ising model on a special kind of Sierpinski Carpets (SC's) by means of graph expansion and a combinatorial approach and investigated the asymptotic behaviour in the limit of long distance. The result show there is no long range correlation between spins at any finite temperature which indicates no existence of phase transition and thus finally confirms the conclusion produced by the renormalization group method and other physical arguments. (author). 7 refs, 6 figs
Reframing Rigor: A Modern Look at Challenge and Support in Higher Education
Campbell, Corbin M.; Dortch, Deniece; Burt, Brian A.
2018-01-01
This chapter describes the limitations of the traditional notions of academic rigor in higher education, and brings forth a new form of rigor that has the potential to support student success and equity.
Design of a Fused-Silica Subwavelength Polarizing Beam Splitter Grating Based on the Modal Method
International Nuclear Information System (INIS)
Hua-Jun, Zhao; Dai-Rong, Yuan; Pei, Wang; Yong-Hua, Lu; Hai, Ming
2010-01-01
A polarizing beam splitter (PBS) design based on a fused-silica lamellar subwavelength transmission grating is demonstrated with the modal method, where TE- and TM-polarized waves are mainly diffracted in the −1 st and 0 th orders, respectively. The physical explanation of the grating diffraction is illustrated by the interference of the corresponding parts of the two propagating modes, which is very similar to a Mach–Zehnder interferometer. It is shown that diffraction efficients over 99% for a TM-polarized wave in the −1 st order and 90% for a TE-polarized wave in the 0 th order are obtained at the wavelength of 1.053 μm. The polarization transmission extinction ratios are better than 33 dB and 51 dB for the order 0 th and the −1 st order, respectively. The splitting properties of the PBS grating designed by the modal method are in good agreement with the results simulated by the rigorous coupled wave analysis method. (fundamental areas of phenomenology(including applications))
Fast and Rigorous Assignment Algorithm Multiple Preference and Calculation
Directory of Open Access Journals (Sweden)
Ümit Çiftçi
2010-03-01
Full Text Available The goal of paper is to develop an algorithm that evaluates students then places them depending on their desired choices according to dependant preferences. The developed algorithm is also used to implement software. The success and accuracy of the software as well as the algorithm are tested by applying it to ability test at Beykent University. This ability test is repeated several times in order to fill all available places at Fine Art Faculty departments in every academic year. It has been shown that this algorithm is very fast and rigorous after application of 2008-2009 and 2009-20010 academic years.Key Words: Assignment algorithm, student placement, ability test
Student’s rigorous mathematical thinking based on cognitive style
Fitriyani, H.; Khasanah, U.
2017-12-01
The purpose of this research was to determine the rigorous mathematical thinking (RMT) of mathematics education students in solving math problems in terms of reflective and impulsive cognitive styles. The research used descriptive qualitative approach. Subjects in this research were 4 students of the reflective and impulsive cognitive style which was each consisting male and female subjects. Data collection techniques used problem-solving test and interview. Analysis of research data used Miles and Huberman model that was reduction of data, presentation of data, and conclusion. The results showed that impulsive male subjects used three levels of the cognitive function required for RMT that were qualitative thinking, quantitative thinking with precision, and relational thinking completely while the other three subjects were only able to use cognitive function at qualitative thinking level of RMT. Therefore the subject of impulsive male has a better RMT ability than the other three research subjects.
Desarrollo constitucional, legal y jurisprudencia del principio de rigor subsidiario
Directory of Open Access Journals (Sweden)
Germán Eduardo Cifuentes Sandoval
2013-09-01
Full Text Available In colombia the environment state administration is in charge of environmental national system, SINA, SINA is made up of states entities that coexist beneath a mixed organization of centralization and decentralization. SINA decentralization express itself in a administrative and territorial level, and is waited that entities that function under this structure act in a coordinated way in order to reach suggested objectives in the environmental national politicy. To achieve the coordinated environmental administration through entities that define the SINA, the environmental legislation of Colombia has include three basic principles: 1. The principle of “armorial regional” 2. The principle of “gradationnormative” 3. The principle of “rigorsubsidiaries”. These principles belong to the article 63, law 99 of 1933, and even in the case of the two first, it is possible to find equivalents in other norms that integrate the Colombian legal system, it does not happen in that way with the “ rigor subsidiaries” because its elements are uniques of the environmental normativity and do not seem to be similar to those that make part of the principle of “ subsidiaridad” present in the article 288 of the politic constitution. The “ rigor subsidiaries” give to decentralizates entities certain type of special ability to modify the current environmental legislation to defend the local ecological patrimony. It is an administrative ability with a foundation in the decentralization autonomy that allows to take place of the reglamentary denied of the legislative power with the condition that the new normativity be more demanding that the one that belongs to the central level
Rigorous patient-prosthesis matching of Perimount Magna aortic bioprosthesis.
Nakamura, Hiromasa; Yamaguchi, Hiroki; Takagaki, Masami; Kadowaki, Tasuku; Nakao, Tatsuya; Amano, Atsushi
2015-03-01
Severe patient-prosthesis mismatch, defined as effective orifice area index ≤0.65 cm(2) m(-2), has demonstrated poor long-term survival after aortic valve replacement. Reported rates of severe mismatch involving the Perimount Magna aortic bioprosthesis range from 4% to 20% in patients with a small annulus. Between June 2008 and August 2011, 251 patients (mean age 70.5 ± 10.2 years; mean body surface area 1.55 ± 0.19 m(2)) underwent aortic valve replacement with a Perimount Magna bioprosthesis, with or without concomitant procedures. We performed our procedure with rigorous patient-prosthesis matching to implant a valve appropriately sized to each patient, and carried out annular enlargement when a 19-mm valve did not fit. The bioprosthetic performance was evaluated by transthoracic echocardiography predischarge and at 1 and 2 years after surgery. Overall hospital mortality was 1.6%. Only 5 (2.0%) patients required annular enlargement. The mean follow-up period was 19.1 ± 10.7 months with a 98.4% completion rate. Predischarge data showed a mean effective orifice area index of 1.21 ± 0.20 cm(2) m(-2). Moderate mismatch, defined as effective orifice area index ≤0.85 cm(2) m(-2), developed in 4 (1.6%) patients. None developed severe mismatch. Data at 1 and 2 years showed only two cases of moderate mismatch; neither was severe. Rigorous patient-prosthesis matching maximized the performance of the Perimount Magna, and no severe mismatch resulted in this Japanese population of aortic valve replacement patients. © The Author(s) 2014 Reprints and permissions: sagepub.co.uk/journalsPermissions.nav.
Rigorous high-precision enclosures of fixed points and their invariant manifolds
Wittig, Alexander N.
The well established concept of Taylor Models is introduced, which offer highly accurate C0 enclosures of functional dependencies, combining high-order polynomial approximation of functions and rigorous estimates of the truncation error, performed using verified arithmetic. The focus of this work is on the application of Taylor Models in algorithms for strongly non-linear dynamical systems. A method is proposed to extend the existing implementation of Taylor Models in COSY INFINITY from double precision coefficients to arbitrary precision coefficients. Great care is taken to maintain the highest efficiency possible by adaptively adjusting the precision of higher order coefficients in the polynomial expansion. High precision operations are based on clever combinations of elementary floating point operations yielding exact values for round-off errors. An experimental high precision interval data type is developed and implemented. Algorithms for the verified computation of intrinsic functions based on the High Precision Interval datatype are developed and described in detail. The application of these operations in the implementation of High Precision Taylor Models is discussed. An application of Taylor Model methods to the verification of fixed points is presented by verifying the existence of a period 15 fixed point in a near standard Henon map. Verification is performed using different verified methods such as double precision Taylor Models, High Precision intervals and High Precision Taylor Models. Results and performance of each method are compared. An automated rigorous fixed point finder is implemented, allowing the fully automated search for all fixed points of a function within a given domain. It returns a list of verified enclosures of each fixed point, optionally verifying uniqueness within these enclosures. An application of the fixed point finder to the rigorous analysis of beam transfer maps in accelerator physics is presented. Previous work done by
Erikson, U; Misimi, E
2008-03-01
The changes in skin and fillet color of anesthetized and exhausted Atlantic salmon were determined immediately after killing, during rigor mortis, and after ice storage for 7 d. Skin color (CIE L*, a*, b*, and related values) was determined by a Minolta Chroma Meter. Roche SalmoFan Lineal and Roche Color Card values were determined by a computer vision method and a sensory panel. Before color assessment, the stress levels of the 2 fish groups were characterized in terms of white muscle parameters (pH, rigor mortis, and core temperature). The results showed that perimortem handling stress initially significantly affected several color parameters of skin and fillets. Significant transient fillet color changes also occurred in the prerigor phase and during the development of rigor mortis. Our results suggested that fillet color was affected by postmortem glycolysis (pH drop, particularly in anesthetized fillets), then by onset and development of rigor mortis. The color change patterns during storage were different for the 2 groups of fish. The computer vision method was considered suitable for automated (online) quality control and grading of salmonid fillets according to color.
Biomedical text mining for research rigor and integrity: tasks, challenges, directions.
Kilicoglu, Halil
2017-06-13
An estimated quarter of a trillion US dollars is invested in the biomedical research enterprise annually. There is growing alarm that a significant portion of this investment is wasted because of problems in reproducibility of research findings and in the rigor and integrity of research conduct and reporting. Recent years have seen a flurry of activities focusing on standardization and guideline development to enhance the reproducibility and rigor of biomedical research. Research activity is primarily communicated via textual artifacts, ranging from grant applications to journal publications. These artifacts can be both the source and the manifestation of practices leading to research waste. For example, an article may describe a poorly designed experiment, or the authors may reach conclusions not supported by the evidence presented. In this article, we pose the question of whether biomedical text mining techniques can assist the stakeholders in the biomedical research enterprise in doing their part toward enhancing research integrity and rigor. In particular, we identify four key areas in which text mining techniques can make a significant contribution: plagiarism/fraud detection, ensuring adherence to reporting guidelines, managing information overload and accurate citation/enhanced bibliometrics. We review the existing methods and tools for specific tasks, if they exist, or discuss relevant research that can provide guidance for future work. With the exponential increase in biomedical research output and the ability of text mining approaches to perform automatic tasks at large scale, we propose that such approaches can support tools that promote responsible research practices, providing significant benefits for the biomedical research enterprise. Published by Oxford University Press 2017. This work is written by a US Government employee and is in the public domain in the US.
Emergency cricothyrotomy for trismus caused by instantaneous rigor in cardiac arrest patients.
Lee, Jae Hee; Jung, Koo Young
2012-07-01
Instantaneous rigor as muscle stiffening occurring in the moment of death (or cardiac arrest) can be confused with rigor mortis. If trismus is caused by instantaneous rigor, orotracheal intubation is impossible and a surgical airway should be secured. Here, we report 2 patients who had emergency cricothyrotomy for trismus caused by instantaneous rigor. This case report aims to help physicians understand instantaneous rigor and to emphasize the importance of securing a surgical airway quickly on the occurrence of trismus. Copyright © 2012 Elsevier Inc. All rights reserved.
Differential algebras with remainder and rigorous proofs of long-term stability
International Nuclear Information System (INIS)
Berz, Martin
1997-01-01
It is shown how in addition to determining Taylor maps of general optical systems, it is possible to obtain rigorous interval bounds for the remainder term of the n-th order Taylor expansion. To this end, the three elementary operations of addition, multiplication, and differentiation in the Differential Algebraic approach are augmented by suitable interval operations in such a way that a remainder bound of the sum, product, and derivative is obtained from the Taylor polynomial and remainder bound of the operands. The method can be used to obtain bounds for the accuracy with which a Taylor map represents the true map of the particle optical system. In a more general sense, it is also useful for a variety of other numerical problems, including rigorous global optimization of highly complex functions. Combined with methods to obtain pseudo-invariants of repetitive motion and extensions of the Lyapunov- and Nekhoroshev stability theory, the latter can be used to guarantee stability for storage rings and other weakly nonlinear systems
Rigorous Combination of GNSS and VLBI: How it Improves Earth Orientation and Reference Frames
Lambert, S. B.; Richard, J. Y.; Bizouard, C.; Becker, O.
2017-12-01
Current reference series (C04) of the International Earth Rotation and Reference Systems Service (IERS) are produced by a weighted combination of Earth orientation parameters (EOP) time series built up by combination centers of each technique (VLBI, GNSS, Laser ranging, DORIS). In the future, we plan to derive EOP from a rigorous combination of the normal equation systems of the four techniques.We present here the results of a rigorous combination of VLBI and GNSS pre-reduced, constraint-free, normal equations with the DYNAMO geodetic analysis software package developed and maintained by the French GRGS (Groupe de Recherche en GeÌodeÌsie Spatiale). The used normal equations are those produced separately by the IVS and IGS combination centers to which we apply our own minimal constraints.We address the usefulness of such a method with respect to the classical, a posteriori, combination method, and we show whether EOP determinations are improved.Especially, we implement external validations of the EOP series based on comparison with geophysical excitation and examination of the covariance matrices. Finally, we address the potential of the technique for the next generation celestial reference frames, which are currently determined by VLBI only.
Rigorous Results for the Distribution of Money on Connected Graphs
Lanchier, Nicolas; Reed, Stephanie
2018-05-01
This paper is concerned with general spatially explicit versions of three stochastic models for the dynamics of money that have been introduced and studied numerically by statistical physicists: the uniform reshuffling model, the immediate exchange model and the model with saving propensity. All three models consist of systems of economical agents that consecutively engage in pairwise monetary transactions. Computer simulations performed in the physics literature suggest that, when the number of agents and the average amount of money per agent are large, the limiting distribution of money as time goes to infinity approaches the exponential distribution for the first model, the gamma distribution with shape parameter two for the second model and a distribution similar but not exactly equal to a gamma distribution whose shape parameter depends on the saving propensity for the third model. The main objective of this paper is to give rigorous proofs of these conjectures and also extend these conjectures to generalizations of the first two models and a variant of the third model that include local rather than global interactions, i.e., instead of choosing the two interacting agents uniformly at random from the system, the agents are located on the vertex set of a general connected graph and can only interact with their neighbors.
Rigorous vector wave propagation for arbitrary flat media
Bos, Steven P.; Haffert, Sebastiaan Y.; Keller, Christoph U.
2017-08-01
Precise modelling of the (off-axis) point spread function (PSF) to identify geometrical and polarization aberrations is important for many optical systems. In order to characterise the PSF of the system in all Stokes parameters, an end-to-end simulation of the system has to be performed in which Maxwell's equations are rigorously solved. We present the first results of a python code that we are developing to perform multiscale end-to-end wave propagation simulations that include all relevant physics. Currently we can handle plane-parallel near- and far-field vector diffraction effects of propagating waves in homogeneous isotropic and anisotropic materials, refraction and reflection of flat parallel surfaces, interference effects in thin films and unpolarized light. We show that the code has a numerical precision on the order of 10-16 for non-absorbing isotropic and anisotropic materials. For absorbing materials the precision is on the order of 10-8. The capabilities of the code are demonstrated by simulating a converging beam reflecting from a flat aluminium mirror at normal incidence.
Dynamics of harmonically-confined systems: Some rigorous results
Energy Technology Data Exchange (ETDEWEB)
Wu, Zhigang, E-mail: zwu@physics.queensu.ca; Zaremba, Eugene, E-mail: zaremba@sparky.phy.queensu.ca
2014-03-15
In this paper we consider the dynamics of harmonically-confined atomic gases. We present various general results which are independent of particle statistics, interatomic interactions and dimensionality. Of particular interest is the response of the system to external perturbations which can be either static or dynamic in nature. We prove an extended Harmonic Potential Theorem which is useful in determining the damping of the centre of mass motion when the system is prepared initially in a highly nonequilibrium state. We also study the response of the gas to a dynamic external potential whose position is made to oscillate sinusoidally in a given direction. We show in this case that either the energy absorption rate or the centre of mass dynamics can serve as a probe of the optical conductivity of the system. -- Highlights: •We derive various rigorous results on the dynamics of harmonically-confined atomic gases. •We derive an extension of the Harmonic Potential Theorem. •We demonstrate the link between the energy absorption rate in a harmonically-confined system and the optical conductivity.
Rigorous derivation of porous-media phase-field equations
Schmuck, Markus; Kalliadasis, Serafim
2017-11-01
The evolution of interfaces in Complex heterogeneous Multiphase Systems (CheMSs) plays a fundamental role in a wide range of scientific fields such as thermodynamic modelling of phase transitions, materials science, or as a computational tool for interfacial flow studies or material design. Here, we focus on phase-field equations in CheMSs such as porous media. To the best of our knowledge, we present the first rigorous derivation of error estimates for fourth order, upscaled, and nonlinear evolution equations. For CheMs with heterogeneity ɛ, we obtain the convergence rate ɛ 1 / 4 , which governs the error between the solution of the new upscaled formulation and the solution of the microscopic phase-field problem. This error behaviour has recently been validated computationally in. Due to the wide range of application of phase-field equations, we expect this upscaled formulation to allow for new modelling, analytic, and computational perspectives for interfacial transport and phase transformations in CheMSs. This work was supported by EPSRC, UK, through Grant Nos. EP/H034587/1, EP/L027186/1, EP/L025159/1, EP/L020564/1, EP/K008595/1, and EP/P011713/1 and from ERC via Advanced Grant No. 247031.
Spectral radiative property control method based on filling solution
International Nuclear Information System (INIS)
Jiao, Y.; Liu, L.H.; Hsu, P.-F.
2014-01-01
Controlling thermal radiation by tailoring spectral properties of microstructure is a promising method, can be applied in many industrial systems and have been widely researched recently. Among various property tailoring schemes, geometry design of microstructures is a commonly used method. However, the existing radiation property tailoring is limited by adjustability of processed microstructures. In other words, the spectral radiative properties of microscale structures are not possible to change after the gratings are fabricated. In this paper, we propose a method that adjusts the grating spectral properties by means of injecting filling solution, which could modify the thermal radiation in a fabricated microstructure. Therefore, this method overcomes the limitation mentioned above. Both mercury and water are adopted as the filling solution in this study. Aluminum and silver are selected as the grating materials to investigate the generality and limitation of this control method. The rigorous coupled-wave analysis is used to investigate the spectral radiative properties of these filling solution grating structures. A magnetic polaritons mechanism identification method is proposed based on LC circuit model principle. It is found that this control method could be used by different grating materials. Different filling solutions would enable the high absorption peak to move to longer or shorter wavelength band. The results show that the filling solution grating structures are promising for active control of spectral radiative properties. -- Highlights: • A filling solution grating structure is designed to adjust spectral radiative properties. • The mechanism of radiative property control is studied for engineering utilization. • Different grating materials are studied to find multi-functions for grating
International Nuclear Information System (INIS)
Ahmadi, A.; Meyer, M.; Rouzineau, D.; Prevost, M.; Alix, P.; Laloue, N.
2010-01-01
This paper gives the first step of the development of a rigorous multicomponent reactive separation model. Such a model is highly essential to further the optimization of acid gases removal plants (CO 2 capture, gas treating, etc.) in terms of size and energy consumption, since chemical solvents are conventionally used. Firstly, two main modelling approaches are presented: the equilibrium-based and the rate-based approaches. Secondly, an extended rate-based model with rigorous modelling methodology for diffusion-reaction phenomena is proposed. The film theory and the generalized Maxwell-Stefan equations are used in order to characterize multicomponent interactions. The complete chain of chemical reactions is taken into account. The reactions can be kinetically controlled or at chemical equilibrium, and they are considered for both liquid film and liquid bulk. Thirdly, the method of numerical resolution is described. Coupling the generalized Maxwell-Stefan equations with chemical equilibrium equations leads to a highly non-linear Differential-Algebraic Equations system known as DAE index 3. The set of equations is discretized with finite-differences as its integration by Gear method is complex. The resulting algebraic system is resolved by the Newton- Raphson method. Finally, the present model and the associated methods of numerical resolution are validated for the example of esterification of methanol. This archetype non-electrolytic system permits an interesting analysis of reaction impact on mass transfer, especially near the phase interface. The numerical resolution of the model by Newton-Raphson method gives good results in terms of calculation time and convergence. The simulations show that the impact of reactions at chemical equilibrium and that of kinetically controlled reactions with high kinetics on mass transfer is relatively similar. Moreover, the Fick's law is less adapted for multicomponent mixtures where some abnormalities such as counter
Nugraheni, Z.; Budiyono, B.; Slamet, I.
2018-03-01
To reach higher order thinking skill, needed to be mastered the conceptual understanding and strategic competence as they are two basic parts of high order thinking skill (HOTS). RMT is a unique realization of the cognitive conceptual construction approach based on Feurstein with his theory of Mediated Learning Experience (MLE) and Vygotsky’s sociocultural theory. This was quasi-experimental research which compared the experimental class that was given Rigorous Mathematical Thinking (RMT) as learning method and the control class that was given Direct Learning (DL) as the conventional learning activity. This study examined whether there was different effect of two learning model toward conceptual understanding and strategic competence of Junior High School Students. The data was analyzed by using Multivariate Analysis of Variance (MANOVA) and obtained a significant difference between experimental and control class when considered jointly on the mathematics conceptual understanding and strategic competence (shown by Wilk’s Λ = 0.84). Further, by independent t-test is known that there was significant difference between two classes both on mathematical conceptual understanding and strategic competence. By this result is known that Rigorous Mathematical Thinking (RMT) had positive impact toward Mathematics conceptual understanding and strategic competence.
Investigating the performance of directional boundary layer model through staged modeling method
Jeong, Moon-Gyu; Lee, Won-Chan; Yang, Seung-Hune; Jang, Sung-Hoon; Shim, Seong-Bo; Kim, Young-Chang; Suh, Chun-Suk; Choi, Seong-Woon; Kim, Young-Hee
2011-04-01
Generally speaking, the models used in the optical proximity effect correction (OPC) can be divided into three parts, mask part, optic part, and resist part. For the excellent quality of the OPC model, each part has to be described by the first principles. However, OPC model can't take the all of the principles since it should cover the full chip level calculation during the correction. Moreover, the calculation has to be done iteratively during the correction until the cost function we want to minimize converges. Normally the optic part in OPC model is described with the sum of coherent system (SOCS[1]) method. Thanks to this method we can calculate the aerial image so fast without the significant loss of accuracy. As for the resist part, the first principle is too complex to implement in detail, so it is normally expressed in a simple way, such as the approximation of the first principles, and the linear combinations of factors which is highly correlated with the chemistries in the resist. The quality of this kind of the resist model depends on how well we train the model through fitting to the empirical data. The most popular way of making the mask function is based on the Kirchhoff's thin mask approximation. This method works well when the feature size on the mask is sufficiently large, but as the line width of the semiconductor circuit becomes smaller, this method causes significant error due to the mask topography effect. To consider the mask topography effect accurately, we have to use rigorous methods of calculating the mask function, such as finite difference time domain (FDTD[2]) and rigorous coupled-wave analysis (RCWA[3]). But these methods are too time-consuming to be used as a part of the OPC model. Until now many alternatives have been suggested as the efficient way of considering the mask topography effect. Among them we focused on the boundary layer model (BLM) in this paper. We mainly investigated the way of optimization of the parameters for the
Arriaza, Pablo; Nedjat-Haiem, Frances; Lee, Hee Yun; Martin, Shadi S
2015-01-01
The purpose of this article is to synthesize and chronicle the authors' experiences as four bilingual and bicultural researchers, each experienced in conducting cross-cultural/cross-language qualitative research. Through narrative descriptions of experiences with Latinos, Iranians, and Hmong refugees, the authors discuss their rewards, challenges, and methods of enhancing rigor, trustworthiness, and transparency when conducting cross-cultural/cross-language research. The authors discuss and explore how to effectively manage cross-cultural qualitative data, how to effectively use interpreters and translators, how to identify best methods of transcribing data, and the role of creating strong community relationships. The authors provide guidelines for health care professionals to consider when engaging in cross-cultural qualitative research.
A rigorous phenomenological analysis of the ππ scattering lengths
International Nuclear Information System (INIS)
Caprini, I.; Dita, P.; Sararu, M.
1979-11-01
The constraining power of the present experimental data, combined with the general theoretical knowledge about ππ scattering, upon the scattering lengths of this process, is investigated by means of a rigorous functional method. We take as input the experimental phase shifts and make no hypotheses about the high energy behaviour of the amplitudes, using only absolute bounds derived from axiomatic field theory and exact consequences of crossing symmetry. In the simplest application of the method, involving only the π 0 π 0 S-wave, we explored numerically a number of values proposed by various authors for the scattering lengths a 0 and a 2 and found that no one appears to be especially favoured. (author)
Stochastic Geometry and Quantum Gravity: Some Rigorous Results
Zessin, H.
The aim of these lectures is a short introduction into some recent developments in stochastic geometry which have one of its origins in simplicial gravity theory (see Regge Nuovo Cimento 19: 558-571, 1961). The aim is to define and construct rigorously point processes on spaces of Euclidean simplices in such a way that the configurations of these simplices are simplicial complexes. The main interest then is concentrated on their curvature properties. We illustrate certain basic ideas from a mathematical point of view. An excellent representation of this area can be found in Schneider and Weil (Stochastic and Integral Geometry, Springer, Berlin, 2008. German edition: Stochastische Geometrie, Teubner, 2000). In Ambjørn et al. (Quantum Geometry Cambridge University Press, Cambridge, 1997) you find a beautiful account from the physical point of view. More recent developments in this direction can be found in Ambjørn et al. ("Quantum gravity as sum over spacetimes", Lect. Notes Phys. 807. Springer, Heidelberg, 2010). After an informal axiomatic introduction into the conceptual foundations of Regge's approach the first lecture recalls the concepts and notations used. It presents the fundamental zero-infinity law of stochastic geometry and the construction of cluster processes based on it. The second lecture presents the main mathematical object, i.e. Poisson-Delaunay surfaces possessing an intrinsic random metric structure. The third and fourth lectures discuss their ergodic behaviour and present the two-dimensional Regge model of pure simplicial quantum gravity. We terminate with the formulation of basic open problems. Proofs are given in detail only in a few cases. In general the main ideas are developed. Sufficiently complete references are given.
RIGOROUS GEOREFERENCING OF ALSAT-2A PANCHROMATIC AND MULTISPECTRAL IMAGERY
Directory of Open Access Journals (Sweden)
I. Boukerch
2013-04-01
Full Text Available The exploitation of the full geometric capabilities of the High-Resolution Satellite Imagery (HRSI, require the development of an appropriate sensor orientation model. Several authors studied this problem; generally we have two categories of geometric models: physical and empirical models. Based on the analysis of the metadata provided with ALSAT-2A, a rigorous pushbroom camera model can be developed. This model has been successfully applied to many very high resolution imagery systems. The relation between the image and ground coordinates by the time dependant collinearity involving many coordinates systems has been tested. The interior orientation parameters must be integrated in the model, the interior parameters can be estimated from the viewing angles corresponding to the pointing directions of any detector, these values are derived from cubic polynomials provided in the metadata. The developed model integrates all the necessary elements with 33 unknown. All the approximate values of the 33 unknowns parameters may be derived from the informations contained in the metadata files provided with the imagery technical specifications or they are simply fixed to zero, so the condition equation is linearized and solved using SVD in a least square sense in order to correct the initial values using a suitable number of well-distributed GCPs. Using Alsat-2A images over the town of Toulouse in the south west of France, three experiments are done. The first is about 2D accuracy analysis using several sets of parameters. The second is about GCPs number and distribution. The third experiment is about georeferencing multispectral image by applying the model calculated from panchromatic image.
A rigorous derivation of gravitational self-force
International Nuclear Information System (INIS)
Gralla, Samuel E; Wald, Robert M
2008-01-01
There is general agreement that the MiSaTaQuWa equations should describe the motion of a 'small body' in general relativity, taking into account the leading order self-force effects. However, previous derivations of these equations have made a number of ad hoc assumptions and/or contain a number of unsatisfactory features. For example, all previous derivations have invoked, without proper justification, the step of 'Lorenz gauge relaxation', wherein the linearized Einstein equation is written in the form appropriate to the Lorenz gauge, but the Lorenz gauge condition is then not imposed-thereby making the resulting equations for the metric perturbation inequivalent to the linearized Einstein equations. (Such a 'relaxation' of the linearized Einstein equations is essential in order to avoid the conclusion that 'point particles' move on geodesics.) In this paper, we analyze the issue of 'particle motion' in general relativity in a systematic and rigorous way by considering a one-parameter family of metrics, g ab (λ), corresponding to having a body (or black hole) that is 'scaled down' to zero size and mass in an appropriate manner. We prove that the limiting worldline of such a one-parameter family must be a geodesic of the background metric, g ab (λ = 0). Gravitational self-force-as well as the force due to coupling of the spin of the body to curvature-then arises as a first-order perturbative correction in λ to this worldline. No assumptions are made in our analysis apart from the smoothness and limit properties of the one-parameter family of metrics, g ab (λ). Our approach should provide a framework for systematically calculating higher order corrections to gravitational self-force, including higher multipole effects, although we do not attempt to go beyond first-order calculations here. The status of the MiSaTaQuWa equations is explained
Rigorous covariance propagation of geoid errors to geodetic MDT estimates
Pail, R.; Albertella, A.; Fecher, T.; Savcenko, R.
2012-04-01
The mean dynamic topography (MDT) is defined as the difference between the mean sea surface (MSS) derived from satellite altimetry, averaged over several years, and the static geoid. Assuming geostrophic conditions, from the MDT the ocean surface velocities as important component of global ocean circulation can be derived from it. Due to the availability of GOCE gravity field models, for the very first time MDT can now be derived solely from satellite observations (altimetry and gravity) down to spatial length-scales of 100 km and even below. Global gravity field models, parameterized in terms of spherical harmonic coefficients, are complemented by the full variance-covariance matrix (VCM). Therefore, for the geoid component a realistic statistical error estimate is available, while the error description of the altimetric component is still an open issue and is, if at all, attacked empirically. In this study we make the attempt to perform, based on the full gravity VCM, rigorous error propagation to derived geostrophic surface velocities, thus also considering all correlations. For the definition of the static geoid we use the third release of the time-wise GOCE model, as well as the satellite-only combination model GOCO03S. In detail, we will investigate the velocity errors resulting from the geoid component in dependence of the harmonic degree, and the impact of using/no using covariances on the MDT errors and its correlations. When deriving an MDT, it is spectrally filtered to a certain maximum degree, which is usually driven by the signal content of the geoid model, by applying isotropic or non-isotropic filters. Since this filtering is acting also on the geoid component, the consistent integration of this filter process into the covariance propagation shall be performed, and its impact shall be quantified. The study will be performed for MDT estimates in specific test areas of particular oceanographic interest.
Directory of Open Access Journals (Sweden)
Kelly Cristina Mucio Marques
2015-04-01
Full Text Available This study aims to assess the methodological rigor of case studies in management accounting published in Brazilian journals. The study is descriptive. The data were collected using documentary research and content analysis, and 180 papers published from 2008 to 2012 in accounting journals rated as A2, B1, and B2 that were classified as case studies were selected. Based on the literature, we established a set of 15 criteria that we expected to be identified (either explicitly or implicitly in the case studies to classify those case studies as appropriate from the standpoint of methodological rigor. These criteria were partially met by the papers analyzed. The aspects less aligned with those proposed in the literature were the following: little emphasis on justifying the need to understand phenomena in context; lack of explanation of the reason for choosing the case study strategy; the predominant use of questions that do not enable deeper analysis; many studies based on only one source of evidence; little use of data and information triangulation; little emphasis on the data collection method; a high number of cases in which confusion between case study as a research strategy and as data collection method were detected; a low number of papers reporting the method of data analysis; few reports on a study's contributions; and a minority highlighting the issues requiring further research. In conclusion, the method used to apply case studies to management accounting must be improved because few studies showed rigorous application of the procedures that this strategy requires.
How Individual Scholars Can Reduce the Rigor-Relevance Gap in Management Research
Wolf, Joachim; Rosenberg, Timo
2012-01-01
This paper discusses a number of avenues management scholars could follow to reduce the existing gap between scientific rigor and practical relevance without relativizing the importance of the first goal dimension. Such changes are necessary because many management studies do not fully exploit the possibilities to increase their practical relevance while maintaining scientific rigor. We argue that this rigor-relevance gap is not only the consequence of the currently prevailing institutional c...
RIGOR MORTIS AND THE INFLUENCE OF CALCIUM AND MAGNESIUM SALTS UPON ITS DEVELOPMENT.
Meltzer, S J; Auer, J
1908-01-01
Calcium salts hasten and magnesium salts retard the development of rigor mortis, that is, when these salts are administered subcutaneously or intravenously. When injected intra-arterially, concentrated solutions of both kinds of salts cause nearly an immediate onset of a strong stiffness of the muscles which is apparently a contraction, brought on by a stimulation caused by these salts and due to osmosis. This contraction, if strong, passes over without a relaxation into a real rigor. This form of rigor may be classed as work-rigor (Arbeitsstarre). In animals, at least in frogs, with intact cords, the early contraction and the following rigor are stronger than in animals with destroyed cord. If M/8 solutions-nearly equimolecular to "physiological" solutions of sodium chloride-are used, even when injected intra-arterially, calcium salts hasten and magnesium salts retard the onset of rigor. The hastening and retardation in this case as well as in the cases of subcutaneous and intravenous injections, are ion effects and essentially due to the cations, calcium and magnesium. In the rigor hastened by calcium the effects of the extensor muscles mostly prevail; in the rigor following magnesium injection, on the other hand, either the flexor muscles prevail or the muscles become stiff in the original position of the animal at death. There seems to be no difference in the degree of stiffness in the final rigor, only the onset and development of the rigor is hastened in the case of the one salt and retarded in the other. Calcium hastens also the development of heat rigor. No positive facts were obtained with regard to the effect of magnesium upon heat vigor. Calcium also hastens and magnesium retards the onset of rigor in the left ventricle of the heart. No definite data were gathered with regard to the effects of these salts upon the right ventricle.
Rigorous Numerics for ill-posed PDEs: Periodic Orbits in the Boussinesq Equation
Castelli, Roberto; Gameiro, Marcio; Lessard, Jean-Philippe
2018-04-01
In this paper, we develop computer-assisted techniques for the analysis of periodic orbits of ill-posed partial differential equations. As a case study, our proposed method is applied to the Boussinesq equation, which has been investigated extensively because of its role in the theory of shallow water waves. The idea is to use the symmetry of the solutions and a Newton-Kantorovich type argument (the radii polynomial approach) to obtain rigorous proofs of existence of the periodic orbits in a weighted ℓ1 Banach space of space-time Fourier coefficients with exponential decay. We present several computer-assisted proofs of the existence of periodic orbits at different parameter values.
Guilak, Farshid
2017-03-21
We are currently in one of the most exciting times for science and engineering as we witness unprecedented growth in our computational and experimental capabilities to generate new data and models. To facilitate data and model sharing, and to enhance reproducibility and rigor in biomechanics research, the Journal of Biomechanics has introduced a number of tools for Content Innovation to allow presentation, sharing, and archiving of methods, models, and data in our articles. The tools include an Interactive Plot Viewer, 3D Geometric Shape and Model Viewer, Virtual Microscope, Interactive MATLAB Figure Viewer, and Audioslides. Authors are highly encouraged to make use of these in upcoming journal submissions. Copyright © 2017 Elsevier Ltd. All rights reserved.
Rigorous upper bounds for transport due to passive advection by inhomogeneous turbulence
International Nuclear Information System (INIS)
Krommes, J.A.; Smith, R.A.
1987-05-01
A variational procedure, due originally to Howard and explored by Busse and others for self-consistent turbulence problems, is employed to determine rigorous upper bounds for the advection of a passive scalar through an inhomogeneous turbulent slab with arbitrary generalized Reynolds number R and Kubo number K. In the basic version of the method, the steady-state energy balance is used as a constraint; the resulting bound, though rigorous, is independent of K. A pedagogical reference model (one dimension, K = ∞) is described in detail; the bound compares favorably with the exact solution. The direct-interaction approximation is also worked out for this model; it is somewhat more accurate than the bound, but requires considerably more labor to solve. For the basic bound, a general formalism is presented for several dimensions, finite correlation length, and reasonably general boundary conditions. Part of the general method, in which a Green's function technique is employed, applies to self-consistent as well as to passive problems, and thereby generalizes previous results in the fluid literature. The formalism is extended for the first time to include time-dependent constraints, and a bound is deduced which explicitly depends on K and has the correct physical scalings in all regimes of R and K. Two applications from the theory of turbulent plasmas ae described: flux in velocity space, and test particle transport in stochastic magnetic fields. For the velocity space problem the simplest bound reproduces Dupree's original scaling for the strong turbulence diffusion coefficient. For the case of stochastic magnetic fields, the scaling of the bounds is described for the magnetic diffusion coefficient as well as for the particle diffusion coefficient in the so-called collisionless, fluid, and double-streaming regimes
Dosimetric effects of edema in permanent prostate seed implants: a rigorous solution
International Nuclear Information System (INIS)
Chen Zhe; Yue Ning; Wang Xiaohong; Roberts, Kenneth B.; Peschel, Richard; Nath, Ravinder
2000-01-01
Purpose: To derive a rigorous analytic solution to the dosimetric effects of prostate edema so that its impact on the conventional pre-implant and post-implant dosimetry can be studied for any given radioactive isotope and edema characteristics. Methods and Materials: The edema characteristics observed by Waterman et al (Int. J. Rad. Onc. Biol. Phys, 41:1069-1077; 1998) was used to model the time evolution of the prostate and the seed locations. The total dose to any part of prostate tissue from a seed implant was calculated analytically by parameterizing the dose fall-off from a radioactive seed as a single inverse power function of distance, with proper account of the edema-induced time evolution. The dosimetric impact of prostate edema was determined by comparing the dose calculated with full consideration of prostate edema to that calculated with the conventional dosimetry approach where the seed locations and the target volume are assumed to be stationary. Results: A rigorous analytic solution on the relative dosimetric effects of prostate edema was obtained. This solution proved explicitly that the relative dosimetric effects of edema, as found in the previous numerical studies by Yue et. al. (Int. J. Radiat. Oncol. Biol. Phys. 43, 447-454, 1999), are independent of the size and the shape of the implant target volume and are independent of the number and the locations of the seeds implanted. It also showed that the magnitude of relative dosimetric effects is independent of the location of dose evaluation point within the edematous target volume. It implies that the relative dosimetric effects of prostate edema are universal with respect to a given isotope and edema characteristic. A set of master tables for the relative dosimetric effects of edema were obtained for a wide range of edema characteristics for both 125 I and 103 Pd prostate seed implants. Conclusions: A rigorous analytic solution of the relative dosimetric effects of prostate edema has been
Trends in Methodological Rigor in Intervention Research Published in School Psychology Journals
Burns, Matthew K.; Klingbeil, David A.; Ysseldyke, James E.; Petersen-Brown, Shawna
2012-01-01
Methodological rigor in intervention research is important for documenting evidence-based practices and has been a recent focus in legislation, including the No Child Left Behind Act. The current study examined the methodological rigor of intervention research in four school psychology journals since the 1960s. Intervention research has increased…
Kim, Hyun-Wook; Hwang, Ko-Eun; Song, Dong-Heon; Kim, Yong-Jae; Ham, Youn-Kyung; Yeo, Eui-Joo; Jeong, Tae-Jun; Choi, Yun-Sang; Kim, Cheon-Jei
2015-01-01
This study was conducted to evaluate the effect of pre-rigor salting level (0-4% NaCl concentration) on physicochemical and textural properties of pre-rigor chicken breast muscles. The pre-rigor chicken breast muscles were de-boned 10 min post-mortem and salted within 25 min post-mortem. An increase in pre-rigor salting level led to the formation of high ultimate pH of chicken breast muscles at post-mortem 24 h. The addition of minimum of 2% NaCl significantly improved water holding capacity, cooking loss, protein solubility, and hardness when compared to the non-salting chicken breast muscle (prigor salting level caused the inhibition of myofibrillar protein degradation and the acceleration of lipid oxidation. However, the difference in NaCl concentration between 3% and 4% had no great differences in the results of physicochemical and textural properties due to pre-rigor salting effects (p>0.05). Therefore, our study certified the pre-rigor salting effect of chicken breast muscle salted with 2% NaCl when compared to post-rigor muscle salted with equal NaCl concentration, and suggests that the 2% NaCl concentration is minimally required to ensure the definite pre-rigor salting effect on chicken breast muscle.
Choi, Yun-Sang
2015-01-01
This study was conducted to evaluate the effect of pre-rigor salting level (0-4% NaCl concentration) on physicochemical and textural properties of pre-rigor chicken breast muscles. The pre-rigor chicken breast muscles were de-boned 10 min post-mortem and salted within 25 min post-mortem. An increase in pre-rigor salting level led to the formation of high ultimate pH of chicken breast muscles at post-mortem 24 h. The addition of minimum of 2% NaCl significantly improved water holding capacity, cooking loss, protein solubility, and hardness when compared to the non-salting chicken breast muscle (psalting level caused the inhibition of myofibrillar protein degradation and the acceleration of lipid oxidation. However, the difference in NaCl concentration between 3% and 4% had no great differences in the results of physicochemical and textural properties due to pre-rigor salting effects (p>0.05). Therefore, our study certified the pre-rigor salting effect of chicken breast muscle salted with 2% NaCl when compared to post-rigor muscle salted with equal NaCl concentration, and suggests that the 2% NaCl concentration is minimally required to ensure the definite pre-rigor salting effect on chicken breast muscle. PMID:26761884
Rigorous bounds on the free energy of electron-phonon models
Raedt, Hans De; Michielsen, Kristel
1997-01-01
We present a collection of rigorous upper and lower bounds to the free energy of electron-phonon models with linear electron-phonon interaction. These bounds are used to compare different variational approaches. It is shown rigorously that the ground states corresponding to the sharpest bounds do
The Relationship between Project-Based Learning and Rigor in STEM-Focused High Schools
Edmunds, Julie; Arshavsky, Nina; Glennie, Elizabeth; Charles, Karen; Rice, Olivia
2016-01-01
Project-based learning (PjBL) is an approach often favored in STEM classrooms, yet some studies have shown that teachers struggle to implement it with academic rigor. This paper explores the relationship between PjBL and rigor in the classrooms of ten STEM-oriented high schools. Utilizing three different data sources reflecting three different…
Moving beyond Data Transcription: Rigor as Issue in Representation of Digital Literacies
Hagood, Margaret Carmody; Skinner, Emily Neil
2015-01-01
Rigor in qualitative research has been based upon criteria of credibility, dependability, confirmability, and transferability. Drawing upon articles published during our editorship of the "Journal of Adolescent & Adult Literacy," we illustrate how the use of digital data in research study reporting may enhance these areas of rigor,…
Bringing scientific rigor to community-developed programs in Hong Kong
Directory of Open Access Journals (Sweden)
Fabrizio Cecilia S
2012-12-01
Full Text Available Abstract Background This paper describes efforts to generate evidence for community-developed programs to enhance family relationships in the Chinese culture of Hong Kong, within the framework of community-based participatory research (CBPR. Methods The CBPR framework was applied to help maximize the development of the intervention and the public health impact of the studies, while enhancing the capabilities of the social service sector partners. Results Four academic-community research teams explored the process of designing and implementing randomized controlled trials in the community. In addition to the expected cultural barriers between teams of academics and community practitioners, with their different outlooks, concerns and languages, the team navigated issues in utilizing the principles of CBPR unique to this Chinese culture. Eventually the team developed tools for adaptation, such as an emphasis on building the relationship while respecting role delineation and an iterative process of defining the non-negotiable parameters of research design while maintaining scientific rigor. Lessons learned include the risk of underemphasizing the size of the operational and skills shift between usual agency practices and research studies, the importance of minimizing non-negotiable parameters in implementing rigorous research designs in the community, and the need to view community capacity enhancement as a long term process. Conclusions The four pilot studies under the FAMILY Project demonstrated that nuanced design adaptations, such as wait list controls and shorter assessments, better served the needs of the community and led to the successful development and vigorous evaluation of a series of preventive, family-oriented interventions in the Chinese culture of Hong Kong.
Directory of Open Access Journals (Sweden)
Tao Liu
2017-08-01
Full Text Available Abstract Background Dengue fever is a severe public heath challenge in south China. A dengue outbreak was reported in Chaozhou city, China in 2015. Intensified interventions were implemented by the government to control the epidemic. However, it is still unknown the degree to which intensified control measures reduced the size of the epidemics, and when should such measures be initiated to reduce the risk of large dengue outbreaks developing? Methods We selected Xiangqiao district as study setting because the majority of the indigenous cases (90.6% in Chaozhou city were from this district. The numbers of daily indigenous dengue cases in 2015 were collected through the national infectious diseases and vectors surveillance system, and daily Breteau Index (BI data were reported by local public health department. We used a compartmental dynamic SEIR (Susceptible, Exposed, Infected and Removed model to assess the effectiveness of control interventions, and evaluate the control effect of intervention timing on dengue epidemic. Results A total of 1250 indigenous dengue cases was reported from Xiangqiao district. The results of SEIR modeling using BI as an indicator of actual control interventions showed a total of 1255 dengue cases, which is close to the reported number (n = 1250. The size and duration of the outbreak were highly sensitive to the intensity and timing of interventions. The more rigorous and earlier the control interventions implemented, the more effective it yielded. Even if the interventions were initiated several weeks after the onset of the dengue outbreak, the interventions were shown to greatly impact the prevalence and duration of dengue outbreak. Conclusions This study suggests that early implementation of rigorous dengue interventions can effectively reduce the epidemic size and shorten the epidemic duration.
Onset of rigor mortis is earlier in red muscle than in white muscle.
Kobayashi, M; Takatori, T; Nakajima, M; Sakurada, K; Hatanaka, K; Ikegaya, H; Matsuda, Y; Iwase, H
2000-01-01
Rigor mortis is thought to be related to falling ATP levels in muscles postmortem. We measured rigor mortis as tension determined isometrically in three rat leg muscles in liquid paraffin kept at 37 degrees C or 25 degrees C--two red muscles, red gastrocnemius (RG) and soleus (SO) and one white muscle, white gastrocnemius (WG). Onset, half and full rigor mortis occurred earlier in RG and SO than in WG both at 37 degrees C and at 25 degrees C even though RG and WG were portions of the same muscle. This suggests that rigor mortis directly reflects the postmortem intramuscular ATP level, which decreases more rapidly in red muscle than in white muscle after death. Rigor mortis was more retarded at 25 degrees C than at 37 degrees C in each type of muscle.
High and low rigor temperature effects on sheep meat tenderness and ageing.
Devine, Carrick E; Payne, Steven R; Peachey, Bridget M; Lowe, Timothy E; Ingram, John R; Cook, Christian J
2002-02-01
Immediately after electrical stimulation, the paired m. longissimus thoracis et lumborum (LT) of 40 sheep were boned out and wrapped tightly with a polyethylene cling film. One of the paired LT's was chilled in 15°C air to reach a rigor mortis (rigor) temperature of 18°C and the other side was placed in a water bath at 35°C and achieved rigor at this temperature. Wrapping reduced rigor shortening and mimicked meat left on the carcass. After rigor, the meat was aged at 15°C for 0, 8, 26 and 72 h and then frozen. The frozen meat was cooked to 75°C in an 85°C water bath and shear force values obtained from a 1×1 cm cross-section. The shear force values of meat for 18 and 35°C rigor were similar at zero ageing, but as ageing progressed, the 18 rigor meat aged faster and became more tender than meat that went into rigor at 35°C (Prigor at each ageing time were significantly different (Prigor were still significantly greater. Thus the toughness of 35°C meat was not a consequence of muscle shortening and appears to be due to both a faster rate of tenderisation and the meat tenderising to a greater extent at the lower temperature. The cook loss at 35°C rigor (30.5%) was greater than that at 18°C rigor (28.4%) (P<0.01) and the colour Hunter L values were higher at 35°C (P<0.01) compared with 18°C, but there were no significant differences in a or b values.
International Nuclear Information System (INIS)
Chen, Xin
2014-01-01
Understanding the roles of the temporary and spatial structures of quantum functional noise in open multilevel quantum molecular systems attracts a lot of theoretical interests. I want to establish a rigorous and general framework for functional quantum noises from the constructive and computational perspectives, i.e., how to generate the random trajectories to reproduce the kernel and path ordering of the influence functional with effective Monte Carlo methods for arbitrary spectral densities. This construction approach aims to unify the existing stochastic models to rigorously describe the temporary and spatial structure of Gaussian quantum noises. In this paper, I review the Euclidean imaginary time influence functional and propose the stochastic matrix multiplication scheme to calculate reduced equilibrium density matrices (REDM). In addition, I review and discuss the Feynman-Vernon influence functional according to the Gaussian quadratic integral, particularly its imaginary part which is critical to the rigorous description of the quantum detailed balance. As a result, I establish the conditions under which the influence functional can be interpreted as the average of exponential functional operator over real-valued Gaussian processes for open multilevel quantum systems. I also show the difference between the local and nonlocal phonons within this framework. With the stochastic matrix multiplication scheme, I compare the normalized REDM with the Boltzmann equilibrium distribution for open multilevel quantum systems
Computation of Quasiperiodic Normally Hyperbolic Invariant Tori: Rigorous Results
Canadell, Marta; Haro, Àlex
2017-12-01
The development of efficient methods for detecting quasiperiodic oscillations and computing the corresponding invariant tori is a subject of great importance in dynamical systems and their applications in science and engineering. In this paper, we prove the convergence of a new Newton-like method for computing quasiperiodic normally hyperbolic invariant tori carrying quasiperiodic motion in smooth families of real-analytic dynamical systems. The main result is stated as an a posteriori KAM-like theorem that allows controlling the inner dynamics on the torus with appropriate detuning parameters, in order to obtain a prescribed quasiperiodic motion. The Newton-like method leads to several fast and efficient computational algorithms, which are discussed and tested in a companion paper (Canadell and Haro in J Nonlinear Sci, 2017. doi: 10.1007/s00332-017-9388-z), in which new mechanisms of breakdown are presented.
Rigorous Verification for the Solution of Nonlinear Interval System ...
African Journals Online (AJOL)
We survey a general method for solving nonlinear interval systems of equations. In particular, we paid special attention to the computational aspects of linear interval systems since the bulk of computations are done during the stage of computing outer estimation of the including linear interval systems. The height of our ...
Rigorous determination of stratospheric water vapor trends from MIPAS observations.
Ceccherini, Simone; Carli, Bruno; Raspollini, Piera; Ridolfi, Marco
2011-05-09
The trend of stratospheric water vapor as a function of latitude is estimated by the MIPAS measurements by means of a new method that uses the measurement space solution. The method uses all the information provided by the observations avoiding the artifacts introduced by the a priori information and by the interpolation to different vertical grids. The analysis provides very precise values of the trends that, however, are limited by a relatively large systematic error induced by the radiometric calibration error of the instrument. The results show in the five years from 2005 to 2009 a dependence on latitude of the stratospheric (from 37 to 53 km) water vapor trend with a positive value of (0.41 ± 0.16)%yr-1 in the northern hemisphere and less than 0.16%yr-1 in the southern hemisphere.
PRO development: rigorous qualitative research as the crucial foundation
Lasch, Kathryn Eilene; Marquis, Patrick; Vigneux, Marc; Abetz, Linda; Arnould, Benoit; Bayliss, Martha; Crawford, Bruce; Rosa, Kathleen
2010-01-01
Recently published articles have described criteria to assess qualitative research in the health field in general, but very few articles have delineated qualitative methods to be used in the development of Patient-Reported Outcomes (PROs). In fact, how PROs are developed with subject input through focus groups and interviews has been given relatively short shrift in the PRO literature when compared to the plethora of quantitative articles on the psychometric properties of PROs. If documented ...
Cumulative prospect theory and mean variance analysis. A rigorous comparison
Hens, Thorsten; Mayer, Janos
2012-01-01
We compare asset allocations derived for cumulative prospect theory(CPT) based on two different methods: Maximizing CPT along the mean–variance efficient frontier and maximizing it without that restriction. We find that with normally distributed returns the difference is negligible. However, using standard asset allocation data of pension funds the difference is considerable. Moreover, with derivatives like call options the restriction to the mean-variance efficient frontier results in a siza...
A rigorous semantics for BPMN 2.0 process diagrams
Kossak, Felix; Geist, Verena; Kubovy, Jan; Natschläger, Christine; Ziebermayr, Thomas; Kopetzky, Theodorich; Freudenthaler, Bernhard; Schewe, Klaus-Dieter
2015-01-01
This book provides the most complete formal specification of the semantics of the Business Process Model and Notation 2.0 standard (BPMN) available to date, in a style that is easily understandable for a wide range of readers - not only for experts in formal methods, but e.g. also for developers of modeling tools, software architects, or graduate students specializing in business process management. BPMN - issued by the Object Management Group - is a widely used standard for business process modeling. However, major drawbacks of BPMN include its limited support for organizational modeling, i
A Development of Advanced Rigorous 2 Step System for the High Resolution Residual Dose Evaluation
Energy Technology Data Exchange (ETDEWEB)
Kim, Do Hyun; Kim, Jong Woo; Kim, Jea Hyun; Lee, Jae Yong; Shin, Chang Ho [Hanyang Univ., Seoul (Korea, Republic of); Kim, Song Hyun [Kyoto University, Sennan (Japan)
2016-10-15
In these days, an activation problem such as residual radiation is one of the important issues. The activated devices and structures can emit the residual radiation. Therefore, the activation should be properly analyzed to make a plan for design, operation, and decontamination of nuclear facilities. For activation calculation, Rigorous 2 Step (R2S) method is introduced as following strategy: (1) the particle transport calculation is performed for an object geometry to get particle spectra and total fluxes; (2) inventories of each cell are calculated by using flux information according to irradiation and decay history; (3) the residual gamma distribution was evaluated by transport code, if needed. This scheme is based on cell calculation of used geometry. In this method, the particle spectra and total fluxes are obtained by mesh tally for activation calculation. It is useful to reduce the effects of gradient flux information. Nevertheless, several limitations are known as follows: Firstly, high relative error of spectra, when lots of meshes were used; secondly, different flux information from spectrum of void in mesh-tally. To calculate high resolution residual dose, several method are developed such as R2Smesh and MCR2S unstructured mesh. The R2Smesh method products better efficiency for obtaining neutron spectra by using fine/coarse mesh. Also, the MCR2S unstructured mesh can effectively separate void spectrum. In this study, the AR2S system was developed to combine the features of those mesh based R2S method. To confirm the AR2S system, the simple activation problem was evaluated and compared with R2S method using same division. Those results have good agreement within 0.83 %. Therefore, it is expected that the AR2S system can properly estimate an activation problem.
Rigorous Training of Dogs Leads to High Accuracy in Human Scent Matching-To-Sample Performance.
Directory of Open Access Journals (Sweden)
Sophie Marchal
Full Text Available Human scent identification is based on a matching-to-sample task in which trained dogs are required to compare a scent sample collected from an object found at a crime scene to that of a suspect. Based on dogs' greater olfactory ability to detect and process odours, this method has been used in forensic investigations to identify the odour of a suspect at a crime scene. The excellent reliability and reproducibility of the method largely depend on rigor in dog training. The present study describes the various steps of training that lead to high sensitivity scores, with dogs matching samples with 90% efficiency when the complexity of the scents presented during the task in the sample is similar to that presented in the in lineups, and specificity reaching a ceiling, with no false alarms in human scent matching-to-sample tasks. This high level of accuracy ensures reliable results in judicial human scent identification tests. Also, our data should convince law enforcement authorities to use these results as official forensic evidence when dogs are trained appropriately.
Applying rigorous decision analysis methodology to optimization of a tertiary recovery project
International Nuclear Information System (INIS)
Wackowski, R.K.; Stevens, C.E.; Masoner, L.O.; Attanucci, V.; Larson, J.L.; Aslesen, K.S.
1992-01-01
This paper reports that the intent of this study was to rigorously look at all of the possible expansion, investment, operational, and CO 2 purchase/recompression scenarios (over 2500) to yield a strategy that would maximize net present value of the CO 2 project at the Rangely Weber Sand Unit. Traditional methods of project management, which involve analyzing large numbers of single case economic evaluations, was found to be too cumbersome and inaccurate for an analysis of this scope. The decision analysis methodology utilized a statistical approach which resulted in a range of economic outcomes. Advantages of the decision analysis methodology included: a more organized approach to classification of decisions and uncertainties; a clear sensitivity method to identify the key uncertainties; an application of probabilistic analysis through the decision tree; and a comprehensive display of the range of possible outcomes for communication to decision makers. This range made it possible to consider the upside and downside potential of the options and to weight these against the Unit's strategies. Savings in time and manpower required to complete the study were also realized
Kornhaber, Rachel Anne; de Jong, A E E; McLean, L
2015-12-01
Qualitative methods are progressively being implemented by researchers for exploration within healthcare. However, there has been a longstanding and wide-ranging debate concerning the relative merits of qualitative research within the health care literature. This integrative review aimed to exam the contribution of qualitative research in burns care and subsequent rehabilitation. Studies were identified using an electronic search strategy using the databases PubMed, Cumulative Index of Nursing and Allied Health Literature (CINAHL), Excerpta Medica database (EMBASE) and Scopus of peer reviewed primary research in English between 2009 to April 2014 using Whittemore and Knaﬂ's integrative review method as a guide for analysis. From the 298 papers identified, 26 research papers met the inclusion criteria. Across all studies there was an average of 22 participants involved in each study with a range of 6-53 participants conducted across 12 nations that focussed on burns prevention, paediatric burns, appropriate acquisition and delivery of burns care, pain and psychosocial implications of burns trauma. Careful and rigorous application of qualitative methodologies promotes and enriches the development of burns knowledge. In particular, the key elements in qualitative methodological process and its publication are critical in disseminating credible and methodologically sound qualitative research. Copyright © 2015 Elsevier Ltd and ISBI. All rights reserved.
Nyitrai, M; Hild, G; Lukács, A; Bódis, E; Somogyi, B
2000-01-28
Cyclic conformational changes in the myosin head are considered essential for muscle contraction. We hereby show that the extension of the fluorescence resonance energy transfer method described originally by Taylor et al. (Taylor, D. L., Reidler, J., Spudich, J. A., and Stryer, L. (1981) J. Cell Biol. 89, 362-367) allows determination of the position of a labeled point outside the actin filament in supramolecular complexes and also characterization of the conformational heterogeneity of an actin-binding protein while considering donor-acceptor distance distributions. Using this method we analyzed proximity relationships between two labeled points of S1 and the actin filament in the acto-S1 rigor complex. The donor (N-[[(iodoacetyl)amino]ethyl]-5-naphthylamine-1-sulfonate) was attached to either the catalytic domain (Cys-707) or the essential light chain (Cys-177) of S1, whereas the acceptor (5-(iodoacetamido)fluorescein) was attached to the actin filament (Cys-374). In contrast to the narrow positional distribution (assumed as being Gaussian) of Cys-707 (5 +/- 3 A), the positional distribution of Cys-177 was found to be broad (102 +/- 4 A). Such a broad positional distribution of the label on the essential light chain of S1 may be important in accommodating the helically arranged acto-myosin binding relative to the filament axis.
Kobayashi, Masahiko; Takemori, Shigeru; Yamaguchi, Maki
2004-02-10
Based on the molecular mechanism of rigor mortis, we have proposed that stiffness (elastic modulus evaluated with tension response against minute length perturbations) can be a suitable index of post-mortem rigidity in skeletal muscle. To trace the developmental process of rigor mortis, we measured stiffness and tension in both red and white rat skeletal muscle kept in liquid paraffin at 37 and 25 degrees C. White muscle (in which type IIB fibres predominate) developed stiffness and tension significantly more slowly than red muscle, except for soleus red muscle at 25 degrees C, which showed disproportionately slow rigor development. In each of the examined muscles, stiffness and tension developed more slowly at 25 degrees C than at 37 degrees C. In each specimen, tension always reached its maximum level earlier than stiffness, and then decreased more rapidly and markedly than stiffness. These phenomena may account for the sequential progress of rigor mortis in human cadavers.
2010-05-27
... rigorous knowledge and skills in English- language arts and mathematics that employers and colleges expect... specialists and to access the student outcome data needed to meet annual evaluation and reporting requirements...
2016-04-01
AU/ACSC/2016 AIR COMMAND AND STAFF COLLEGE AIR UNIVERSITY MASTERS OF ANALYTICAL TRADECRAFT: CERTIFYING THE STANDARDS AND ANALYTIC RIGOR OF...establishing unit level certified Masters of Analytic Tradecraft (MAT) analysts to be trained and entrusted to evaluate and rate the standards and...cues) ideally should meet or exceed effective rigor (based on analytical process).4 To accomplish this, decision makers should not be left to their
A Generic Model for Relative Adjustment Between Optical Sensors Using Rigorous Orbit Mechanics
Directory of Open Access Journals (Sweden)
B. Islam
2008-06-01
Full Text Available The classical calibration or space resection is the fundamental task in photogrammetry. The lack of sufficient knowledge of interior and exterior orientation parameters lead to unreliable results in the photogrammetric process. One of the earliest in approaches using in photogrammetry was the plumb line calibration method. This method is suitable to recover the radial and decentering lens distortion coefficients, while the remaining interior(focal length and principal point coordinates and exterior orientation parameters have to be determined by a complimentary method. As the lens distortion remains very less it not considered as the interior orientation parameters, in the present rigorous sensor model. There are several other available methods based on the photogrammetric collinearity equations, which consider the determination of exterior orientation parameters, with no mention to the simultaneous determination of inner orientation parameters. Normal space resection methods solve the problem using control points, whose coordinates are known both in image and object reference systems. The non-linearity of the model and the problems, in point location in digital images and identifying the maximum GPS measured control points are the main drawbacks of the classical approaches. This paper addresses mathematical model based on the fundamental assumption of collineariy of three points of two Along-Track Stereo imagery sensors and independent object point. Assuming this condition it is possible to extract the exterior orientation (EO parameters for a long strip and single image together, without and with using the control points. Moreover, after extracting the EO parameters the accuracy for satellite data products are compared in with using single and with no control points.
Healthcare market research examined. Relevant, rigorous and highly regulated
Directory of Open Access Journals (Sweden)
Bob Douglas
2011-10-01
Full Text Available [The abstract of this article is not available. Here are the first sentences of the article. The full text is freely available upon registration]Market research is invariably confused with marketing – but, in fact, the two disciplines are very different. Put in its simplest terms, marketing is about promotion whilst market research is about understanding. Accordingly, data collected for market research purposes are used in a completely different way to that gathered for marketing, with research practices heavily regulated to ensure high ethical standards.Let’s begin with a definition of what, exactly, market research is. According to the ICC/ESOMAR International Code 2007 (a definition also adopted by the European Pharmaceutical Market Research Association, it is: «the systematic gathering and interpretation of information about individuals or organisations using the statistical and analytical methods and techniques of the applied social sciences to gain insight or support decision-making. The identity of respondents will not be revealed to the user of the information without explicit consent and no sales approach will be made to them as a direct result of their having provided information».
Rigorous results in quantum theory of stimulated Raman scattering
International Nuclear Information System (INIS)
Rupasov, V.I.
1993-01-01
The modern theory of stimulated Raman scattering (SRS) of light in resonant media is based on the investigations of appropriate integrable models of the classical field theory by means of the inverse problem method. But, strictly speaking, Raman scattering is a pure spontaneous process and, hence, it is necessary to take into account a quantum nature of the phenomenon. Moreover, there are some questions and problems, for example, the problem of scattered photons statistics, which can be studied only within the framework of the quantum field theory. We have developed an exact quantum theory of SRS for the case of point-like geometry of resonant media (two-level atoms or harmonic oscillators) of the radius r much-lt λ 0 , where λ 0 is the typical wavelength of the light, but all our results are also valid for the case of short extended medium of the length L much-lt l p (l p is the typical size of pulses) when the spatially homogeneous approximation is valid
Di, K.; Liu, Y.; Liu, B.; Peng, M.
2012-07-01
Chang'E-1(CE-1) and Chang'E-2(CE-2) are the two lunar orbiters of China's lunar exploration program. Topographic mapping using CE-1 and CE-2 images is of great importance for scientific research as well as for preparation of landing and surface operation of Chang'E-3 lunar rover. In this research, we developed rigorous sensor models of CE-1 and CE-2 CCD cameras based on push-broom imaging principle with interior and exterior orientation parameters. Based on the rigorous sensor model, the 3D coordinate of a ground point in lunar body-fixed (LBF) coordinate system can be calculated by space intersection from the image coordinates of con-jugate points in stereo images, and the image coordinates can be calculated from 3D coordinates by back-projection. Due to uncer-tainties of the orbit and the camera, the back-projected image points are different from the measured points. In order to reduce these inconsistencies and improve precision, we proposed two methods to refine the rigorous sensor model: 1) refining EOPs by correcting the attitude angle bias, 2) refining the interior orientation model by calibration of the relative position of the two linear CCD arrays. Experimental results show that the mean back-projection residuals of CE-1 images are reduced to better than 1/100 pixel by method 1 and the mean back-projection residuals of CE-2 images are reduced from over 20 pixels to 0.02 pixel by method 2. Consequently, high precision DEM (Digital Elevation Model) and DOM (Digital Ortho Map) are automatically generated.
Directory of Open Access Journals (Sweden)
K. Di
2012-07-01
Full Text Available Chang'E-1(CE-1 and Chang'E-2(CE-2 are the two lunar orbiters of China's lunar exploration program. Topographic mapping using CE-1 and CE-2 images is of great importance for scientific research as well as for preparation of landing and surface operation of Chang'E-3 lunar rover. In this research, we developed rigorous sensor models of CE-1 and CE-2 CCD cameras based on push-broom imaging principle with interior and exterior orientation parameters. Based on the rigorous sensor model, the 3D coordinate of a ground point in lunar body-fixed (LBF coordinate system can be calculated by space intersection from the image coordinates of con-jugate points in stereo images, and the image coordinates can be calculated from 3D coordinates by back-projection. Due to uncer-tainties of the orbit and the camera, the back-projected image points are different from the measured points. In order to reduce these inconsistencies and improve precision, we proposed two methods to refine the rigorous sensor model: 1 refining EOPs by correcting the attitude angle bias, 2 refining the interior orientation model by calibration of the relative position of the two linear CCD arrays. Experimental results show that the mean back-projection residuals of CE-1 images are reduced to better than 1/100 pixel by method 1 and the mean back-projection residuals of CE-2 images are reduced from over 20 pixels to 0.02 pixel by method 2. Consequently, high precision DEM (Digital Elevation Model and DOM (Digital Ortho Map are automatically generated.
Rigorous classification and carbon accounting principles for low and Zero Carbon Cities
International Nuclear Information System (INIS)
Kennedy, Scott; Sgouridis, Sgouris
2011-01-01
A large number of communities, new developments, and regions aim to lower their carbon footprint and aspire to become 'zero carbon' or 'Carbon Neutral.' Yet there are neither clear definitions for the scope of emissions that such a label would address on an urban scale, nor is there a process for qualifying the carbon reduction claims. This paper addresses the question of how to define a zero carbon, Low Carbon, or Carbon Neutral urban development by proposing hierarchical emissions categories with three levels: Internal Emissions based on the geographical boundary, external emissions directly caused by core municipal activities, and internal or external emissions due to non-core activities. Each level implies a different carbon management strategy (eliminating, balancing, and minimizing, respectively) needed to meet a Net Zero Carbon designation. The trade-offs, implications, and difficulties of implementing carbon debt accounting based upon these definitions are further analyzed. - Highlights: → A gap exists in comprehensive and standardized accounting methods for urban carbon emissions. → We propose a comprehensive and rigorous City Framework for Carbon Accounting (CiFCA). → CiFCA classifies emissions hierarchically with corresponding carbon management strategies. → Adoption of CiFCA allows for meaningful comparisons of claimed performance of eco-cities.
Revisiting the constant growth angle: Estimation and verification via rigorous thermal modeling
Virozub, Alexander; Rasin, Igal G.; Brandon, Simon
2008-12-01
Methods for estimating growth angle ( θgr) values, based on the a posteriori analysis of directionally solidified material (e.g. drops) often involve assumptions of negligible gravitational effects as well as a planar solid/liquid interface during solidification. We relax both of these assumptions when using experimental drop shapes from the literature to estimate the relevant growth angles at the initial stages of solidification. Assumed to be constant, we use these values as input into a rigorous heat transfer and solidification model of the growth process. This model, which is shown to reproduce the experimental shape of a solidified sessile water drop using the literature value of θgr=0∘, yields excellent agreement with experimental profiles using our estimated values for silicon ( θgr=10∘) and germanium ( θgr=14.3∘) solidifying on an isotropic crystalline surface. The effect of gravity on the solidified drop shape is found to be significant in the case of germanium, suggesting that gravity should either be included in the analysis or that care should be taken that the relevant Bond number is truly small enough in each measurement. The planar solidification interface assumption is found to be unjustified. Although this issue is important when simulating the inflection point in the profile of the solidified water drop, there are indications that solidified drop shapes (at least in the case of silicon) may be fairly insensitive to the shape of this interface.
A Rigorous Investigation on the Ground State of the Penson-Kolb Model
Yang, Kai-Hua; Tian, Guang-Shan; Han, Ru-Qi
2003-05-01
By using either numerical calculations or analytical methods, such as the bosonization technique, the ground state of the Penson-Kolb model has been previously studied by several groups. Some physicists argued that, as far as the existence of superconductivity in this model is concerned, it is canonically equivalent to the negative-U Hubbard model. However, others did not agree. In the present paper, we shall investigate this model by an independent and rigorous approach. We show that the ground state of the Penson-Kolb model is nondegenerate and has a nonvanishing overlap with the ground state of the negative-U Hubbard model. Furthermore, we also show that the ground states of both the models have the same good quantum numbers and may have superconducting long-range order at the same momentum q = 0. Our results support the equivalence between these models. The project partially supported by the Special Funds for Major State Basic Research Projects (G20000365) and National Natural Science Foundation of China under Grant No. 10174002
Hamid, H.
2018-01-01
The purpose of this study is to analyze an improvement of students’ mathematical critical thinking (CT) ability in Real Analysis course by using Rigorous Teaching and Learning (RTL) model with informal argument. In addition, this research also attempted to understand students’ CT on their initial mathematical ability (IMA). This study was conducted at a private university in academic year 2015/2016. The study employed the quasi-experimental method with pretest-posttest control group design. The participants of the study were 83 students in which 43 students were in the experimental group and 40 students were in the control group. The finding of the study showed that students in experimental group outperformed students in control group on mathematical CT ability based on their IMA (high, medium, low) in learning Real Analysis. In addition, based on medium IMA the improvement of mathematical CT ability of students who were exposed to RTL model with informal argument was greater than that of students who were exposed to CI (conventional instruction). There was also no effect of interaction between RTL model and CI model with both (high, medium, and low) IMA increased mathematical CT ability. Finally, based on (high, medium, and low) IMA there was a significant improvement in the achievement of all indicators of mathematical CT ability of students who were exposed to RTL model with informal argument than that of students who were exposed to CI.
Chen, Xinzhong; Lo, Chiu Fan Bowen; Zheng, William; Hu, Hai; Dai, Qing; Liu, Mengkun
2017-11-01
Over the last decade, scattering-type scanning near-field optical microscopy and spectroscopy have been widely used in nano-photonics and material research due to their fine spatial resolution and broad spectral range. A number of simplified analytical models have been proposed to quantitatively understand the tip-scattered near-field signal. However, a rigorous interpretation of the experimental results is still lacking at this stage. Numerical modelings, on the other hand, are mostly done by simulating the local electric field slightly above the sample surface, which only qualitatively represents the near-field signal rendered by the tip-sample interaction. In this work, we performed a more comprehensive numerical simulation which is based on realistic experimental parameters and signal extraction procedures. By directly comparing to the experiments as well as other simulation efforts, our methods offer a more accurate quantitative description of the near-field signal, paving the way for future studies of complex systems at the nanoscale.
Bounding Averages Rigorously Using Semidefinite Programming: Mean Moments of the Lorenz System
Goluskin, David
2018-04-01
We describe methods for proving bounds on infinite-time averages in differential dynamical systems. The methods rely on the construction of nonnegative polynomials with certain properties, similarly to the way nonlinear stability can be proved using Lyapunov functions. Nonnegativity is enforced by requiring the polynomials to be sums of squares, a condition which is then formulated as a semidefinite program (SDP) that can be solved computationally. Although such computations are subject to numerical error, we demonstrate two ways to obtain rigorous results: using interval arithmetic to control the error of an approximate SDP solution, and finding exact analytical solutions to relatively small SDPs. Previous formulations are extended to allow for bounds depending analytically on parametric variables. These methods are illustrated using the Lorenz equations, a system with three state variables ( x, y, z) and three parameters (β ,σ ,r). Bounds are reported for infinite-time averages of all eighteen moments x^ly^mz^n up to quartic degree that are symmetric under (x,y)\\mapsto (-x,-y). These bounds apply to all solutions regardless of stability, including chaotic trajectories, periodic orbits, and equilibrium points. The analytical approach yields two novel bounds that are sharp: the mean of z^3 can be no larger than its value of (r-1)^3 at the nonzero equilibria, and the mean of xy^3 must be nonnegative. The interval arithmetic approach is applied at the standard chaotic parameters to bound eleven average moments that all appear to be maximized on the shortest periodic orbit. Our best upper bound on each such average exceeds its value on the maximizing orbit by less than 1%. Many bounds reported here are much tighter than would be possible without computer assistance.
Hans-Erik Andersen; Stephen E. Reutebuch; Robert J. McGaughey
2006-01-01
Tree height is an important variable in forest inventory programs but is typically time-consuming and costly to measure in the field using conventional techniques. Airborne light detection and ranging (LIDAR) provides individual tree height measurements that are highly correlated with field-derived measurements, but the imprecision of conventional field techniques does...
Cypress, Brigitte S
Issues are still raised even now in the 21st century by the persistent concern with achieving rigor in qualitative research. There is also a continuing debate about the analogous terms reliability and validity in naturalistic inquiries as opposed to quantitative investigations. This article presents the concept of rigor in qualitative research using a phenomenological study as an exemplar to further illustrate the process. Elaborating on epistemological and theoretical conceptualizations by Lincoln and Guba, strategies congruent with qualitative perspective for ensuring validity to establish the credibility of the study are described. A synthesis of the historical development of validity criteria evident in the literature during the years is explored. Recommendations are made for use of the term rigor instead of trustworthiness and the reconceptualization and renewed use of the concept of reliability and validity in qualitative research, that strategies for ensuring rigor must be built into the qualitative research process rather than evaluated only after the inquiry, and that qualitative researchers and students alike must be proactive and take responsibility in ensuring the rigor of a research study. The insights garnered here will move novice researchers and doctoral students to a better conceptual grasp of the complexity of reliability and validity and its ramifications for qualitative inquiry.
Single-case synthesis tools I: Comparing tools to evaluate SCD quality and rigor.
Zimmerman, Kathleen N; Ledford, Jennifer R; Severini, Katherine E; Pustejovsky, James E; Barton, Erin E; Lloyd, Blair P
2018-03-03
Tools for evaluating the quality and rigor of single case research designs (SCD) are often used when conducting SCD syntheses. Preferred components include evaluations of design features related to the internal validity of SCD to obtain quality and/or rigor ratings. Three tools for evaluating the quality and rigor of SCD (Council for Exceptional Children, What Works Clearinghouse, and Single-Case Analysis and Design Framework) were compared to determine if conclusions regarding the effectiveness of antecedent sensory-based interventions for young children changed based on choice of quality evaluation tool. Evaluation of SCD quality differed across tools, suggesting selection of quality evaluation tools impacts evaluation findings. Suggestions for selecting an appropriate quality and rigor assessment tool are provided and across-tool conclusions are drawn regarding the quality and rigor of studies. Finally, authors provide guidance for using quality evaluations in conjunction with outcome analyses when conducting syntheses of interventions evaluated in the context of SCD. Copyright © 2018 Elsevier Ltd. All rights reserved.
Directory of Open Access Journals (Sweden)
Andrea Sebastiano Staiti
2012-01-01
Full Text Available In this paper I present and assess Husserl's arguments against epistomological and psychological naturalism in his essay Philosophy as a Rigorous Science. I show that his critique is directed against positions that are generally more extreme than most currently debated variants of naturalism. However, Husserl has interesting thoughts to contribute to philosophy today. First, he shows that there is an important connection between naturalism in epistemology (which in his view amounts to the position that the validity of logic can be reduced to the validity natural laws of thinking and naturalism in psychology (which in his view amounts to the position that all psychic occurrences are merely parallel accompaniments of physiological occurrences. Second, he shows that a strong version of epistemological naturalism is self-undermining and fails to translate the cogency of logic in psychological terms. Third, and most importantly for current debates, he attacks Cartesianism as a form of psychological naturalism because of its construal of the psyche as a substance. Against this position, Husserl defends the necessity to formulate new epistemic aims for the investigation of consciousness. He contends that what is most interesting about consciousness is not its empirical fact but its transcendental function of granting cognitive access to all kinds of objects (both empirical and ideal. The study of this function requires a specific method (eidetics that cannot be conflated with empirical methods. I conclude that Husserl's analyses offer much-needed insight into the fabric of consciousness and compelling arguments against unwarranted metaphysical speculations about the relationship between mind and body.
The effect of temperature on the mechanical aspects of rigor mortis in a liquid paraffin model.
Ozawa, Masayoshi; Iwadate, Kimiharu; Matsumoto, Sari; Asakura, Kumiko; Ochiai, Eriko; Maebashi, Kyoko
2013-11-01
Rigor mortis is an important phenomenon to estimate the postmortem interval in forensic medicine. Rigor mortis is affected by temperature. We measured stiffness of rat muscles using a liquid paraffin model to monitor the mechanical aspects of rigor mortis at five temperatures (37, 25, 10, 5 and 0°C). At 37, 25 and 10°C, the progression of stiffness was slower in cooler conditions. At 5 and 0°C, the muscle stiffness increased immediately after the muscles were soaked in cooled liquid paraffin and then muscles gradually became rigid without going through a relaxed state. This phenomenon suggests that it is important to be careful when estimating the postmortem interval in cold seasons. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.
Sikes, Anita L; Mawson, Raymond; Stark, Janet; Warner, Robyn
2014-11-01
The delivery of a consistent quality product to the consumer is vitally important for the food industry. The aim of this study was to investigate the potential for using high frequency ultrasound applied to pre- and post-rigor beef muscle on the metabolism and subsequent quality. High frequency ultrasound (600kHz at 48kPa and 65kPa acoustic pressure) applied to post-rigor beef striploin steaks resulted in no significant effect on the texture (peak force value) of cooked steaks as measured by a Tenderometer. There was no added benefit of ultrasound treatment above that of the normal ageing process after ageing of the steaks for 7days at 4°C. Ultrasound treatment of post-rigor beef steaks resulted in a darkening of fresh steaks but after ageing for 7days at 4°C, the ultrasound-treated steaks were similar in colour to that of the aged, untreated steaks. High frequency ultrasound (2MHz at 48kPa acoustic pressure) applied to pre-rigor beef neck muscle had no effect on the pH, but the calculated exhaustion factor suggested that there was some effect on metabolism and actin-myosin interaction. However, the resultant texture of cooked, ultrasound-treated muscle was lower in tenderness compared to the control sample. After ageing for 3weeks at 0°C, the ultrasound-treated samples had the same peak force value as the control. High frequency ultrasound had no significant effect on the colour parameters of pre-rigor beef neck muscle. This proof-of-concept study showed no effect of ultrasound on quality but did indicate that the application of high frequency ultrasound to pre-rigor beef muscle shows potential for modifying ATP turnover and further investigation is warranted. Crown Copyright © 2014. Published by Elsevier B.V. All rights reserved.
Mathematical framework for fast and rigorous track fit for the ZEUS detector
Energy Technology Data Exchange (ETDEWEB)
Spiridonov, Alexander
2008-12-15
In this note we present a mathematical framework for a rigorous approach to a common track fit for trackers located in the inner region of the ZEUS detector. The approach makes use of the Kalman filter and offers a rigorous treatment of magnetic field inhomogeneity, multiple scattering and energy loss. We describe mathematical details of the implementation of the Kalman filter technique with a reduced amount of computations for a cylindrical drift chamber, barrel and forward silicon strip detectors and a forward straw drift chamber. Options with homogeneous and inhomogeneous field are discussed. The fitting of tracks in one ZEUS event takes about of 20ms on standard PC. (orig.)
Electrocardiogram artifact caused by rigors mimicking narrow complex tachycardia: a case report.
Matthias, Anne Thushara; Indrakumar, Jegarajah
2014-02-04
The electrocardiogram (ECG) is useful in the diagnosis of cardiac and non-cardiac conditions. Rigors due to shivering can cause electrocardiogram artifacts mimicking various cardiac rhythm abnormalities. We describe an 80-year-old Sri Lankan man with an abnormal electrocardiogram mimicking narrow complex tachycardia during the immediate post-operative period. Electrocardiogram changes caused by muscle tremor during rigors could mimic a narrow complex tachycardia. Identification of muscle tremor as a cause of electrocardiogram artifact can avoid unnecessary pharmacological and non-pharmacological intervention to prevent arrhythmias.
Peridynamics as a rigorous coarse-graining of atomistics for multiscale materials design
International Nuclear Information System (INIS)
Lehoucq, Richard B.; Aidun, John Bahram; Silling, Stewart Andrew; Sears, Mark P.; Kamm, James R.; Parks, Michael L.
2010-01-01
This report summarizes activities undertaken during FY08-FY10 for the LDRD Peridynamics as a Rigorous Coarse-Graining of Atomistics for Multiscale Materials Design. The goal of our project was to develop a coarse-graining of finite temperature molecular dynamics (MD) that successfully transitions from statistical mechanics to continuum mechanics. The goal of our project is to develop a coarse-graining of finite temperature molecular dynamics (MD) that successfully transitions from statistical mechanics to continuum mechanics. Our coarse-graining overcomes the intrinsic limitation of coupling atomistics with classical continuum mechanics via the FEM (finite element method), SPH (smoothed particle hydrodynamics), or MPM (material point method); namely, that classical continuum mechanics assumes a local force interaction that is incompatible with the nonlocal force model of atomistic methods. Therefore FEM, SPH, and MPM inherit this limitation. This seemingly innocuous dichotomy has far reaching consequences; for example, classical continuum mechanics cannot resolve the short wavelength behavior associated with atomistics. Other consequences include spurious forces, invalid phonon dispersion relationships, and irreconcilable descriptions/treatments of temperature. We propose a statistically based coarse-graining of atomistics via peridynamics and so develop a first of a kind mesoscopic capability to enable consistent, thermodynamically sound, atomistic-to-continuum (AtC) multiscale material simulation. Peridynamics (PD) is a microcontinuum theory that assumes nonlocal forces for describing long-range material interaction. The force interactions occurring at finite distances are naturally accounted for in PD. Moreover, PDs nonlocal force model is entirely consistent with those used by atomistics methods, in stark contrast to classical continuum mechanics. Hence, PD can be employed for mesoscopic phenomena that are beyond the realms of classical continuum mechanics and
Miyazaki, Hideki T; Miyazaki, Hiroshi; Miyano, Kenjiro
2003-09-01
We have recently identified the resonant scattering from dielectric bispheres in the specular direction, which has long been known as the specular resonance, to be a type of rainbow (a caustic) and a general phenomenon for bispheres. We discuss the details of the specular resonance on the basis of systematic calculations. In addition to the rigorous theory, which precisely describes the scattering even in the resonance regime, the ray-tracing method, which gives the scattering in the geometrical-optics limit, is used. Specular resonance is explicitly defined as strong scattering in the direction of the specular reflection from the symmetrical axis of the bisphere whose intensity exceeds that of the scattering from noninteracting bispheres. Then the range of parameters for computing a particular specular resonance is specified. This resonance becomes prominent in a wide range of refractive indices (from 1.2 to 2.2) in a wide range of size parameters (from five to infinity) and for an arbitrarily polarized light incident within an angle of 40 degrees to the symmetrical axis. This particular scattering can stay evident even when the spheres are not in contact or the sizes of the spheres are different. Thus specular resonance is a common and robust phenomenon in dielectric bispheres. Furthermore, we demonstrate that various characteristic features in the scattering from bispheres can be explained successfully by using intuitive and simple representations. Most of the significant scatterings other than the specular resonance are also understandable as caustics in geometrical-optics theory. The specular resonance becomes striking at the smallest size parameter among these caustics because its optical trajectory is composed of only the refractions at the surfaces and has an exceptionally large intensity. However, some characteristics are not accounted for by geometrical optics. In particular, the oscillatory behaviors of their scattering intensity are well described by
Kline, Joshua C.
2014-01-01
Over the past four decades, various methods have been implemented to measure synchronization of motor-unit firings. In this work, we provide evidence that prior reports of the existence of universal common inputs to all motoneurons and the presence of long-term synchronization are misleading, because they did not use sufficiently rigorous statistical tests to detect synchronization. We developed a statistically based method (SigMax) for computing synchronization and tested it with data from 17,736 motor-unit pairs containing 1,035,225 firing instances from the first dorsal interosseous and vastus lateralis muscles—a data set one order of magnitude greater than that reported in previous studies. Only firing data, obtained from surface electromyographic signal decomposition with >95% accuracy, were used in the study. The data were not subjectively selected in any manner. Because of the size of our data set and the statistical rigor inherent to SigMax, we have confidence that the synchronization values that we calculated provide an improved estimate of physiologically driven synchronization. Compared with three other commonly used techniques, ours revealed three types of discrepancies that result from failing to use sufficient statistical tests necessary to detect synchronization. 1) On average, the z-score method falsely detected synchronization at 16 separate latencies in each motor-unit pair. 2) The cumulative sum method missed one out of every four synchronization identifications found by SigMax. 3) The common input assumption method identified synchronization from 100% of motor-unit pairs studied. SigMax revealed that only 50% of motor-unit pairs actually manifested synchronization. PMID:25210152
Ar-Ar_Redux: rigorous error propagation of 40Ar/39Ar data, including covariances
Vermeesch, P.
2015-12-01
Rigorous data reduction and error propagation algorithms are needed to realise Earthtime's objective to improve the interlaboratory accuracy of 40Ar/39Ar dating to better than 1% and thereby facilitate the comparison and combination of the K-Ar and U-Pb chronometers. Ar-Ar_Redux is a new data reduction protocol and software program for 40Ar/39Ar geochronology which takes into account two previously underappreciated aspects of the method: 1. 40Ar/39Ar measurements are compositional dataIn its simplest form, the 40Ar/39Ar age equation can be written as: t = log(1+J [40Ar/39Ar-298.5636Ar/39Ar])/λ = log(1 + JR)/λ Where λ is the 40K decay constant and J is the irradiation parameter. The age t does not depend on the absolute abundances of the three argon isotopes but only on their relative ratios. Thus, the 36Ar, 39Ar and 40Ar abundances can be normalised to unity and plotted on a ternary diagram or 'simplex'. Argon isotopic data are therefore subject to the peculiar mathematics of 'compositional data', sensu Aitchison (1986, The Statistical Analysis of Compositional Data, Chapman & Hall). 2. Correlated errors are pervasive throughout the 40Ar/39Ar methodCurrent data reduction protocols for 40Ar/39Ar geochronology propagate the age uncertainty as follows: σ2(t) = [J2 σ2(R) + R2 σ2(J)] / [λ2 (1 + R J)], which implies zero covariance between R and J. In reality, however, significant error correlations are found in every step of the 40Ar/39Ar data acquisition and processing, in both single and multi collector instruments, during blank, interference and decay corrections, age calculation etc. Ar-Ar_Redux revisits every aspect of the 40Ar/39Ar method by casting the raw mass spectrometer data into a contingency table of logratios, which automatically keeps track of all covariances in a compositional context. Application of the method to real data reveals strong correlations (r2 of up to 0.9) between age measurements within a single irradiation batch. Propertly taking
Hughes, Brianna H; Greenberg, Neil J; Yang, Tom C; Skonberg, Denise I
2015-01-01
High-pressure processing (HPP) is used to increase meat safety and shelf-life, with conflicting quality effects depending on rigor status during HPP. In the seafood industry, HPP is used to shuck and pasteurize oysters, but its use on abalones has only been minimally evaluated and the effect of rigor status during HPP on abalone quality has not been reported. Farm-raised abalones (Haliotis rufescens) were divided into 12 HPP treatments and 1 unprocessed control treatment. Treatments were processed pre-rigor or post-rigor at 2 pressures (100 and 300 MPa) and 3 processing times (1, 3, and 5 min). The control was analyzed post-rigor. Uniform plugs were cut from adductor and foot meat for texture profile analysis, shear force, and color analysis. Subsamples were used for scanning electron microscopy of muscle ultrastructure. Texture profile analysis revealed that post-rigor processed abalone was significantly (P abalone meat was more tender than pre-rigor processed meat, and post-rigor processed foot meat was lighter in color than pre-rigor processed foot meat, suggesting that waiting for rigor to resolve prior to processing abalones may improve consumer perceptions of quality and market value. © 2014 Institute of Food Technologists®
Rigorous lower bound on the dynamic critical exponent of some multilevel Swendsen-Wang algorithms
International Nuclear Information System (INIS)
Li, X.; Sokal, A.D.
1991-01-01
We prove the rigorous lower bound z exp ≥α/ν for the dynamic critical exponent of a broad class of multilevel (or ''multigrid'') variants of the Swendsen-Wang algorithm. This proves that such algorithms do suffer from critical slowing down. We conjecture that such algorithms in fact lie in the same dynamic universality class as the stanard Swendsen-Wang algorithm
International Nuclear Information System (INIS)
García-Valenzuela, A; Contreras-Tello, H; Márquez-Islas, R; Sánchez-Pérez, C
2013-01-01
We derive an optical model for the light intensity distribution around the critical angle in a standard Abbe refractometer when used on absorbing homogenous fluids. The model is developed using rigorous electromagnetic optics. The obtained formula is very simple and can be used suitably in the analysis and design of optical sensors relying on Abbe type refractometry.
International Nuclear Information System (INIS)
Galatolo, Stefano; Monge, Maurizio; Nisoli, Isaia
2016-01-01
We study the problem of the rigorous computation of the stationary measure and of the rate of convergence to equilibrium of an iterated function system described by a stochastic mixture of two or more dynamical systems that are either all uniformly expanding on the interval, either all contracting. In the expanding case, the associated transfer operators satisfy a Lasota–Yorke inequality, we show how to compute a rigorous approximations of the stationary measure in the L "1 norm and an estimate for the rate of convergence. The rigorous computation requires a computer-aided proof of the contraction of the transfer operators for the maps, and we show that this property propagates to the transfer operators of the IFS. In the contracting case we perform a rigorous approximation of the stationary measure in the Wasserstein–Kantorovich distance and rate of convergence, using the same functional analytic approach. We show that a finite computation can produce a realistic computation of all contraction rates for the whole parameter space. We conclude with a description of the implementation and numerical experiments. (paper)
Muroya, Susumu; Ohnishi-Kameyama, Mayumi; Oe, Mika; Nakajima, Ikuyo; Shibata, Masahiro; Chikuni, Koichi
2007-05-16
To investigate changes in myosin light chains (MyLCs) during postmortem aging of the bovine longissimus muscle, we performed two-dimensional gel electrophoresis followed by identification with matrix-assisted laser desorption ionization time-of-flight mass spectrometry. The results of fluorescent differential gel electrophoresis showed that two spots of the myosin regulatory light chain (MyLC2) at pI values of 4.6 and 4.7 shifted toward those at pI values of 4.5 and 4.6, respectively, by 24 h postmortem when rigor mortis was completed. Meanwhile, the MyLC1 and MyLC3 spots did not change during the 14 days postmortem. Phosphoprotein-specific staining of the gels demonstrated that the MyLC2 proteins at pI values of 4.5 and 4.6 were phosphorylated. Furthermore, possible N-terminal region peptides containing one and two phosphoserine residues were detected in each mass spectrum of the MyLC2 spots at pI values of 4.5 and 4.6, respectively. These results demonstrated that MyLC2 became doubly phosphorylated during rigor formation of the bovine longissimus, suggesting involvement of the MyLC2 phosphorylation in the progress of beef rigor mortis. Bovine; myosin regulatory light chain (RLC, MyLC2); phosphorylation; rigor mortis; skeletal muscle.
Some comments on rigorous quantum field path integrals in the analytical regularization scheme
Energy Technology Data Exchange (ETDEWEB)
Botelho, Luiz C.L. [Universidade Federal Fluminense (UFF), Niteroi, RJ (Brazil). Dept. de Matematica Aplicada]. E-mail: botelho.luiz@superig.com.br
2008-07-01
Through the systematic use of the Minlos theorem on the support of cylindrical measures on R{sup {infinity}}, we produce several mathematically rigorous path integrals in interacting euclidean quantum fields with Gaussian free measures defined by generalized powers of the Laplacian operator. (author)
Some comments on rigorous quantum field path integrals in the analytical regularization scheme
International Nuclear Information System (INIS)
Botelho, Luiz C.L.
2008-01-01
Through the systematic use of the Minlos theorem on the support of cylindrical measures on R ∞ , we produce several mathematically rigorous path integrals in interacting euclidean quantum fields with Gaussian free measures defined by generalized powers of the Laplacian operator. (author)
Community historians and the dilemma of rigor vs relevance : A comment on Danziger and van Rappard
Dehue, Trudy
1998-01-01
Since the transition from finalism to contextualism, the history of science seems to be caught up in a basic dilemma. Many historians fear that with the new contextualist standards of rigorous historiography, historical research can no longer be relevant to working scientists themselves. The present
Goodman, Lisa A.; Epstein, Deborah; Sullivan, Cris M.
2018-01-01
Programs for domestic violence (DV) victims and their families have grown exponentially over the last four decades. The evidence demonstrating the extent of their effectiveness, however, often has been criticized as stemming from studies lacking scientific rigor. A core reason for this critique is the widespread belief that credible evidence can…
College Readiness in California: A Look at Rigorous High School Course-Taking
Gao, Niu
2016-01-01
Recognizing the educational and economic benefits of a college degree, education policymakers at the federal, state, and local levels have made college preparation a priority. There are many ways to measure college readiness, but one key component is rigorous high school coursework. California has not yet adopted a statewide college readiness…
Warriss, P D; Brown, S N; Knowles, T G
2003-12-13
The degree of development of rigor mortis in the carcases of slaughter pigs was assessed subjectively on a three-point scale 35 minutes after they were exsanguinated, and related to the levels of cortisol, lactate and creatine kinase in blood collected at exsanguination. Earlier rigor development was associated with higher concentrations of these stress indicators in the blood. This relationship suggests that the mean rigor score, and the frequency distribution of carcases that had or had not entered rigor, could be used as an index of the degree of stress to which the pigs had been subjected.
Lewis, Cara C; Stanick, Cameo F; Martinez, Ruben G; Weiner, Bryan J; Kim, Mimi; Barwick, Melanie; Comtois, Katherine A
2015-01-08
Identification of psychometrically strong instruments for the field of implementation science is a high priority underscored in a recent National Institutes of Health working meeting (October 2013). Existing instrument reviews are limited in scope, methods, and findings. The Society for Implementation Research Collaboration Instrument Review Project's objectives address these limitations by identifying and applying a unique methodology to conduct a systematic and comprehensive review of quantitative instruments assessing constructs delineated in two of the field's most widely used frameworks, adopt a systematic search process (using standard search strings), and engage an international team of experts to assess the full range of psychometric criteria (reliability, construct and criterion validity). Although this work focuses on implementation of psychosocial interventions in mental health and health-care settings, the methodology and results will likely be useful across a broad spectrum of settings. This effort has culminated in a centralized online open-access repository of instruments depicting graphical head-to-head comparisons of their psychometric properties. This article describes the methodology and preliminary outcomes. The seven stages of the review, synthesis, and evaluation methodology include (1) setting the scope for the review, (2) identifying frameworks to organize and complete the review, (3) generating a search protocol for the literature review of constructs, (4) literature review of specific instruments, (5) development of an evidence-based assessment rating criteria, (6) data extraction and rating instrument quality by a task force of implementation experts to inform knowledge synthesis, and (7) the creation of a website repository. To date, this multi-faceted and collaborative search and synthesis methodology has identified over 420 instruments related to 34 constructs (total 48 including subconstructs) that are relevant to implementation
Rigorous study of the gap equation for an inhomogeneous superconducting state near T/sub c/
International Nuclear Information System (INIS)
Hu, C.
1975-01-01
A rigorous analytic study of the self-consistent gap equation (symobolically Δ=F/sub T/Δ), for an inhomogeneous superconducting state, is presented in the Bogoliubov formulation. The gap function Δ (r) is taken to simulate a planar normal-superconducting phase boundary: Δ (r) =Δ/sub infinity/ tanh(αΔ/sub infinity/z/v/sub F/) THETA (z), where Δ/sub infinity/(T) is the equilibrium gap, v/subF/ is the Fermi velocity, and THETA (z) is a unit step function. First a special space integral of the gap equation proportional∫ 0 /sub +//sup infinity/(F/sub T/-Δ)(dΔ/dz) dz is evaluated essentially exactly, except for a nonperturbative WKBJ approximation used in solving the Bogoliubov--de Gennes equations. It is then expanded near the transition temperature T/sub c/ in power of Δ/sub infinity/proportional (1-T/T/sub c/) 1 / 2 , demonstrating an exact cancellation of a subseries of ''anomalous-order'' terms. The leading surviving term is found to agree in order, but not in magnitude, with the Ginzburg-Landau-Gor'kov (GLG) approximation. The discrepancy is found to be linked to the slope discontinuity in our chosen Δ. A contour-integral technique in a complex-energy plane is then devised to evaluate the local value of F/sub T/-Δ exactly. Our result reveals that near T/sub c/ this method can reproduce the GLG result essentially everywhere, except within a BCS coherence length not xi (T) exclamation from a singularity in Δ, where F/sub T/-Δ can have a singular contribution with an ''anomalous'' local magnitude, not expected from the GLG approach. This anomalous term precisely accounts for the discrepancy found in the special integral of the gap equation as mentioned above, and likely explains the ultimate origin of the anomalous terms found in the free energy of an isolated vortex line by Cleary
Diouf, Boucar; Rioux, Pierre
1999-01-01
Presents the rigor mortis process in brook charr (Salvelinus fontinalis) as a tool for better understanding skeletal muscle metabolism. Describes an activity that demonstrates how rigor mortis is related to the post-mortem decrease of muscular glycogen and ATP, how glycogen degradation produces lactic acid that lowers muscle pH, and how…
Evans, Mark I; Krantz, David A; Hallahan, Terrence; Sherwin, John; Britt, David W
2013-01-01
To determine if nuchal translucency (NT) quality correlates with the extent to which clinics vary in rigor and quality control. We correlated NT performance quality (bias and precision) of 246,000 patients with two alternative measures of clinic culture - % of cases for whom nasal bone (NB) measurements were performed and % of requisitions correctly filled for race-ethnicity and weight. When requisition errors occurred in 5% (33%), the curve lowered to 0.93 MoM (p 90%, MoM was 0.99 compared to those quality exists independent of individual variation in NT quality, and two divergent indices of program rigor are associated with NT quality. Quality control must be program wide, and to effect continued improvement in the quality of NT results across time, the cultures of clinics must become a target for intervention. Copyright © 2013 S. Karger AG, Basel.
A Framework for Rigorously Identifying Research Gaps in Qualitative Literature Reviews
DEFF Research Database (Denmark)
Müller-Bloch, Christoph; Kranz, Johann
2015-01-01
Identifying research gaps is a fundamental goal of literature reviewing. While it is widely acknowledged that literature reviews should identify research gaps, there are no methodological guidelines for how to identify research gaps in qualitative literature reviews ensuring rigor and replicability....... Our study addresses this gap and proposes a framework that should help scholars in this endeavor without stifling creativity. To develop the framework we thoroughly analyze the state-of-the-art procedure of identifying research gaps in 40 recent literature reviews using a grounded theory approach....... Based on the data, we subsequently derive a framework for identifying research gaps in qualitative literature reviews and demonstrate its application with an example. Our results provide a modus operandi for identifying research gaps, thus enabling scholars to conduct literature reviews more rigorously...
International Nuclear Information System (INIS)
Kinoshita, Masahiro; Naruse, Yuji
1981-08-01
The basic equations are derived for rigorous dynamic simulation of cryogenic distillation columns for hydrogen isotope separation. The model accounts for such factors as differences in latent heat of vaporization among the six isotopic species of molecular hydrogen, decay heat of tritium, heat transfer through the column wall and nonideality of the solutions. Provision is also made for simulation of columns with multiple feeds and multiple sidestreams. (author)
Rigor mortis development in turkey breast muscle and the effect of electrical stunning.
Alvarado, C Z; Sams, A R
2000-11-01
Rigor mortis development in turkey breast muscle and the effect of electrical stunning on this process are not well characterized. Some electrical stunning procedures have been known to inhibit postmortem (PM) biochemical reactions, thereby delaying the onset of rigor mortis in broilers. Therefore, this study was designed to characterize rigor mortis development in stunned and unstunned turkeys. A total of 154 turkey toms in two trials were conventionally processed at 20 to 22 wk of age. Turkeys were either stunned with a pulsed direct current (500 Hz, 50% duty cycle) at 35 mA (40 V) in a saline bath for 12 seconds or left unstunned as controls. At 15 min and 1, 2, 4, 8, 12, and 24 h PM, pectoralis samples were collected to determine pH, R-value, L* value, sarcomere length, and shear value. In Trial 1, the samples obtained for pH, R-value, and sarcomere length were divided into surface and interior samples. There were no significant differences between the surface and interior samples among any parameters measured. Muscle pH significantly decreased over time in stunned and unstunned birds through 2 h PM. The R-values increased to 8 h PM in unstunned birds and 24 h PM in stunned birds. The L* values increased over time, with no significant differences after 1 h PM for the controls and 2 h PM for the stunned birds. Sarcomere length increased through 2 h PM in the controls and 12 h PM in the stunned fillets. Cooked meat shear values decreased through the 1 h PM deboning time in the control fillets and 2 h PM in the stunned fillets. These results suggest that stunning delayed the development of rigor mortis through 2 h PM, but had no significant effect on the measured parameters at later time points, and that deboning turkey breasts at 2 h PM or later will not significantly impair meat tenderness.
Learning from Science and Sport - How we, Safety, "Engage with Rigor"
Herd, A.
2012-01-01
As the world of spaceflight safety is relatively small and potentially inward-looking, we need to be aware of the "outside world". We should then try to remind ourselves to be open to the possibility that data, knowledge or experience from outside of the spaceflight community may provide some constructive alternate perspectives. This paper will assess aspects from two seemingly tangential fields, science and sport, and align these with the world of safety. In doing so some useful insights will be given to the challenges we face and may provide solutions relevant in our everyday (of safety engineering). Sport, particularly a contact sport such as rugby union, requires direct interaction between members of two (opposing) teams. Professional, accurately timed and positioned interaction for a desired outcome. These interactions, whilst an essential part of the game, are however not without their constraints. The rugby scrum has constraints as to the formation and engagement of the two teams. The controlled engagement provides for an interaction between the two teams in a safe manner. The constraints arising from the reality that an incorrect engagement could cause serious injury to members of either team. In academia, scientific rigor is applied to assure that the arguments provided and the conclusions drawn in academic papers presented for publication are valid, legitimate and credible. The scientific goal of the need for rigor may be expressed in the example of achieving a statistically relevant sample size, n, in order to assure analysis validity of the data pool. A failure to apply rigor could then place the entire study at risk of failing to have the respective paper published. This paper will consider the merits of these two different aspects, scientific rigor and sports engagement, and offer a reflective look at how this may provide a "modus operandi" for safety engineers at any level whether at their desks (creating or reviewing safety assessments) or in a
Rigor force responses of permeabilized fibres from fast and slow skeletal muscles of aged rats.
Plant, D R; Lynch, G S
2001-09-01
1. Ageing is generally associated with a decline in skeletal muscle mass and strength and a slowing of muscle contraction, factors that impact upon the quality of life for the elderly. The mechanisms underlying this age-related muscle weakness have not been fully resolved. The purpose of the present study was to determine whether the decrease in muscle force as a consequence of age could be attributed partly to a decrease in the number of cross-bridges participating during contraction. 2. Given that the rigor force is proportional to the approximate total number of interacting sites between the actin and myosin filaments, we tested the null hypothesis that the rigor force of permeabilized muscle fibres from young and old rats would not be different. 3. Permeabilized fibres from the extensor digitorum longus (fast-twitch; EDL) and soleus (predominantly slow-twitch) muscles of young (6 months of age) and old (27 months of age) male F344 rats were activated in Ca2+-buffered solutions to determine force-pCa characteristics (where pCa = -log(10)[Ca2+]) and then in solutions lacking ATP and Ca2+ to determine rigor force levels. 4. The rigor forces for EDL and soleus muscle fibres were not different between young and old rats, indicating that the approximate total number of cross-bridges that can be formed between filaments did not decline with age. We conclude that the age-related decrease in force output is more likely attributed to a decrease in the force per cross-bridge and/or decreases in the efficiency of excitation-contraction coupling.
Rigorous Integration of Non-Linear Ordinary Differential Equations in Chebyshev Basis
Czech Academy of Sciences Publication Activity Database
Dzetkulič, Tomáš
2015-01-01
Roč. 69, č. 1 (2015), s. 183-205 ISSN 1017-1398 R&D Projects: GA MŠk OC10048; GA ČR GD201/09/H057 Institutional research plan: CEZ:AV0Z10300504 Keywords : Initial value problem * Rigorous integration * Taylor model * Chebyshev basis Subject RIV: IN - Informatics, Computer Science Impact factor: 1.366, year: 2015
A rigorous proof of the Landau-Peierls formula and much more
DEFF Research Database (Denmark)
Briet, Philippe; Cornean, Horia; Savoie, Baptiste
2012-01-01
We present a rigorous mathematical treatment of the zero-field orbital magnetic susceptibility of a non-interacting Bloch electron gas, at fixed temperature and density, for both metals and semiconductors/insulators. In particular, we obtain the Landau-Peierls formula in the low temperature and d...... and density limit as conjectured by Kjeldaas and Kohn (Phys Rev 105:806–813, 1957)....
International Nuclear Information System (INIS)
Cuttino, Laurie W.; Heffernan, Jill; Vera, Robyn; Rosu, Mihaela; Ramakrishnan, V. Ramesh; Arthur, Douglas W.
2011-01-01
Purpose: Multiple investigations have used the skin distance as a surrogate for the skin dose and have shown that distances 4.05 Gy/fraction. Conclusion: The initial skin dose recommendations have been based on safe use and the avoidance of significant toxicity. The results from the present study have suggested that patients might further benefit if more rigorous constraints were applied and if the skin dose were limited to 120% of the prescription dose.
Whitley, Meredith A.
2014-01-01
While the quality and quantity of research on service-learning has increased considerably over the past 20 years, researchers as well as governmental and funding agencies have called for more rigor in service-learning research. One key variable in improving rigor is using relevant existing theories to improve the research. The purpose of this…
Bombaerts, G.; Nickel, P.J.
2017-01-01
We inquire how peer and tutor feedback influences students' optimal rigor, basic needs and motivation. We analyze questionnaires from two courses in two subsequent years. We conclude that feedback in blended learning can contribute to rigor and basic needs, but it is not clear from our data what
Unmet Need: Improving mHealth Evaluation Rigor to Build the Evidence Base.
Mookherji, Sangeeta; Mehl, Garrett; Kaonga, Nadi; Mechael, Patricia
2015-01-01
mHealth-the use of mobile technologies for health-is a growing element of health system activity globally, but evaluation of those activities remains quite scant, and remains an important knowledge gap for advancing mHealth activities. In 2010, the World Health Organization and Columbia University implemented a small-scale survey to generate preliminary data on evaluation activities used by mHealth initiatives. The authors describe self-reported data from 69 projects in 29 countries. The majority (74%) reported some sort of evaluation activity, primarily nonexperimental in design (62%). The authors developed a 6-point scale of evaluation rigor comprising information on use of comparison groups, sample size calculation, data collection timing, and randomization. The mean score was low (2.4); half (47%) were conducting evaluations with a minimum threshold (4+) of rigor, indicating use of a comparison group, while less than 20% had randomized the mHealth intervention. The authors were unable to assess whether the rigor score was appropriate for the type of mHealth activity being evaluated. What was clear was that although most data came from mHealth projects pilots aimed for scale-up, few had designed evaluations that would support crucial decisions on whether to scale up and how. Whether the mHealth activity is a strategy to improve health or a tool for achieving intermediate outcomes that should lead to better health, mHealth evaluations must be improved to generate robust evidence for cost-effectiveness assessment and to allow for accurate identification of the contribution of mHealth initiatives to health systems strengthening and the impact on actual health outcomes.
Effect of muscle restraint on sheep meat tenderness with rigor mortis at 18°C.
Devine, Carrick E; Payne, Steven R; Wells, Robyn W
2002-02-01
The effect on shear force of skeletal restraint and removing muscles from lamb m. longissimus thoracis et lumborum (LT) immediately after slaughter and electrical stimulation was undertaken at a rigor temperature of 18°C (n=15). The temperature of 18°C was achieved through chilling of electrically stimulated sheep carcasses in air at 12°C, air flow 1-1.5 ms(-2). In other groups, the muscle was removed at 2.5 h post-mortem and either wrapped or left non-wrapped before being placed back on the carcass to follow carcass cooling regimes. Following rigor mortis, the meat was aged for 0, 16, 40 and 65 h at 15°C and frozen. For the non-stimulated samples, the meat was aged for 0, 12, 36 and 60 h before being frozen. The frozen meat was cooked to 75°C in an 85°C water bath and shear force values obtained from a 1 × 1 cm cross-section. Commencement of ageing was considered to take place at rigor mortis and this was taken as zero aged meat. There were no significant differences in the rate of tenderisation and initial shear force for all treatments. The 23% cook loss was similar for all wrapped and non-wrapped situations and the values decreased slightly with longer ageing durations. Wrapping was shown to mimic meat left intact on the carcass, as it prevented significant prerigor shortening. Such techniques allows muscles to be removed and placed in a controlled temperature environment to enable precise studies of ageing processes.
Study Design Rigor in Animal-Experimental Research Published in Anesthesia Journals.
Hoerauf, Janine M; Moss, Angela F; Fernandez-Bustamante, Ana; Bartels, Karsten
2018-01-01
Lack of reproducibility of preclinical studies has been identified as an impediment for translation of basic mechanistic research into effective clinical therapies. Indeed, the National Institutes of Health has revised its grant application process to require more rigorous study design, including sample size calculations, blinding procedures, and randomization steps. We hypothesized that the reporting of such metrics of study design rigor has increased over time for animal-experimental research published in anesthesia journals. PubMed was searched for animal-experimental studies published in 2005, 2010, and 2015 in primarily English-language anesthesia journals. A total of 1466 publications were graded on the performance of sample size estimation, randomization, and blinding. Cochran-Armitage test was used to assess linear trends over time for the primary outcome of whether or not a metric was reported. Interrater agreement for each of the 3 metrics (power, randomization, and blinding) was assessed using the weighted κ coefficient in a 10% random sample of articles rerated by a second investigator blinded to the ratings of the first investigator. A total of 1466 manuscripts were analyzed. Reporting for all 3 metrics of experimental design rigor increased over time (2005 to 2010 to 2015): for power analysis, from 5% (27/516), to 12% (59/485), to 17% (77/465); for randomization, from 41% (213/516), to 50% (243/485), to 54% (253/465); and for blinding, from 26% (135/516), to 38% (186/485), to 47% (217/465). The weighted κ coefficients and 98.3% confidence interval indicate almost perfect agreement between the 2 raters beyond that which occurs by chance alone (power, 0.93 [0.85, 1.0], randomization, 0.91 [0.85, 0.98], and blinding, 0.90 [0.84, 0.96]). Our hypothesis that reported metrics of rigor in animal-experimental studies in anesthesia journals have increased during the past decade was confirmed. More consistent reporting, or explicit justification for absence
DEFF Research Database (Denmark)
Gaspar, Jozsef; Ritschel, Tobias Kasper Skovborg; Jørgensen, John Bagterp
2017-01-01
-linear model based control to achieve optimal techno-economic performance. Accordingly, this work presents a computationally efficient and novel approach for solving a tray-by-tray equilibrium model and its implementation for open-loop optimal-control of a cryogenic distillation column. Here, the optimisation...... objective is to reduce the cost of compression in a volatile electricity market while meeting the production requirements, i.e. product flow rate and purity. This model is implemented in Matlab and uses the ThermoLib rigorous thermodynamic library. The present work represents a first step towards plant...
Supersymmetry and the Parisi-Sourlas dimensional reduction: A rigorous proof
International Nuclear Information System (INIS)
Klein, A.; Landau, L.J.; Perez, J.F.
1984-01-01
Functional integrals that are formally related to the average correlation functions of a classical field theory in the presence of random external sources are given a rigorous meaning. Their dimensional reduction to the Schwinger functions of the corresponding quantum field theory in two fewer dimensions is proven. This is done by reexpressing those functional integrals as expectations of a supersymmetric field theory. The Parisi-Sourlas dimensional reduction of a supersymmetric field theory to a usual quantum field theory in two fewer dimensions is proven. (orig.)
A Rigorous Treatment of Energy Extraction from a Rotating Black Hole
Finster, F.; Kamran, N.; Smoller, J.; Yau, S.-T.
2009-05-01
The Cauchy problem is considered for the scalar wave equation in the Kerr geometry. We prove that by choosing a suitable wave packet as initial data, one can extract energy from the black hole, thereby putting supperradiance, the wave analogue of the Penrose process, into a rigorous mathematical framework. We quantify the maximal energy gain. We also compute the infinitesimal change of mass and angular momentum of the black hole, in agreement with Christodoulou’s result for the Penrose process. The main mathematical tool is our previously derived integral representation of the wave propagator.
Devine, Carrick; Wells, Robyn; Lowe, Tim; Waller, John
2014-01-01
The M. longissimus from lambs electrically stimulated at 15 min post-mortem were removed after grading, wrapped in polythene film and held at 4 (n=6), 7 (n=6), 15 (n=6, n=8) and 35°C (n=6), until rigor mortis then aged at 15°C for 0, 4, 24 and 72 h post-rigor. Centrifuged free water increased exponentially, and bound water, dry matter and shear force decreased exponentially over time. Decreases in shear force and increases in free water were closely related (r(2)=0.52) and were unaffected by pre-rigor temperatures. © 2013.
Parent Management Training-Oregon Model: Adapting Intervention with Rigorous Research.
Forgatch, Marion S; Kjøbli, John
2016-09-01
Parent Management Training-Oregon Model (PMTO(®) ) is a set of theory-based parenting programs with status as evidence-based treatments. PMTO has been rigorously tested in efficacy and effectiveness trials in different contexts, cultures, and formats. Parents, the presumed agents of change, learn core parenting practices, specifically skill encouragement, limit setting, monitoring/supervision, interpersonal problem solving, and positive involvement. The intervention effectively prevents and ameliorates children's behavior problems by replacing coercive interactions with positive parenting practices. Delivery format includes sessions with individual families in agencies or families' homes, parent groups, and web-based and telehealth communication. Mediational models have tested parenting practices as mechanisms of change for children's behavior and found support for the theory underlying PMTO programs. Moderating effects include children's age, maternal depression, and social disadvantage. The Norwegian PMTO implementation is presented as an example of how PMTO has been tailored to reach diverse populations as delivered by multiple systems of care throughout the nation. An implementation and research center in Oslo provides infrastructure and promotes collaboration between practitioners and researchers to conduct rigorous intervention research. Although evidence-based and tested within a wide array of contexts and populations, PMTO must continue to adapt to an ever-changing world. © 2016 Family Process Institute.
Efficiency versus speed in quantum heat engines: Rigorous constraint from Lieb-Robinson bound
Shiraishi, Naoto; Tajima, Hiroyasu
2017-08-01
A long-standing open problem whether a heat engine with finite power achieves the Carnot efficiency is investgated. We rigorously prove a general trade-off inequality on thermodynamic efficiency and time interval of a cyclic process with quantum heat engines. In a first step, employing the Lieb-Robinson bound we establish an inequality on the change in a local observable caused by an operation far from support of the local observable. This inequality provides a rigorous characterization of the following intuitive picture that most of the energy emitted from the engine to the cold bath remains near the engine when the cyclic process is finished. Using this description, we prove an upper bound on efficiency with the aid of quantum information geometry. Our result generally excludes the possibility of a process with finite speed at the Carnot efficiency in quantum heat engines. In particular, the obtained constraint covers engines evolving with non-Markovian dynamics, which almost all previous studies on this topic fail to address.
Rigorous RG Algorithms and Area Laws for Low Energy Eigenstates in 1D
Arad, Itai; Landau, Zeph; Vazirani, Umesh; Vidick, Thomas
2017-11-01
One of the central challenges in the study of quantum many-body systems is the complexity of simulating them on a classical computer. A recent advance (Landau et al. in Nat Phys, 2015) gave a polynomial time algorithm to compute a succinct classical description for unique ground states of gapped 1D quantum systems. Despite this progress many questions remained unsolved, including whether there exist efficient algorithms when the ground space is degenerate (and of polynomial dimension in the system size), or for the polynomially many lowest energy states, or even whether such states admit succinct classical descriptions or area laws. In this paper we give a new algorithm, based on a rigorously justified RG type transformation, for finding low energy states for 1D Hamiltonians acting on a chain of n particles. In the process we resolve some of the aforementioned open questions, including giving a polynomial time algorithm for poly( n) degenerate ground spaces and an n O(log n) algorithm for the poly( n) lowest energy states (under a mild density condition). For these classes of systems the existence of a succinct classical description and area laws were not rigorously proved before this work. The algorithms are natural and efficient, and for the case of finding unique ground states for frustration-free Hamiltonians the running time is {\\tilde{O}(nM(n))} , where M( n) is the time required to multiply two n × n matrices.
"Snow White" Coating Protects SpaceX Dragon's Trunk Against Rigors of Space
McMahan, Tracy
2013-01-01
He described it as "snow white." But NASA astronaut Don Pettit was not referring to the popular children's fairy tale. Rather, he was talking about the white coating of the Space Exploration Technologies Corp. (SpaceX) Dragon spacecraft that reflected from the International Space Station s light. As it approached the station for the first time in May 2012, the Dragon s trunk might have been described as the "fairest of them all," for its pristine coating, allowing Pettit to clearly see to maneuver the robotic arm to grab the Dragon for a successful nighttime berthing. This protective thermal control coating, developed by Alion Science and Technology Corp., based in McLean, Va., made its bright appearance again with the March 1 launch of SpaceX's second commercial resupply mission. Named Z-93C55, the coating was applied to the cargo portion of the Dragon to protect it from the rigors of space. "For decades, Alion has produced coatings to protect against the rigors of space," said Michael Kenny, senior chemist with Alion. "As space missions evolved, there was a growing need to dissipate electrical charges that build up on the exteriors of spacecraft, or there could be damage to the spacecraft s electronics. Alion's research led us to develop materials that would meet this goal while also providing thermal controls. The outcome of this research was Alion's proprietary Z-93C55 coating."
Rigorous Screening Technology for Identifying Suitable CO2 Storage Sites II
Energy Technology Data Exchange (ETDEWEB)
George J. Koperna Jr.; Vello A. Kuuskraa; David E. Riestenberg; Aiysha Sultana; Tyler Van Leeuwen
2009-06-01
This report serves as the final technical report and users manual for the 'Rigorous Screening Technology for Identifying Suitable CO2 Storage Sites II SBIR project. Advanced Resources International has developed a screening tool by which users can technically screen, assess the storage capacity and quantify the costs of CO2 storage in four types of CO2 storage reservoirs. These include CO2-enhanced oil recovery reservoirs, depleted oil and gas fields (non-enhanced oil recovery candidates), deep coal seems that are amenable to CO2-enhanced methane recovery, and saline reservoirs. The screening function assessed whether the reservoir could likely serve as a safe, long-term CO2 storage reservoir. The storage capacity assessment uses rigorous reservoir simulation models to determine the timing, ultimate storage capacity, and potential for enhanced hydrocarbon recovery. Finally, the economic assessment function determines both the field-level and pipeline (transportation) costs for CO2 sequestration in a given reservoir. The screening tool has been peer reviewed at an Electrical Power Research Institute (EPRI) technical meeting in March 2009. A number of useful observations and recommendations emerged from the Workshop on the costs of CO2 transport and storage that could be readily incorporated into a commercial version of the Screening Tool in a Phase III SBIR.
International Nuclear Information System (INIS)
Hoffstaetter, G.H.
1994-12-01
Analyzing stability of particle motion in storage rings contributes to the general field of stability analysis in weakly nonlinear motion. A method which we call pseudo invariant estimation (PIE) is used to compute lower bounds on the survival time in circular accelerators. The pseudeo invariants needed for this approach are computed via nonlinear perturbative normal form theory and the required global maxima of the highly complicated multivariate functions could only be rigorously bound with an extension of interval arithmetic. The bounds on the survival times are large enough to the relevant; the same is true for the lower bounds on dynamical aperatures, which can be computed. The PIE method can lead to novel design criteria with the objective of maximizing the survival time. A major effort in the direction of rigourous predictions only makes sense if accurate models of accelerators are available. Fringe fields often have a significant influence on optical properties, but the computation of fringe-field maps by DA based integration is slower by several orders of magnitude than DA evaluation of the propagator for main-field maps. A novel computation of fringe-field effects called symplectic scaling (SYSCA) is introduced. It exploits the advantages of Lie transformations, generating functions, and scaling properties and is extremely accurate. The computation of fringe-field maps is typically made nearly two orders of magnitude faster. (orig.)
Liu, Gang; Zhang, Jian; Bao, Jie
2016-01-01
Cost reduction on cellulase enzyme usage has been the central effort in the commercialization of fuel ethanol production from lignocellulose biomass. Therefore, establishing an accurate evaluation method on cellulase enzyme cost is crucially important to support the health development of the future biorefinery industry. Currently, the cellulase cost evaluation methods were complicated and various controversial or even conflict results were presented. To give a reliable evaluation on this important topic, a rigorous analysis based on the Aspen Plus flowsheet simulation in the commercial scale ethanol plant was proposed in this study. The minimum ethanol selling price (MESP) was used as the indicator to show the impacts of varying enzyme supply modes, enzyme prices, process parameters, as well as enzyme loading on the enzyme cost. The results reveal that the enzyme cost drives the cellulosic ethanol price below the minimum profit point when the enzyme is purchased from the current industrial enzyme market. An innovative production of cellulase enzyme such as on-site enzyme production should be explored and tested in the industrial scale to yield an economically sound enzyme supply for the future cellulosic ethanol production.
Birkeland, S; Akse, L
2010-01-01
Improved slaughtering procedures in the salmon industry have caused a delayed onset of rigor mortis and, thus, a potential for pre-rigor secondary processing. The aim of this study was to investigate the effect of rigor status at time of processing on quality traits color, texture, sensory, microbiological, in injection salted, and cold-smoked Atlantic salmon (Salmo salar). Injection of pre-rigor fillets caused a significant (Prigor processed fillets; however, post-rigor (1477 ± 38 g) fillets had a significant (P>0.05) higher fracturability than pre-rigor fillets (1369 ± 71 g). Pre-rigor fillets were significantly (Prigor fillets (37.8 ± 0.8) and had significantly lower (Prigor processed fillets. This study showed that similar quality characteristics can be obtained in cold-smoked products processed either pre- or post-rigor when using suitable injection salting protocols and smoking techniques. © 2010 Institute of Food Technologists®
Paulo Leminski : um estudo sobre o rigor e o relaxo em suas poesias
Dhynarte de Borba e Albuquerque
2005-01-01
O trabalho examina a trajetória da poesia de Paulo Leminski, buscando estabelecer os termos do humor, da pesquisa metalingüística e do eu-lírico, e que não deixa de exibir traços da poesia marginal dos 70. Um autor que trabalhou com a busca do rigor concretista mediante os procedimentos da fala cotidiana mais ou menos relaxada. O esforço poético do curitibano Leminski é uma “linha que nunca termina” – ele escreveu poesias, romances, peças de publicidade, letras de música e fez traduções. Em t...
Rigorous decoupling between edge states in frustrated spin chains and ladders
Chepiga, Natalia; Mila, Frédéric
2018-05-01
We investigate the occurrence of exact zero modes in one-dimensional quantum magnets of finite length that possess edge states. Building on conclusions first reached in the context of the spin-1/2 X Y chain in a field and then for the spin-1 J1-J2 Heisenberg model, we show that the development of incommensurate correlations in the bulk invariably leads to oscillations in the sign of the coupling between edge states, and hence to exact zero energy modes at the crossing points where the coupling between the edge states rigorously vanishes. This is true regardless of the origin of the frustration (e.g., next-nearest-neighbor coupling or biquadratic coupling for the spin-1 chain), of the value of the bulk spin (we report on spin-1/2, spin-1, and spin-2 examples), and of the value of the edge-state emergent spin (spin-1/2 or spin-1).
Using Project Complexity Determinations to Establish Required Levels of Project Rigor
Energy Technology Data Exchange (ETDEWEB)
Andrews, Thomas D.
2015-10-01
This presentation discusses the project complexity determination process that was developed by National Security Technologies, LLC, for the U.S. Department of Energy, National Nuclear Security Administration Nevada Field Office for implementation at the Nevada National Security Site (NNSS). The complexity determination process was developed to address the diversity of NNSS project types, size, and complexity; to fill the need for one procedure but with provision for tailoring the level of rigor to the project type, size, and complexity; and to provide consistent, repeatable, effective application of project management processes across the enterprise; and to achieve higher levels of efficiency in project delivery. These needs are illustrated by the wide diversity of NNSS projects: Defense Experimentation, Global Security, weapons tests, military training areas, sensor development and testing, training in realistic environments, intelligence community support, sensor development, environmental restoration/waste management, and disposal of radioactive waste, among others.
Release of major ions during rigor mortis development in kid Longissimus dorsi muscle.
Feidt, C; Brun-Bellut, J
1999-01-01
Ionic strength plays an important role in post mortem muscle changes. Its increase is due to ion release during the development of rigor mortis. Twelve alpine kids were used to study the effects of chilling and meat pH on ion release. Free ions were measured in Longissimus dorsi muscle by capillary electrophoresis after water extraction. All free ion concentrations increased after death, but there were differences between ions. Temperature was not a factor affecting ion release in contrast to ultimate pH value. Three release mechanisms are believed to coexist: a passive binding to proteins, which stops as pH decreases, an active segregation which stops as ATP disappears and the production of metabolites due to anaerobic glycolysis.
Rigorous approach to the comparison between experiment and theory in Casimir force measurements
International Nuclear Information System (INIS)
Klimchitskaya, G L; Chen, F; Decca, R S; Fischbach, E; Krause, D E; Lopez, D; Mohideen, U; Mostepanenko, V M
2006-01-01
In most experiments on the Casimir force the comparison between measurement data and theory was done using the concept of the root-mean-square deviation, a procedure that has been criticized in the literature. Here we propose a special statistical analysis which should be performed separately for the experimental data and for the results of the theoretical computations. In so doing, the random, systematic and total experimental errors are found as functions of separation, taking into account the distribution laws for each error at 95% confidence. Independently, all theoretical errors are combined to obtain the total theoretical error at the same confidence. Finally, the confidence interval for the differences between theoretical and experimental values is obtained as a function of separation. This rigorous approach is applied to two recent experiments on the Casimir effect
Rigorous description of holograms of particles illuminated by an astigmatic elliptical Gaussian beam
Energy Technology Data Exchange (ETDEWEB)
Yuan, Y J; Ren, K F; Coetmellec, S; Lebrun, D, E-mail: fang.ren@coria.f [UMR 6614/CORIA, CNRS and Universite et INSA de Rouen Avenue de l' Universite BP 12, 76801 Saint Etienne du Rouvray (France)
2009-02-01
The digital holography is a non-intrusive optical metrology and well adapted for the measurement of the size and velocity field of particles in the spray of a fluid. The simplified model of an opaque disk is often used in the treatment of the diagrams and therefore the refraction and the third dimension diffraction of the particle are not taken into account. We present in this paper a rigorous description of the holographic diagrams and evaluate the effects of the refraction and the third dimension diffraction by comparison to the opaque disk model. It is found that the effects are important when the real part of the refractive index is near unity or the imaginary part is non zero but small.
Study design elements for rigorous quasi-experimental comparative effectiveness research.
Maciejewski, Matthew L; Curtis, Lesley H; Dowd, Bryan
2013-03-01
Quasi-experiments are likely to be the workhorse study design used to generate evidence about the comparative effectiveness of alternative treatments, because of their feasibility, timeliness, affordability and external validity compared with randomized trials. In this review, we outline potential sources of discordance in results between quasi-experiments and experiments, review study design choices that can improve the internal validity of quasi-experiments, and outline innovative data linkage strategies that may be particularly useful in quasi-experimental comparative effectiveness research. There is an urgent need to resolve the debate about the evidentiary value of quasi-experiments since equal consideration of rigorous quasi-experiments will broaden the base of evidence that can be brought to bear in clinical decision-making and governmental policy-making.
Increasing rigor in NMR-based metabolomics through validated and open source tools.
Eghbalnia, Hamid R; Romero, Pedro R; Westler, William M; Baskaran, Kumaran; Ulrich, Eldon L; Markley, John L
2017-02-01
The metabolome, the collection of small molecules associated with an organism, is a growing subject of inquiry, with the data utilized for data-intensive systems biology, disease diagnostics, biomarker discovery, and the broader characterization of small molecules in mixtures. Owing to their close proximity to the functional endpoints that govern an organism's phenotype, metabolites are highly informative about functional states. The field of metabolomics identifies and quantifies endogenous and exogenous metabolites in biological samples. Information acquired from nuclear magnetic spectroscopy (NMR), mass spectrometry (MS), and the published literature, as processed by statistical approaches, are driving increasingly wider applications of metabolomics. This review focuses on the role of databases and software tools in advancing the rigor, robustness, reproducibility, and validation of metabolomics studies. Copyright © 2016. Published by Elsevier Ltd.
Liu, Tao; Zhu, Guanghu; He, Jianfeng; Song, Tie; Zhang, Meng; Lin, Hualiang; Xiao, Jianpeng; Zeng, Weilin; Li, Xing; Li, Zhihao; Xie, Runsheng; Zhong, Haojie; Wu, Xiaocheng; Hu, Wenbiao; Zhang, Yonghui; Ma, Wenjun
2017-08-02
Dengue fever is a severe public heath challenge in south China. A dengue outbreak was reported in Chaozhou city, China in 2015. Intensified interventions were implemented by the government to control the epidemic. However, it is still unknown the degree to which intensified control measures reduced the size of the epidemics, and when should such measures be initiated to reduce the risk of large dengue outbreaks developing? We selected Xiangqiao district as study setting because the majority of the indigenous cases (90.6%) in Chaozhou city were from this district. The numbers of daily indigenous dengue cases in 2015 were collected through the national infectious diseases and vectors surveillance system, and daily Breteau Index (BI) data were reported by local public health department. We used a compartmental dynamic SEIR (Susceptible, Exposed, Infected and Removed) model to assess the effectiveness of control interventions, and evaluate the control effect of intervention timing on dengue epidemic. A total of 1250 indigenous dengue cases was reported from Xiangqiao district. The results of SEIR modeling using BI as an indicator of actual control interventions showed a total of 1255 dengue cases, which is close to the reported number (n = 1250). The size and duration of the outbreak were highly sensitive to the intensity and timing of interventions. The more rigorous and earlier the control interventions implemented, the more effective it yielded. Even if the interventions were initiated several weeks after the onset of the dengue outbreak, the interventions were shown to greatly impact the prevalence and duration of dengue outbreak. This study suggests that early implementation of rigorous dengue interventions can effectively reduce the epidemic size and shorten the epidemic duration.
Bringing scientific rigor to community-developed programs in Hong Kong.
Fabrizio, Cecilia S; Hirschmann, Malia R; Lam, Tai Hing; Cheung, Teresa; Pang, Irene; Chan, Sophia; Stewart, Sunita M
2012-12-31
This paper describes efforts to generate evidence for community-developed programs to enhance family relationships in the Chinese culture of Hong Kong, within the framework of community-based participatory research (CBPR). The CBPR framework was applied to help maximize the development of the intervention and the public health impact of the studies, while enhancing the capabilities of the social service sector partners. Four academic-community research teams explored the process of designing and implementing randomized controlled trials in the community. In addition to the expected cultural barriers between teams of academics and community practitioners, with their different outlooks, concerns and languages, the team navigated issues in utilizing the principles of CBPR unique to this Chinese culture. Eventually the team developed tools for adaptation, such as an emphasis on building the relationship while respecting role delineation and an iterative process of defining the non-negotiable parameters of research design while maintaining scientific rigor. Lessons learned include the risk of underemphasizing the size of the operational and skills shift between usual agency practices and research studies, the importance of minimizing non-negotiable parameters in implementing rigorous research designs in the community, and the need to view community capacity enhancement as a long term process. The four pilot studies under the FAMILY Project demonstrated that nuanced design adaptations, such as wait list controls and shorter assessments, better served the needs of the community and led to the successful development and vigorous evaluation of a series of preventive, family-oriented interventions in the Chinese culture of Hong Kong.
McKee, S R; Sams, A R
1998-01-01
Development of rigor mortis at elevated post-mortem temperatures may contribute to turkey meat characteristics that are similar to those found in pale, soft, exudative pork. To evaluate this effect, 36 Nicholas tom turkeys were processed at 19 wk of age and placed in water at 40, 20, and 0 C immediately after evisceration. Pectoralis muscle samples were taken at 15 min, 30 min, 1 h, 2 h, and 4 h post-mortem and analyzed for R-value (an indirect measure of adenosine triphosphate), glycogen, pH, color, and sarcomere length. At 4 h, the remaining intact Pectoralis muscle was harvested, and aged on ice 23 h, and analyzed for drip loss, cook loss, shear values, and sarcomere length. By 15 min post-mortem, the 40 C treatment had higher R-values, which persisted through 4 h. By 1 h, the 40 C treatment pH and glycogen levels were lower than the 0 C treatment; however, they did not differ from those of the 20 C treatment. Increased L* values indicated that color became more pale by 2 h post-mortem in the 40 C treatment when compared to the 20 and 0 C treatments. Drip loss, cook loss, and shear value were increased whereas sarcomere lengths were decreased as a result of the 40 C treatment. These findings suggested that elevated post-mortem temperatures during processing resulted in acceleration of rigor mortis and biochemical changes in the muscle that produced pale, exudative meat characteristics in turkey.
Provencher, Steeve; Archer, Stephen L; Ramirez, F Daniel; Hibbert, Benjamin; Paulin, Roxane; Boucherat, Olivier; Lacasse, Yves; Bonnet, Sébastien
2018-03-30
Despite advances in our understanding of the pathophysiology and the management of pulmonary arterial hypertension (PAH), significant therapeutic gaps remain for this devastating disease. Yet, few innovative therapies beyond the traditional pathways of endothelial dysfunction have reached clinical trial phases in PAH. Although there are inherent limitations of the currently available models of PAH, the leaky pipeline of innovative therapies relates, in part, to flawed preclinical research methodology, including lack of rigour in trial design, incomplete invasive hemodynamic assessment, and lack of careful translational studies that replicate randomized controlled trials in humans with attention to adverse effects and benefits. Rigorous methodology should include the use of prespecified eligibility criteria, sample sizes that permit valid statistical analysis, randomization, blinded assessment of standardized outcomes, and transparent reporting of results. Better design and implementation of preclinical studies can minimize inherent flaws in the models of PAH, reduce the risk of bias, and enhance external validity and our ability to distinguish truly promising therapies form many false-positive or overstated leads. Ideally, preclinical studies should use advanced imaging, study several preclinical pulmonary hypertension models, or correlate rodent and human findings and consider the fate of the right ventricle, which is the major determinant of prognosis in human PAH. Although these principles are widely endorsed, empirical evidence suggests that such rigor is often lacking in pulmonary hypertension preclinical research. The present article discusses the pitfalls in the design of preclinical pulmonary hypertension trials and discusses opportunities to create preclinical trials with improved predictive value in guiding early-phase drug development in patients with PAH, which will need support not only from researchers, peer reviewers, and editors but also from
Pavlacky, David C; Lukacs, Paul M; Blakesley, Jennifer A; Skorkowsky, Robert C; Klute, David S; Hahn, Beth A; Dreitz, Victoria J; George, T Luke; Hanni, David J
2017-01-01
Monitoring is an essential component of wildlife management and conservation. However, the usefulness of monitoring data is often undermined by the lack of 1) coordination across organizations and regions, 2) meaningful management and conservation objectives, and 3) rigorous sampling designs. Although many improvements to avian monitoring have been discussed, the recommendations have been slow to emerge in large-scale programs. We introduce the Integrated Monitoring in Bird Conservation Regions (IMBCR) program designed to overcome the above limitations. Our objectives are to outline the development of a statistically defensible sampling design to increase the value of large-scale monitoring data and provide example applications to demonstrate the ability of the design to meet multiple conservation and management objectives. We outline the sampling process for the IMBCR program with a focus on the Badlands and Prairies Bird Conservation Region (BCR 17). We provide two examples for the Brewer's sparrow (Spizella breweri) in BCR 17 demonstrating the ability of the design to 1) determine hierarchical population responses to landscape change and 2) estimate hierarchical habitat relationships to predict the response of the Brewer's sparrow to conservation efforts at multiple spatial scales. The collaboration across organizations and regions provided economy of scale by leveraging a common data platform over large spatial scales to promote the efficient use of monitoring resources. We designed the IMBCR program to address the information needs and core conservation and management objectives of the participating partner organizations. Although it has been argued that probabilistic sampling designs are not practical for large-scale monitoring, the IMBCR program provides a precedent for implementing a statistically defensible sampling design from local to bioregional scales. We demonstrate that integrating conservation and management objectives with rigorous statistical
Directory of Open Access Journals (Sweden)
David C Pavlacky
Full Text Available Monitoring is an essential component of wildlife management and conservation. However, the usefulness of monitoring data is often undermined by the lack of 1 coordination across organizations and regions, 2 meaningful management and conservation objectives, and 3 rigorous sampling designs. Although many improvements to avian monitoring have been discussed, the recommendations have been slow to emerge in large-scale programs. We introduce the Integrated Monitoring in Bird Conservation Regions (IMBCR program designed to overcome the above limitations. Our objectives are to outline the development of a statistically defensible sampling design to increase the value of large-scale monitoring data and provide example applications to demonstrate the ability of the design to meet multiple conservation and management objectives. We outline the sampling process for the IMBCR program with a focus on the Badlands and Prairies Bird Conservation Region (BCR 17. We provide two examples for the Brewer's sparrow (Spizella breweri in BCR 17 demonstrating the ability of the design to 1 determine hierarchical population responses to landscape change and 2 estimate hierarchical habitat relationships to predict the response of the Brewer's sparrow to conservation efforts at multiple spatial scales. The collaboration across organizations and regions provided economy of scale by leveraging a common data platform over large spatial scales to promote the efficient use of monitoring resources. We designed the IMBCR program to address the information needs and core conservation and management objectives of the participating partner organizations. Although it has been argued that probabilistic sampling designs are not practical for large-scale monitoring, the IMBCR program provides a precedent for implementing a statistically defensible sampling design from local to bioregional scales. We demonstrate that integrating conservation and management objectives with rigorous
International Nuclear Information System (INIS)
Lim, S C; Teo, L P
2007-01-01
We derive rigorously explicit formulae of the Casimir free energy at finite temperature for massless scalar field and electromagnetic field confined in a closed rectangular cavity with different boundary conditions by a zeta regularization method. We study both the low and high temperature expansions of the free energy. In each case, we write the free energy as a sum of a polynomial in temperature plus exponentially decay terms. We show that the free energy is always a decreasing function of temperature. In the cases of massless scalar field with the Dirichlet boundary condition and electromagnetic field, the zero temperature Casimir free energy might be positive. In each of these cases, there is a unique transition temperature (as a function of the side lengths of the cavity) where the Casimir energy changes from positive to negative. When the space dimension is equal to two and three, we show graphically the dependence of this transition temperature on the side lengths of the cavity. Finally we also show that we can obtain the results for a non-closed rectangular cavity by letting the size of some directions of a closed cavity go to infinity, and we find that these results agree with the usual integration prescription adopted by other authors
Directory of Open Access Journals (Sweden)
Maria Jesús Algar
2013-01-01
Full Text Available This paper presents a complete assessment to the interferences caused in the nearby radio systems by wind turbines. Three different parameters have been considered: the scattered field of a wind turbine, its radar cross-section (RCS, and the Doppler shift generated by the rotating movements of the blades. These predictions are very useful for the study of the influence of wind farms in radio systems. To achieve this, both high-frequency techniques, such as Geometrical Theory of Diffraction/Uniform Theory of Diffraction (GTD/UTD and Physical Optics (PO, and rigorous techniques, like Method of Moments (MoM, have been used. In the analysis of the scattered field, conductor and dielectric models of the wind turbine have been analyzed. In this way, realistic results can be obtained. For all cases under analysis, the wind turbine has been modeled with NURBS (Non-Uniform Rational B-Spline surfaces since they allow the real shape of the object to be accurately replicated with very little information.
Fadıloğlu, Eylem Ezgi; Serdaroğlu, Meltem
2018-01-01
Abstract This study was conducted to evaluate the effects of pre and post-rigor marinade injections on some quality parameters of Longissimus dorsi (LD) muscles. Three marinade formulations were prepared with 2% NaCl, 2% NaCl+0.5 M lactic acid and 2% NaCl+0.5 M sodium lactate. In this study marinade uptake, pH, free water, cooking loss, drip loss and color properties were analyzed. Injection time had significant effect on marinade uptake levels of samples. Regardless of marinate formulation, marinade uptake of pre-rigor samples injected with marinade solutions were higher than post rigor samples. Injection of sodium lactate increased pH values of samples whereas lactic acid injection decreased pH. Marinade treatment and storage period had significant effect on cooking loss. At each evaluation period interaction between marinade treatment and injection time showed different effect on free water content. Storage period and marinade application had significant effect on drip loss values. Drip loss in all samples increased during the storage. During all storage days, lowest CIE L* value was found in pre-rigor samples injected with sodium lactate. Lactic acid injection caused color fade in pre-rigor and post-rigor samples. Interaction between marinade treatment and storage period was found statistically significant (p<0.05). At day 0 and 3, the lowest CIE b* values obtained pre-rigor samples injected with sodium lactate and there were no differences were found in other samples. At day 6, no significant differences were found in CIE b* values of all samples. PMID:29805282
Nevalainen, T J; Gavin, J B; Seelye, R N; Whitehouse, S; Donnell, M
1978-07-01
The effect of normal and artificially induced rigor mortis on the vascular passage of erythrocytes and fluid through isolated dog hearts was studied. Increased rigidity of 6-mm thick transmural sections through the centre of the posterior papillary muscle was used as an indication of rigor. The perfusibility of the myocardium was tested by injecting 10 ml of 1% sodium fluorescein in Hanks solution into the circumflex branch of the left coronary artery. In prerigor hearts (20 minute incubation) fluorescein perfused the myocardium evenly whether or not it was preceded by an injection of 10 ml of heparinized dog blood. Rigor mortis developed in all hearts after 90 minutes incubation or within 20 minutes of perfusing the heart with 50 ml of 5 mM iodoacetate in Hanks solution. Fluorescein injected into hearts in rigor did not enter the posterior papillary muscle and adjacent subendocardium whether or not it was preceded by heparinized blood. Thus the vascular occlusion caused by rigor in the dog heart appears to be so effective that it prevents flow into the subendocardium of small soluble ions such as fluorescein.
Ogilvie, Karen; Olde Daalhuis, Adri B.
2015-11-01
By application of the theory for second-order linear differential equations with two turning points developed in [Olver F.W.J., Philos. Trans. Roy. Soc. London Ser. A 278 (1975), 137-174], uniform asymptotic approximations are obtained in the first part of this paper for the Lamé and Mathieu functions with a large real parameter. These approximations are expressed in terms of parabolic cylinder functions, and are uniformly valid in their respective real open intervals. In all cases explicit bounds are supplied for the error terms associated with the approximations. Approximations are also obtained for the large order behaviour for the respective eigenvalues. We restrict ourselves to a two term uniform approximation. Theoretically more terms in these approximations could be computed, but the coefficients would be very complicated. In the second part of this paper we use a simplified method to obtain uniform asymptotic expansions for these functions. The coefficients are just polynomials and satisfy simple recurrence relations. The price to pay is that these asymptotic expansions hold only in a shrinking interval as their respective parameters become large; this interval however encapsulates all the interesting oscillatory behaviour of the functions. This simplified method also gives many terms in asymptotic expansions for these eigenvalues, derived simultaneously with the coefficients in the function expansions. We provide rigorous realistic error bounds for the function expansions when truncated and order estimates for the error when the eigenvalue expansions are truncated. With this paper we confirm that many of the formal results in the literature are correct.
Simulation of photonic waveguides with deterministic aperiodic nanostructures for biosensing
DEFF Research Database (Denmark)
Neustock, Lars Thorben; Paulsen, Moritz; Jahns, Sabrina
2016-01-01
Photonic waveguides with deterministic aperiodic corrugations offer rich spectral characteristics under surface-normal illumination. The finite-element method (FEM), the finite-difference time-domain (FDTD) method and a rigorous coupled wave algorithm (RCWA) are compared for computing the near...
Layout optimization of DRAM cells using rigorous simulation model for NTD
Jeon, Jinhyuck; Kim, Shinyoung; Park, Chanha; Yang, Hyunjo; Yim, Donggyu; Kuechler, Bernd; Zimmermann, Rainer; Muelders, Thomas; Klostermann, Ulrich; Schmoeller, Thomas; Do, Mun-hoe; Choi, Jung-Hoe
2014-03-01
scanning electron microscope (SEM) measurements. High resist impact and difficult model data acquisition demand for a simulation model that hat is capable of extrapolating reliably beyond its calibration dataset. We use rigorous simulation models to provide that predictive performance. We have discussed the need of a rigorous mask optimization process for DRAM contact cell layout yielding mask layouts that are optimal in process performance, mask manufacturability and accuracy. In this paper, we have shown the step by step process from analytical illumination source derivation, a NTD and application tailored model calibration to layout optimization such as OPC and SRAF placement. Finally the work has been verified with simulation and experimental results on wafer.
Burrus, Barri B.; Scott, Alicia Richmond
2012-01-01
Adolescent parents and their children are at increased risk for adverse short- and long-term health and social outcomes. Effective interventions are needed to support these young families. We studied the evidence base and found a dearth of rigorously evaluated programs. Strategies from successful interventions are needed to inform both intervention design and policies affecting these adolescents. The lack of rigorous evaluations may be attributable to inadequate emphasis on and sufficient funding for evaluation, as well as to challenges encountered by program evaluators working with this population. More rigorous program evaluations are urgently needed to provide scientifically sound guidance for programming and policy decisions. Evaluation lessons learned have implications for other vulnerable populations. PMID:22897541
Manley, J.; Chegwidden, D.; Mote, A. S.; Ledley, T. S.; Lynds, S. E.; Haddad, N.; Ellins, K.
2016-02-01
EarthLabs, envisioned as a national model for high school Earth or Environmental Science lab courses, is adaptable for both undergraduate middle school students. The collection includes ten online modules that combine to feature a global view of our planet as a dynamic, interconnected system, by engaging learners in extended investigations. EarthLabs support state and national guidelines, including the NGSS, for science content. Four modules directly guide students to discover vital aspects of the oceans while five other modules incorporate ocean sciences in order to complete an understanding of Earth's climate system. Students gain a broad perspective on the key role oceans play in fishing industry, droughts, coral reefs, hurricanes, the carbon cycle, as well as life on land and in the seas to drive our changing climate by interacting with scientific research data, manipulating satellite imagery, numerical data, computer visualizations, experiments, and video tutorials. Students explore Earth system processes and build quantitative skills that enable them to objectively evaluate scientific findings for themselves as they move through ordered sequences that guide the learning. As a robust collection, EarthLabs modules engage students in extended, rigorous investigations allowing a deeper understanding of the ocean, climate and weather. This presentation provides an overview of the ten curriculum modules that comprise the EarthLabs collection developed by TERC and found at http://serc.carleton.edu/earthlabs/index.html. Evaluation data on the effectiveness and use in secondary education classrooms will be summarized.
Directory of Open Access Journals (Sweden)
Surendra P. Singh
2015-11-01
Full Text Available The Himalaya range encompasses enormous variation in elevation, precipitation, biodiversity, and patterns of human livelihoods. These mountains modify the regional climate in complex ways; the ecosystem services they provide influence the lives of almost 1 billion people in 8 countries. However, our understanding of these ecosystems remains rudimentary. The 2007 Intergovernmental Panel on Climate Change report that erroneously predicted a date for widespread glacier loss exposed how little was known of Himalayan glaciers. Recent research shows how variably glaciers respond to climate change in different Himalayan regions. Alarmist theories are not new. In the 1980s, the Theory of Himalayan Degradation warned of complete forest loss and devastation of downstream areas, an eventuality that never occurred. More recently, the debate on hydroelectric construction appears driven by passions rather than science. Poor data, hasty conclusions, and bad science plague Himalayan research. Rigorous sampling, involvement of civil society in data collection, and long-term collaborative research involving institutions from across the Himalaya are essential to improve knowledge of this region.
Hidayat, D.; Nurlaelah, E.; Dahlan, J. A.
2017-09-01
The ability of mathematical creative and critical thinking are two abilities that need to be developed in the learning of mathematics. Therefore, efforts need to be made in the design of learning that is capable of developing both capabilities. The purpose of this research is to examine the mathematical creative and critical thinking ability of students who get rigorous mathematical thinking (RMT) approach and students who get expository approach. This research was quasi experiment with control group pretest-posttest design. The population were all of students grade 11th in one of the senior high school in Bandung. The result showed that: the achievement of mathematical creative and critical thinking abilities of student who obtain RMT is better than students who obtain expository approach. The use of Psychological tools and mediation with criteria of intentionality, reciprocity, and mediated of meaning on RMT helps students in developing condition in critical and creative processes. This achievement contributes to the development of integrated learning design on students’ critical and creative thinking processes.
A Rigorous Theory of Many-Body Prethermalization for Periodically Driven and Closed Quantum Systems
Abanin, Dmitry; De Roeck, Wojciech; Ho, Wen Wei; Huveneers, François
2017-09-01
Prethermalization refers to the transient phenomenon where a system thermalizes according to a Hamiltonian that is not the generator of its evolution. We provide here a rigorous framework for quantum spin systems where prethermalization is exhibited for very long times. First, we consider quantum spin systems under periodic driving at high frequency {ν}. We prove that up to a quasi-exponential time {τ_* ˜ e^{c ν/log^3 ν}}, the system barely absorbs energy. Instead, there is an effective local Hamiltonian {\\widehat D} that governs the time evolution up to {τ_*}, and hence this effective Hamiltonian is a conserved quantity up to {τ_*}. Next, we consider systems without driving, but with a separation of energy scales in the Hamiltonian. A prime example is the Fermi-Hubbard model where the interaction U is much larger than the hopping J. Also here we prove the emergence of an effective conserved quantity, different from the Hamiltonian, up to a time {τ_*} that is (almost) exponential in {U/J}.
Coupling of Rigor Mortis and Intestinal Necrosis during C. elegans Organismal Death
Directory of Open Access Journals (Sweden)
Evgeniy R. Galimov
2018-03-01
Full Text Available Organismal death is a process of systemic collapse whose mechanisms are less well understood than those of cell death. We previously reported that death in C. elegans is accompanied by a calcium-propagated wave of intestinal necrosis, marked by a wave of blue autofluorescence (death fluorescence. Here, we describe another feature of organismal death, a wave of body wall muscle contraction, or death contraction (DC. This phenomenon is accompanied by a wave of intramuscular Ca2+ release and, subsequently, of intestinal necrosis. Correlation of directions of the DC and intestinal necrosis waves implies coupling of these death processes. Long-lived insulin/IGF-1-signaling mutants show reduced DC and delayed intestinal necrosis, suggesting possible resistance to organismal death. DC resembles mammalian rigor mortis, a postmortem necrosis-related process in which Ca2+ influx promotes muscle hyper-contraction. In contrast to mammals, DC is an early rather than a late event in C. elegans organismal death.
Inosine-5'-monophosphate is a candidate agent to resolve rigor mortis of skeletal muscle.
Matsuishi, Masanori; Tsuji, Mariko; Yamaguchi, Megumi; Kitamura, Natsumi; Tanaka, Sachi; Nakamura, Yukinobu; Okitani, Akihiro
2016-11-01
The object of the present study was to reveal the action of inosine-5'-monophosphate (IMP) toward myofibrils in postmortem muscles. IMP solubilized isolated actomyosin within a narrow range of KCl concentration, 0.19-0.20 mol/L, because of the dissociation of actomyosin into actin and myosin, but it did not solubilize the proteins in myofibrils with 0.2 mol/L KCl. However, IMP could solubilize both proteins in myofibrils with 0.2 mol/L KCl in the presence of 1 m mol/L pyrophosphate or 1.0-3.3 m mol/L adenosine-5'-diphosphate (ADP). Thus, we presumed that pyrophosphate and ADP released thin filaments composed of actin, and thick filaments composed of myosin from restraints of myofibrils, and then both filaments were solubilized through the IMP-induced dissociation of actomyosin. Thus, we concluded that IMP is a candidate agent to resolve rigor mortis because of its ability to break the association between thick and thin filaments. © 2016 Japanese Society of Animal Science.
Alternative pre-rigor foreshank positioning can improve beef shoulder muscle tenderness.
Grayson, A L; Lawrence, T E
2013-09-01
Thirty beef carcasses were harvested and the foreshank of each side was independently positioned (cranial, natural, parallel, or caudal) 1h post-mortem to determine the effect of foreshank angle at rigor mortis on the sarcomere length and tenderness of six beef shoulder muscles. The infraspinatus (IS), pectoralis profundus (PP), serratus ventralis (SV), supraspinatus (SS), teres major (TM) and triceps brachii (TB) were excised 48 h post-mortem for Warner-Bratzler shear force (WBSF) and sarcomere length evaluations. All muscles except the SS had altered (P<0.05) sarcomere lengths between positions; the cranial position resulted in the longest sarcomeres for the SV and TB muscles whilst the natural position had longer sarcomeres for the PP and TM muscles. The SV from the cranial position had lower (P<0.05) shear than the caudal position and TB from the natural position had lower (P<0.05) shear than the parallel or caudal positions. Sarcomere length was moderately correlated (r=-0.63; P<0.01) to shear force. Copyright © 2013 Elsevier Ltd. All rights reserved.
Coupling of Rigor Mortis and Intestinal Necrosis during C. elegans Organismal Death.
Galimov, Evgeniy R; Pryor, Rosina E; Poole, Sarah E; Benedetto, Alexandre; Pincus, Zachary; Gems, David
2018-03-06
Organismal death is a process of systemic collapse whose mechanisms are less well understood than those of cell death. We previously reported that death in C. elegans is accompanied by a calcium-propagated wave of intestinal necrosis, marked by a wave of blue autofluorescence (death fluorescence). Here, we describe another feature of organismal death, a wave of body wall muscle contraction, or death contraction (DC). This phenomenon is accompanied by a wave of intramuscular Ca 2+ release and, subsequently, of intestinal necrosis. Correlation of directions of the DC and intestinal necrosis waves implies coupling of these death processes. Long-lived insulin/IGF-1-signaling mutants show reduced DC and delayed intestinal necrosis, suggesting possible resistance to organismal death. DC resembles mammalian rigor mortis, a postmortem necrosis-related process in which Ca 2+ influx promotes muscle hyper-contraction. In contrast to mammals, DC is an early rather than a late event in C. elegans organismal death. VIDEO ABSTRACT. Copyright © 2018 The Author(s). Published by Elsevier Inc. All rights reserved.
Rigor mortis and the epileptology of Charles Bland Radcliffe (1822-1889).
Eadie, M J
2007-03-01
Charles Bland Radcliffe (1822-1889) was one of the physicians who made major contributions to the literature on epilepsy in the mid-19th century, when the modern understanding of the disorder was beginning to emerge, particularly in England. His experimental work was concerned with the electrical properties of frog muscle and nerve. Early in his career he related his experimental findings to the phenomenon of rigor mortis and concluded that, contrary to the general belief of the time, muscle contraction depended on the cessation of nerve input, and muscle relaxation on its presence. He adhered to this counter-intuitive interpretation throughout his life and, based on it, produced an epileptology that was very different from those of his contemporaries and successors. His interpretations were ultimately without any direct influence on the advance of knowledge. However, his idea that withdrawal of an inhibitory process released previously suppressed muscular contractile powers, when applied to the brain rather than the periphery of the nervous system, permitted Hughlings Jackson to explain certain psychological phenomena that accompany or follow some epileptic events. As well, Radcliffe was one of the chief early advocates for potassium bromide, the first effective anticonvulsant.
Control group design: enhancing rigor in research of mind-body therapies for depression.
Kinser, Patricia Anne; Robins, Jo Lynne
2013-01-01
Although a growing body of research suggests that mind-body therapies may be appropriate to integrate into the treatment of depression, studies consistently lack methodological sophistication particularly in the area of control groups. In order to better understand the relationship between control group selection and methodological rigor, we provide a brief review of the literature on control group design in yoga and tai chi studies for depression, and we discuss challenges we have faced in the design of control groups for our recent clinical trials of these mind-body complementary therapies for women with depression. To address the multiple challenges of research about mind-body therapies, we suggest that researchers should consider 4 key questions: whether the study design matches the research question; whether the control group addresses performance, expectation, and detection bias; whether the control group is ethical, feasible, and attractive; and whether the control group is designed to adequately control for nonspecific intervention effects. Based on these questions, we provide specific recommendations about control group design with the goal of minimizing bias and maximizing validity in future research.
Directory of Open Access Journals (Sweden)
MIHAI NOVAC
2012-05-01
Full Text Available According to many of its traditional critics, the main weakness in Kantian moral-political philosophy resides in its impossibility of admitting exceptions. In nuce, all these critical positions have converged, despite their reciprocal heterogeneity, in the so called accuse of moral rigorism (unjustly, I would say directed against Kant’s moral and political perspective. As such, basically, I will seek to defend Kant against this type of criticism, by showing that any perspective attempting to evaluate Kant’s ethics on the grounds of its capacity or incapacity to admit exceptions is apriorily doomed to lack of sense, in its two logical alternatives, i.e. either as nonsense (predicating about empty notions, or as tautology (formulating ad hoc definitions and criteria with respect to Kant’s system and then claiming that it does not hold with respect to them. Essentially, I will try to show that Kantian ethics can organically immunize itself epistemologically against any such so called antirigorist criticism.
Energy Technology Data Exchange (ETDEWEB)
Maksymov, Ivan S., E-mail: ivan.maksymov@uwa.edu.au [School of Physics M013, The University of Western Australia, 35 Stirling Highway, Crawley, WA 6009 (Australia); ARC Centre of Excellence for Nanoscale BioPhotonics, School of Applied Sciences, RMIT University, Melbourne, VIC 3001 (Australia); Hutomo, Jessica; Nam, Donghee; Kostylev, Mikhail [School of Physics M013, The University of Western Australia, 35 Stirling Highway, Crawley, WA 6009 (Australia)
2015-05-21
We demonstrate theoretically a ∼350-fold local enhancement of the intensity of the in-plane microwave magnetic field in multilayered structures made from a magneto-insulating yttrium iron garnet (YIG) layer sandwiched between two non-magnetic layers with a high dielectric constant matching that of YIG. The enhancement is predicted for the excitation regime when the microwave magnetic field is induced inside the multilayer by the transducer of a stripline Broadband Ferromagnetic Resonance (BFMR) setup. By means of a rigorous numerical solution of the Landau-Lifshitz-Gilbert equation consistently with the Maxwell's equations, we investigate the magnetisation dynamics in the multilayer. We reveal a strong photon-magnon coupling, which manifests itself as anti-crossing of the ferromagnetic resonance magnon mode supported by the YIG layer and the electromagnetic resonance mode supported by the whole multilayered structure. The frequency of the magnon mode depends on the external static magnetic field, which in our case is applied tangentially to the multilayer in the direction perpendicular to the microwave magnetic field induced by the stripline of the BFMR setup. The frequency of the electromagnetic mode is independent of the static magnetic field. Consequently, the predicted photon-magnon coupling is sensitive to the applied magnetic field and thus can be used in magnetically tuneable metamaterials based on simultaneously negative permittivity and permeability achievable thanks to the YIG layer. We also suggest that the predicted photon-magnon coupling may find applications in microwave quantum information systems.
Rigorous constraints on the matrix elements of the energy–momentum tensor
Directory of Open Access Journals (Sweden)
Peter Lowdon
2017-11-01
Full Text Available The structure of the matrix elements of the energy–momentum tensor play an important role in determining the properties of the form factors A(q2, B(q2 and C(q2 which appear in the Lorentz covariant decomposition of the matrix elements. In this paper we apply a rigorous frame-independent distributional-matching approach to the matrix elements of the Poincaré generators in order to derive constraints on these form factors as q→0. In contrast to the literature, we explicitly demonstrate that the vanishing of the anomalous gravitomagnetic moment B(0 and the condition A(0=1 are independent of one another, and that these constraints are not related to the specific properties or conservation of the individual Poincaré generators themselves, but are in fact a consequence of the physical on-shell requirement of the states in the matrix elements and the manner in which these states transform under Poincaré transformations.
International Nuclear Information System (INIS)
Calahorra, Yonatan; Mendels, Dan; Epstein, Ariel
2014-01-01
Bounded geometries introduce a fundamental problem in calculating the image force barrier lowering of metal-wrapped semiconductor systems. In bounded geometries, the derivation of the barrier lowering requires calculating the reference energy of the system, when the charge is at the geometry center. In the following, we formulate and rigorously solve this problem; this allows combining the image force electrostatic potential with the band diagram of the bounded geometry. The suggested approach is applied to spheres as well as cylinders. Furthermore, although the expressions governing cylindrical systems are complex and can only be evaluated numerically, we present analytical approximations for the solution, which allow easy implementation in calculated band diagrams. The results are further used to calculate the image force barrier lowering of metal-wrapped cylindrical nanowires; calculations show that although the image force potential is stronger than that of planar systems, taking the complete band-structure into account results in a weaker effect of barrier lowering. Moreover, when considering small diameter nanowires, we find that the electrostatic effects of the image force exceed the barrier region, and influence the electronic properties of the nanowire core. This study is of interest to the nanowire community, and in particular for the analysis of nanowire I−V measurements where wrapped or omega-shaped metallic contacts are used. (paper)
Rigorous Statistical Bounds in Uncertainty Quantification for One-Layer Turbulent Geophysical Flows
Qi, Di; Majda, Andrew J.
2018-04-01
Statistical bounds controlling the total fluctuations in mean and variance about a basic steady-state solution are developed for the truncated barotropic flow over topography. Statistical ensemble prediction is an important topic in weather and climate research. Here, the evolution of an ensemble of trajectories is considered using statistical instability analysis and is compared and contrasted with the classical deterministic instability for the growth of perturbations in one pointwise trajectory. The maximum growth of the total statistics in fluctuations is derived relying on the statistical conservation principle of the pseudo-energy. The saturation bound of the statistical mean fluctuation and variance in the unstable regimes with non-positive-definite pseudo-energy is achieved by linking with a class of stable reference states and minimizing the stable statistical energy. Two cases with dependence on initial statistical uncertainty and on external forcing and dissipation are compared and unified under a consistent statistical stability framework. The flow structures and statistical stability bounds are illustrated and verified by numerical simulations among a wide range of dynamical regimes, where subtle transient statistical instability exists in general with positive short-time exponential growth in the covariance even when the pseudo-energy is positive-definite. Among the various scenarios in this paper, there exist strong forward and backward energy exchanges between different scales which are estimated by the rigorous statistical bounds.
Directory of Open Access Journals (Sweden)
Spiros Pagiatakis
2009-10-01
Full Text Available In this paper, we examine the effect of changing the temperature points on MEMS-based inertial sensor random error. We collect static data under different temperature points using a MEMS-based inertial sensor mounted inside a thermal chamber. Rigorous stochastic models, namely Autoregressive-based Gauss-Markov (AR-based GM models are developed to describe the random error behaviour. The proposed AR-based GM model is initially applied to short stationary inertial data to develop the stochastic model parameters (correlation times. It is shown that the stochastic model parameters of a MEMS-based inertial unit, namely the ADIS16364, are temperature dependent. In addition, field kinematic test data collected at about 17 °C are used to test the performance of the stochastic models at different temperature points in the filtering stage using Unscented Kalman Filter (UKF. It is shown that the stochastic model developed at 20 °C provides a more accurate inertial navigation solution than the ones obtained from the stochastic models developed at −40 °C, −20 °C, 0 °C, +40 °C, and +60 °C. The temperature dependence of the stochastic model is significant and should be considered at all times to obtain optimal navigation solution for MEMS-based INS/GPS integration.
El-Diasty, Mohammed; Pagiatakis, Spiros
2009-01-01
In this paper, we examine the effect of changing the temperature points on MEMS-based inertial sensor random error. We collect static data under different temperature points using a MEMS-based inertial sensor mounted inside a thermal chamber. Rigorous stochastic models, namely Autoregressive-based Gauss-Markov (AR-based GM) models are developed to describe the random error behaviour. The proposed AR-based GM model is initially applied to short stationary inertial data to develop the stochastic model parameters (correlation times). It is shown that the stochastic model parameters of a MEMS-based inertial unit, namely the ADIS16364, are temperature dependent. In addition, field kinematic test data collected at about 17 °C are used to test the performance of the stochastic models at different temperature points in the filtering stage using Unscented Kalman Filter (UKF). It is shown that the stochastic model developed at 20 °C provides a more accurate inertial navigation solution than the ones obtained from the stochastic models developed at -40 °C, -20 °C, 0 °C, +40 °C, and +60 °C. The temperature dependence of the stochastic model is significant and should be considered at all times to obtain optimal navigation solution for MEMS-based INS/GPS integration.
K. Di; Y. Liu; B. Liu; M. Peng
2012-01-01
Chang'E-1(CE-1) and Chang'E-2(CE-2) are the two lunar orbiters of China's lunar exploration program. Topographic mapping using CE-1 and CE-2 images is of great importance for scientific research as well as for preparation of landing and surface operation of Chang'E-3 lunar rover. In this research, we developed rigorous sensor models of CE-1 and CE-2 CCD cameras based on push-broom imaging principle with interior and exterior orientation parameters. Based on the rigorous sensor model, the 3D c...
Energy Technology Data Exchange (ETDEWEB)
Ahn, Hye-Kyung; Kim, Byoung Chan; Jun, Seung-Hyun; Chang, Mun Seock; Lopez-Ferrer, Daniel; Smith, Richard D.; Gu, Man Bock; Lee, Sang-Won; Kim, Beom S.; Kim, Jungbae
2010-12-15
An efficient protein digestion in proteomic analysis requires the stabilization of proteases such as trypsin. In the present work, trypsin was stabilized in the form of enzyme coating on electrospun polymer nanofibers (EC-TR), which crosslinks additional trypsin molecules onto covalently-attached trypsin (CA-TR). EC-TR showed better stability than CA-TR in rigorous conditions, such as at high temperatures of 40 °C and 50 °C, in the presence of organic co-solvents, and at various pH's. For example, the half-lives of CA-TR and EC-TR were 0.24 and 163.20 hours at 40 ºC, respectively. The improved stability of EC-TR can be explained by covalent-linkages on the surface of trypsin molecules, which effectively inhibits the denaturation, autolysis, and leaching of trypsin. The protein digestion was performed at 40 °C by using both CA-TR and EC-TR in digesting a model protein, enolase. EC-TR showed better performance and stability than CA-TR by maintaining good performance of enolase digestion under recycled uses for a period of one week. In the same condition, CA-TR showed poor performance from the beginning, and could not be used for digestion at all after a few usages. The enzyme coating approach is anticipated to be successfully employed not only for protein digestion in proteomic analysis, but also for various other fields where the poor enzyme stability presently hampers the practical applications of enzymes.
Rigorous construction and Hadamard property of the Unruh state in Schwarzschild spacetime
International Nuclear Information System (INIS)
Dappiaggi, Claudio; Pinamonti, Nicola
2009-07-01
The discovery of the radiation properties of black holes prompted the search for a natural candidate quantum ground state for a massless scalar field theory on Schwarzschild spacetime, here considered in the Eddington-Finkelstein representation. Among the several available proposals in the literature, an important physical role is played by the so-called Unruh state which is supposed to be appropriate to capture the physics of a black hole formed by spherically symmetric collapsing matter. Within this respect, we shall consider a massless Klein-Gordon field and we shall rigorously and globally construct such state, that is on the algebra of Weyl observables localised in the union of the static external region, the future event horizon and the non-static black hole region. Eventually, out of a careful use of microlocal techniques, we prove that the built state fulfils, where defined, the so-called Hadamard condition; hence, it is perturbatively stable, in other words realizing the natural candidate with which one could study purely quantum phenomena such as the role of the back reaction of Hawking's radiation. From a geometrical point of view, we shall make a profitable use of a bulk-to-boundary reconstruction technique which carefully exploits the Killing horizon structure as well as the conformal asymptotic behaviour of the underlying background. From an analytical point of view, our tools will range from Hoermander's theorem on propagation of singularities, results on the role of passive states, and a detailed use of the recently discovered peeling behaviour of the solutions of the wave equation in Schwarzschild spacetime. (orig.)
Rigor mortis at the myocardium investigated by post-mortem magnetic resonance imaging.
Bonzon, Jérôme; Schön, Corinna A; Schwendener, Nicole; Zech, Wolf-Dieter; Kara, Levent; Persson, Anders; Jackowski, Christian
2015-12-01
Post-mortem cardiac MR exams present with different contraction appearances of the left ventricle in cardiac short axis images. It was hypothesized that the grade of post-mortem contraction may be related to the post-mortem interval (PMI) or cause of death and a phenomenon caused by internal rigor mortis that may give further insights in the circumstances of death. The cardiac contraction grade was investigated in 71 post-mortem cardiac MR exams (mean age at death 52 y, range 12-89 y; 48 males, 23 females). In cardiac short axis images the left ventricular lumen volume as well as the left ventricular myocardial volume were assessed by manual segmentation. The quotient of both (LVQ) represents the grade of myocardial contraction. LVQ was correlated to the PMI, sex, age, cardiac weight, body mass and height, cause of death and pericardial tamponade when present. In cardiac causes of death a separate correlation was investigated for acute myocardial infarction cases and arrhythmic deaths. LVQ values ranged from 1.99 (maximum dilatation) to 42.91 (maximum contraction) with a mean of 15.13. LVQ decreased slightly with increasing PMI, however without significant correlation. Pericardial tamponade positively correlated with higher LVQ values. Variables such as sex, age, body mass and height, cardiac weight and cause of death did not correlate with LVQ values. There was no difference in LVQ values for myocardial infarction without tamponade and arrhythmic deaths. Based on the observation in our investigated cases, the phenomenon of post-mortem myocardial contraction cannot be explained by the influence of the investigated variables, except for pericardial tamponade cases. Further research addressing post-mortem myocardial contraction has to focus on other, less obvious factors, which may influence the early post-mortem phase too. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Rigorous Performance Evaluation of Smartphone GNSS/IMU Sensors for ITS Applications
Directory of Open Access Journals (Sweden)
Vassilis Gikas
2016-08-01
Full Text Available With the rapid growth in smartphone technologies and improvement in their navigation sensors, an increasing amount of location information is now available, opening the road to the provision of new Intelligent Transportation System (ITS services. Current smartphone devices embody miniaturized Global Navigation Satellite System (GNSS, Inertial Measurement Unit (IMU and other sensors capable of providing user position, velocity and attitude. However, it is hard to characterize their actual positioning and navigation performance capabilities due to the disparate sensor and software technologies adopted among manufacturers and the high influence of environmental conditions, and therefore, a unified certification process is missing. This paper presents the analysis results obtained from the assessment of two modern smartphones regarding their positioning accuracy (i.e., precision and trueness capabilities (i.e., potential and limitations based on a practical but rigorous methodological approach. Our investigation relies on the results of several vehicle tracking (i.e., cruising and maneuvering tests realized through comparing smartphone obtained trajectories and kinematic parameters to those derived using a high-end GNSS/IMU system and advanced filtering techniques. Performance testing is undertaken for the HTC One S (Android and iPhone 5s (iOS. Our findings indicate that the deviation of the smartphone locations from ground truth (trueness deteriorates by a factor of two in obscured environments compared to those derived in open sky conditions. Moreover, it appears that iPhone 5s produces relatively smaller and less dispersed error values compared to those computed for HTC One S. Also, the navigation solution of the HTC One S appears to adapt faster to changes in environmental conditions, suggesting a somewhat different data filtering approach for the iPhone 5s. Testing the accuracy of the accelerometer and gyroscope sensors for a number of
Rigorous construction and Hadamard property of the Unruh state in Schwarzschild spacetime
Energy Technology Data Exchange (ETDEWEB)
Dappiaggi, Claudio; Pinamonti, Nicola [Hamburg Univ. (Germany). II. Inst. fuer Theoretische Physik; Moretti, Valter [Trento Univ., Povo (Italy). Dipt. di Matematica; Istituto Nazionale di Fisica Nucleare, Povo (Italy); Istituto Nazionale di Alta Matematica ' ' F. Severi' ' , GNFM, Sesto Fiorentino (Italy)
2009-07-15
The discovery of the radiation properties of black holes prompted the search for a natural candidate quantum ground state for a massless scalar field theory on Schwarzschild spacetime, here considered in the Eddington-Finkelstein representation. Among the several available proposals in the literature, an important physical role is played by the so-called Unruh state which is supposed to be appropriate to capture the physics of a black hole formed by spherically symmetric collapsing matter. Within this respect, we shall consider a massless Klein-Gordon field and we shall rigorously and globally construct such state, that is on the algebra of Weyl observables localised in the union of the static external region, the future event horizon and the non-static black hole region. Eventually, out of a careful use of microlocal techniques, we prove that the built state fulfils, where defined, the so-called Hadamard condition; hence, it is perturbatively stable, in other words realizing the natural candidate with which one could study purely quantum phenomena such as the role of the back reaction of Hawking's radiation. From a geometrical point of view, we shall make a profitable use of a bulk-to-boundary reconstruction technique which carefully exploits the Killing horizon structure as well as the conformal asymptotic behaviour of the underlying background. From an analytical point of view, our tools will range from Hoermander's theorem on propagation of singularities, results on the role of passive states, and a detailed use of the recently discovered peeling behaviour of the solutions of the wave equation in Schwarzschild spacetime. (orig.)
International Nuclear Information System (INIS)
Piwowarska-Bilska, H.; Supinska, A.; Listewnik, M. H.; Zorga, P.; Birkenfeld, B.
2013-01-01
The appropriate radiation protection measures applied in departments of nuclear medicine should lead to a reduction in doses received by the employees. During 1991-2007, at the Department of Nuclear Medicine of Pomeranian Medical University (Szczecin, Poland), nurses received on average two-times higher (4.6 mSv) annual doses to the whole body than those received by radiopharmacy technicians. The purpose of this work was to examine whether implementation of changes in the radiation protection protocol will considerably influence the reduction in whole-body doses received by the staff that are the most exposed. A reduction in nurses' exposure by ∼63% took place in 2008-11, whereas the exposure of radiopharmacy technicians grew by no more than 22% in comparison with that in the period 1991-2007. Proper reorganisation of the work in departments of nuclear medicine can considerably affect dose reduction and bring about equal distribution of the exposure. (authors)
The seed of soybean (Glycine max L. Merr) is a valuable source of high quality edible oil and protein. Despite dramatic breeding gains over the past 80 years, soybean seed oil continues to be oxidatively unstable. Until recently, the majority of soybean oil underwent partial chemical hydrogenation. ...
Geldenhuys, Greta; Muller, Nina; Frylinck, Lorinda; Hoffman, Louwrens C
2016-01-15
Baseline research on the toughness of Egyptian goose meat is required. This study therefore investigates the post mortem pH and temperature decline (15 min-4 h 15 min post mortem) in the pectoralis muscle (breast portion) of this gamebird species. It also explores the enzyme activity of the Ca(2+)-dependent protease (calpain system) and the lysosomal cathepsins during the rigor mortis period. No differences were found for any of the variables between genders. The pH decline in the pectoralis muscle occurs quite rapidly (c = -0.806; ultimate pH ∼ 5.86) compared with other species and it is speculated that the high rigor temperature (>20 °C) may contribute to the increased toughness. No calpain I was found in Egyptian goose meat and the µ/m-calpain activity remained constant during the rigor period, while a decrease in calpastatin activity was observed. The cathepsin B, B & L and H activity increased over the rigor period. Further research into the connective tissue content and myofibrillar breakdown during aging is required in order to know if the proteolytic enzymes do in actual fact contribute to tenderisation. © 2015 Society of Chemical Industry.
Warren, Mark R.; Calderón, José; Kupscznk, Luke Aubry; Squires, Gregory; Su, Celina
2018-01-01
Contrary to the charge that advocacy-oriented research cannot meet social science research standards because it is inherently biased, the authors of this article argue that collaborative, community-engaged scholarship (CCES) must meet high standards of rigor if it is to be useful to support equity-oriented, social justice agendas. In fact, they…
Directory of Open Access Journals (Sweden)
Kun Hu
2016-09-01
Full Text Available High precision geometric rectification of High Resolution Satellite Imagery (HRSI is the basis of digital mapping and Three-Dimensional (3D modeling. Taking advantage of line features as basic geometric control conditions instead of control points, the Line-Based Transformation Model (LBTM provides a practical and efficient way of image rectification. It is competent to build the mathematical relationship between image space and the corresponding object space accurately, while it reduces the workloads of ground control and feature recognition dramatically. Based on generalization and the analysis of existing LBTMs, a novel rigorous LBTM is proposed in this paper, which can further eliminate the geometric deformation caused by sensor inclination and terrain variation. This improved nonlinear LBTM is constructed based on a generalized point strategy and resolved by least squares overall adjustment. Geo-positioning accuracy experiments with IKONOS, GeoEye-1 and ZiYuan-3 satellite imagery are performed to compare rigorous LBTM with other relevant line-based and point-based transformation models. Both theoretic analysis and experimental results demonstrate that the rigorous LBTM is more accurate and reliable without adding extra ground control. The geo-positioning accuracy of satellite imagery rectified by rigorous LBTM can reach about one pixel with eight control lines and can be further improved by optimizing the horizontal and vertical distribution of control lines.
Vaade, Elizabeth; McCready, Bo
2012-01-01
Traditionally, researchers, policymakers, and practitioners have perceived a tension between rigor and accessibility in quantitative research and evaluation in postsecondary education. However, this study indicates that both producers and consumers of these studies value high-quality work and clear findings that can reach multiple audiences. The…
Dahlquist, Germund
1974-01-01
""Substantial, detailed and rigorous . . . readers for whom the book is intended are admirably served."" - MathSciNet (Mathematical Reviews on the Web), American Mathematical Society.Practical text strikes fine balance between students' requirements for theoretical treatment and needs of practitioners, with best methods for large- and small-scale computing. Prerequisites are minimal (calculus, linear algebra, and preferably some acquaintance with computer programming). Text includes many worked examples, problems, and an extensive bibliography.
Vereecken, Luc; Peeters, Jozef
2003-09-01
The rigorous implementation of transition state theory (TST) for a reaction system with multiple reactant rotamers and multiple transition state conformers is discussed by way of a statistical rate analysis of the 1,5-H-shift in 1-butoxy radicals, a prototype reaction for the important class of H-shift reactions in atmospheric chemistry. Several approaches for deriving a multirotamer TST expression are treated: oscillator versus (hindered) internal rotor models; distinguishable versus indistinguishable atoms; and direct count methods versus degeneracy factors calculated by (simplified) direct count methods or from symmetry numbers and number of enantiomers, where applicable. It is shown that the various treatments are fully consistent, even if the TST expressions themselves appear different. The 1-butoxy H-shift reaction is characterized quantum chemically using B3LYP-DFT; the performance of this level of theory is compared to other methods. Rigorous application of the multirotamer TST methodology in an harmonic oscillator approximation based on this data yields a rate coefficient of k(298 K,1 atm)=1.4×105 s-1, and an Arrhenius expression k(T,1 atm)=1.43×1011 exp(-8.17 kcal mol-1/RT) s-1, which both closely match the experimental recommendations in the literature. The T-dependence is substantially influenced by the multirotamer treatment, as well as by the tunneling and fall-off corrections. The present results are compared to those of simplified TST calculations based solely on the properties of the lowest energy 1-butoxy rotamer.
Causality as a Rigorous Notion and Quantitative Causality Analysis with Time Series
Liang, X. S.
2017-12-01
Given two time series, can one faithfully tell, in a rigorous and quantitative way, the cause and effect between them? Here we show that this important and challenging question (one of the major challenges in the science of big data), which is of interest in a wide variety of disciplines, has a positive answer. Particularly, for linear systems, the maximal likelihood estimator of the causality from a series X2 to another series X1, written T2→1, turns out to be concise in form: T2→1 = [C11 C12 C2,d1 — C112 C1,d1] / [C112 C22 — C11C122] where Cij (i,j=1,2) is the sample covariance between Xi and Xj, and Ci,dj the covariance between Xi and ΔXj/Δt, the difference approximation of dXj/dt using the Euler forward scheme. An immediate corollary is that causation implies correlation, but not vice versa, resolving the long-standing debate over causation versus correlation. The above formula has been validated with touchstone series purportedly generated with one-way causality that evades the classical approaches such as Granger causality test and transfer entropy analysis. It has also been applied successfully to the investigation of many real problems. Through a simple analysis with the stock series of IBM and GE, an unusually strong one-way causality is identified from the former to the latter in their early era, revealing to us an old story, which has almost faded into oblivion, about "Seven Dwarfs" competing with a "Giant" for the computer market. Another example presented here regards the cause-effect relation between the two climate modes, El Niño and Indian Ocean Dipole (IOD). In general, these modes are mutually causal, but the causality is asymmetric. To El Niño, the information flowing from IOD manifests itself as a propagation of uncertainty from the Indian Ocean. In the third example, an unambiguous one-way causality is found between CO2 and the global mean temperature anomaly. While it is confirmed that CO2 indeed drives the recent global warming
Misimi, E; Erikson, U; Digre, H; Skavhaug, A; Mathiassen, J R
2008-03-01
The present study describes the possibilities for using computer vision-based methods for the detection and monitoring of transient 2D and 3D changes in the geometry of a given product. The rigor contractions of unstressed and stressed fillets of Atlantic salmon (Salmo salar) and Atlantic cod (Gadus morhua) were used as a model system. Gradual changes in fillet shape and size (area, length, width, and roundness) were recorded for 7 and 3 d, respectively. Also, changes in fillet area and height (cross-section profiles) were tracked using a laser beam and a 3D digital camera. Another goal was to compare rigor developments of the 2 species of farmed fish, and whether perimortem stress affected the appearance of the fillets. Some significant changes in fillet size and shape were found (length, width, area, roundness, height) between unstressed and stressed fish during the course of rigor mortis as well as after ice storage (postrigor). However, the observed irreversible stress-related changes were small and would hardly mean anything for postrigor fish processors or consumers. The cod were less stressed (as defined by muscle biochemistry) than the salmon after the 2 species had been subjected to similar stress bouts. Consequently, the difference between the rigor courses of unstressed and stressed fish was more extreme in the case of salmon. However, the maximal whole fish rigor strength was judged to be about the same for both species. Moreover, the reductions in fillet area and length, as well as the increases in width, were basically of similar magnitude for both species. In fact, the increases in fillet roundness and cross-section height were larger for the cod. We conclude that the computer vision method can be used effectively for automated monitoring of changes in 2D and 3D shape and size of fish fillets during rigor mortis and ice storage. In addition, it can be used for grading of fillets according to uniformity in size and shape, as well as measurement of
Vada-Kovács, M
1996-01-01
Porcine biceps femoris strips of 10 cm original length were stretched by 50% and fixed within 1 hr post mortem then subjected to temperatures of 4 °, 15 ° or 36 °C until they attained their ultimate pH. Unrestrained control muscle strips, which were left to shorten freely, were similarly treated. Post-mortem metabolism (pH, R-value) and shortening were recorded; thereafter ultimate meat quality traits (pH, lightness, extraction and swelling of myofibrils) were determined. The rate of pH fall at 36 °C, as well as ATP breakdown at 36 and 4 °C, were significantly reduced by pre-rigor stretch. The relationship between R-value and pH indicated cold shortening at 4 °C. Myofibrils isolated from pre-rigor stretched muscle strips kept at 36 °C showed the most severe reduction of hydration capacity, while paleness remained below extreme values. However, pre-rigor stretched myofibrils - when stored at 4 °C - proved to be superior to shortened ones in their extractability and swelling.
Kobayashi, M; Takatori, T; Iwadate, K; Nakajima, M
1996-10-25
We examined the changes in adenosine triphosphate (ATP), lactic acid, adenosine diphosphate (ADP) and adenosine monophosphate (AMP) in five different rat muscles after death. Rigor mortis has been thought to occur simultaneously in dead muscles and hence to start in small muscles sooner than in large muscles. In this study we found that the rate of decrease in ATP was significantly different in each muscle. The greatest drop in ATP was observed in the masseter muscle. These findings contradict the conventional theory of rigor mortis. Similarly, the rates of change in ADP and lactic acid, which are thought to be related to the consumption or production of ATP, were different in each muscle. However, the rate of change of AMP was the same in each muscle.
Authorization of Animal Experiments Is Based on Confidence Rather than Evidence of Scientific Rigor
Nathues, Christina; Würbel, Hanno
2016-01-01
animal experiments are lacking important information about experimental conduct that determines the scientific validity of the findings, which may be critical for the weight attributed to the benefit of the research in the harm–benefit analysis. Similar to manuscripts getting accepted for publication despite poor reporting of measures against bias, applications for animal experiments may often be approved based on implicit confidence rather than explicit evidence of scientific rigor. Our findings shed serious doubt on the current authorization procedure for animal experiments, as well as the peer-review process for scientific publications, which in the long run may undermine the credibility of research. Developing existing authorization procedures that are already in place in many countries towards a preregistration system for animal research is one promising way to reform the system. This would not only benefit the scientific validity of findings from animal experiments but also help to avoid unnecessary harm to animals for inconclusive research. PMID:27911892
Effects of well-boat transportation on the muscle pH and onset of rigor mortis in Atlantic salmon.
Gatica, M C; Monti, G; Gallo, C; Knowles, T G; Warriss, P D
2008-07-26
During the transport of salmon (Salmo salar), in a well-boat, 10 fish were sampled at each of six stages: in cages after crowding at the farm (stage 1), in the well-boat after loading (stage 2), in the well-boat after eight hours transport and before unloading (stage 3), in the resting cages immediately after finishing unloading (stage 4), after 24 hours resting in cages, (stage 5) and in the processing plant after pumping from the resting cages (stage 6). The water in the well-boat was at ambient temperature with recirculation to the sea. At each stage the fish were stunned percussively and bled by gill cutting. Immediately after death, and then every three hours for 18 hours, the muscle pH and rigor index of the fish were measured. At successive stages the initial muscle pH of the fish decreased, except for a slight gain in stage 5, after they had been rested for 24 hours. The lowest initial muscle pH was observed at stage 6. The fishes' rigor index showed that rigor developed more quickly at each successive stage, except for a slight decrease in rate at stage 5, attributable to the recovery of muscle reserves.
Guarino, Heidi; Yoder, Shaun
2015-01-01
"Seizing the Future: How Ohio's Career and Technical Education Programs Fuse Academic Rigor and Real-World Experiences to Prepare Students for College and Work," demonstrates Ohio's progress in developing strong policies for career and technical education (CTE) programs to promote rigor, including college- and career-ready graduation…
Directory of Open Access Journals (Sweden)
Ling Fiona W.M.
2017-01-01
Full Text Available Rapid prototyping of microchannel gain lots of attention from researchers along with the rapid development of microfluidic technology. The conventional methods carried few disadvantages such as high cost, time consuming, required high operating pressure and temperature and involve expertise in operating the equipment. In this work, new method adapting xurography method is introduced to replace the conventional method of fabrication of microchannels. The novelty in this study is replacing the adhesion film with clear plastic film which was used to cut the design of the microchannel as the material is more suitable for fabricating more complex microchannel design. The microchannel was then mold using polymethyldisiloxane (PDMS and bonded with a clean glass to produce a close microchannel. The microchannel produced had a clean edge indicating good master mold was produced using the cutting plotter and the bonding between the PDMS and glass was good where no leakage was observed. The materials used in this method is cheap and the total time consumed is less than 5 hours where this method is suitable for rapid prototyping of microchannel.
Directory of Open Access Journals (Sweden)
L. M. Kimball
2002-01-01
Full Text Available This paper presents an interior point algorithm to solve the multiperiod hydrothermal economic dispatch (HTED. The multiperiod HTED is a large scale nonlinear programming problem. Various optimization methods have been applied to the multiperiod HTED, but most neglect important network characteristics or require decomposition into thermal and hydro subproblems. The algorithm described here exploits the special bordered block diagonal structure and sparsity of the Newton system for the first order necessary conditions to result in a fast efficient algorithm that can account for all network aspects. Applying this new algorithm challenges a conventional method for the use of available hydro resources known as the peak shaving heuristic.
Rigorous upper bounds for fluid and plasma transport due to passive advection
International Nuclear Information System (INIS)
Krommes, J.A.; Smith, R.A.; Kim, C.B.
1987-07-01
The formulation of variational principles for transport due to passive advection is described. A detailed account of the work has been published elsewhere. In the present paper, the motivations, philosophy, and implications of the method are briefly discussed. 15 refs
Energy Technology Data Exchange (ETDEWEB)
Verhaart, Rene F.; Rijnen, Zef; Verduijn, Gerda M.; Paulides, Margarethus M. [Erasmus MC - Cancer Institute, Department of Radiation Oncology, Hyperthermia Unit, Rotterdam (Netherlands); Fortunati, Valerio; Walsum, Theo van; Veenland, Jifke F. [Erasmus MC, Departments of Medical Informatics and Radiology, Biomedical Imaging Group Rotterdam, Rotterdam (Netherlands)
2014-12-15
Hyperthermia treatment planning (HTP) is used in the head and neck region (H and N) for pretreatment optimization, decision making, and real-time HTP-guided adaptive application of hyperthermia. In current clinical practice, HTP is based on power-absorption predictions, but thermal dose-effect relationships advocate its extension to temperature predictions. Exploitation of temperature simulations requires region- and temperature-specific thermal tissue properties due to the strong thermoregulatory response of H and N tissues. The purpose of our work was to develop a technique for patient group-specific optimization of thermal tissue properties based on invasively measured temperatures, and to evaluate the accuracy achievable. Data from 17 treated patients were used to optimize the perfusion and thermal conductivity values for the Pennes bioheat equation-based thermal model. A leave-one-out approach was applied to accurately assess the difference between measured and simulated temperature (∇T). The improvement in ∇T for optimized thermal property values was assessed by comparison with the ∇T for values from the literature, i.e., baseline and under thermal stress. The optimized perfusion and conductivity values of tumor, muscle, and fat led to an improvement in simulation accuracy (∇T: 2.1 ± 1.2 C) compared with the accuracy for baseline (∇T: 12.7 ± 11.1 C) or thermal stress (∇T: 4.4 ± 3.5 C) property values. The presented technique leads to patient group-specific temperature property values that effectively improve simulation accuracy for the challenging H and N region, thereby making simulations an elegant addition to invasive measurements. The rigorous leave-one-out assessment indicates that improvements in accuracy are required to rely only on temperature-based HTP in the clinic. (orig.) [German] Die Hyperthermiebehandlungsplanung (HTP, ''hyperthermia treatment planning'') wird in der Kopf- und Halsregion zur Optimierung der
International Nuclear Information System (INIS)
Terhesiu, Dalia; Froyland, Gary
2008-01-01
It is well known that for different classes of transformations, including the class of piecewise C 2 expanding maps T : [0, 1] O, Ulam's method is an efficient way to numerically approximate the absolutely continuous invariant measure of T. We develop a new extension of Ulam's method and prove that this extension can be used for the numerical approximation of the Ruelle–Perron–Frobenius operator associated with T and the potential φ β = −β log |T | |, where β element of R. In particular, we prove that our extended Ulam's method is a powerful tool for computing the topological pressure P(T, φ β ) and the density of the equilibrium state
A rigorous approach to relativistic corrections of bound state energies for spin-1/2 particles
International Nuclear Information System (INIS)
Gesztesy, F.; Thaller, B.; Grosse, H.
1983-01-01
Under fairly general conditions on the interactions we prove holomorphy of the Dirac resolvent around its nonrelativistic limit. As a consequences, perturbation theory in terms of resolvents (instead of Hamiltonians) yields holomorphy of Dirac eigenvalues and eigenfunctions with respect to c - 1 and a new method of calculating relativistic corrections to bound state energies. Due to a formulation in an abstract setting our method is applicable in many different concrete situation. In particular our approach covers the case of the relavistic hydrogen atom in external electromagnetic fields. (Author)
Gatica, M C; Monti, G E; Knowles, T G; Gallo, C B
2010-01-09
Two systems for transporting live salmon (Salmo salar) were compared in terms of their effects on blood variables, muscle pH and rigor index: an 'open system' well-boat with recirculated sea water at 13.5 degrees C and a stocking density of 107 kg/m3 during an eight-hour journey, and a 'closed system' well-boat with water chilled from 16.7 to 2.1 degrees C and a stocking density of 243.7 kg/m3 during a seven-hour journey. Groups of 10 fish were sampled at each of four stages: in cages at the farm, in the well-boat after loading, in the well-boat after the journey and before unloading, and in the processing plant after they were pumped from the resting cages. At each sampling, the fish were stunned and bled by gill cutting. Blood samples were taken to measure lactate, osmolality, chloride, sodium, cortisol and glucose, and their muscle pH and rigor index were measured at death and three hours later. In the open system well-boat, the initial muscle pH of the fish decreased at each successive stage, and at the final stage they had a significantly lower initial muscle pH and more rapid onset of rigor than the fish transported on the closed system well-boat. At the final stage all the blood variables except glucose were significantly affected in the fish transported on both types of well-boat.
Underhill, Kristen
2013-10-01
The growing evidence base for biomedical HIV prevention interventions - such as oral pre-exposure prophylaxis, microbicides, male circumcision, treatment as prevention, and eventually prevention vaccines - has given rise to concerns about the ways in which users of these biomedical products may adjust their HIV risk behaviors based on the perception that they are prevented from infection. Known as risk compensation, this behavioral adjustment draws on the theory of "risk homeostasis," which has previously been applied to phenomena as diverse as Lyme disease vaccination, insurance mandates, and automobile safety. Little rigorous evidence exists to answer risk compensation concerns in the biomedical HIV prevention literature, in part because the field has not systematically evaluated the study designs available for testing these behaviors. The goals of this Commentary are to explain the origins of risk compensation behavior in risk homeostasis theory, to reframe risk compensation as a testable response to the perception of reduced risk, and to assess the methodological rigor and ethical justification of study designs aiming to isolate risk compensation responses. Although the most rigorous methodological designs for assessing risk compensation behavior may be unavailable due to ethical flaws, several strategies can help investigators identify potential risk compensation behavior during Phase II, Phase III, and Phase IV testing of new technologies. Where concerns arise regarding risk compensation behavior, empirical evidence about the incidence, types, and extent of these behavioral changes can illuminate opportunities to better support the users of new HIV prevention strategies. This Commentary concludes by suggesting a new way to conceptualize risk compensation behavior in the HIV prevention context. Copyright © 2013 Elsevier Ltd. All rights reserved.
Fight the power: the limits of empiricism and the costs of positivistic rigor.
Indick, William
2002-01-01
A summary of the influence of positivistic philosophy and empiricism on the field of psychology is followed by a critique of the empirical method. The dialectic process is advocated as an alternative method of inquiry. The main advantage of the dialectic method is that it is open to any logical argument, including empirical hypotheses, but unlike empiricism, it does not automatically reject arguments that are not based on observable data. Evolutionary and moral psychology are discussed as examples of important fields of study that could benefit from types of arguments that frequently do not conform to the empirical standards of systematic observation and falsifiability of hypotheses. A dialectic method is shown to be a suitable perspective for those fields of research, because it allows for logical arguments that are not empirical and because it fosters a functionalist perspective, which is indispensable for both evolutionary and moral theories. It is suggested that all psychologists may gain from adopting a dialectic approach, rather than restricting themselves to empirical arguments alone.
Towards a Rigorous Formulation of the Space Mapping Technique for Engineering Design
DEFF Research Database (Denmark)
Koziel, Slawek; Bandler, John W.; Madsen, Kaj
2005-01-01
This paper deals with the Space Mapping (SM) approach to engineering design optimization. We attempt here a theoretical justification of methods that have already proven efficient in solving practical problems, especially in the RF and microwave area. A formal definition of optimization algorithm...
A Rigorous Inter-Comparison of Ground-Level Ozone Predictions
Czech Academy of Sciences Publication Activity Database
Schlink, U.; Dorling, S.R.; Pelikán, Emil; Nunnari, G.; Cawley, C.; Junninen, H.; Greig, A.; Foxall, R.; Eben, Kryštof; Chatterton, T.; Vondráček, Jiří; Richter, M.; Dostál, Michal; Bertucco, L.; Kolehmainen, M.; Doyle, M.
2003-01-01
Roč. 37, - (2003), s. 3237-3253 ISSN 1352-2310 Grant - others:APPETISE(XE) IST-1999-11764 Institutional research plan: AV0Z1030915 Keywords : statistical methods * air polution * neural networks Subject RIV: BB - Applied Statistics, Operational Research Impact factor: 2.338, year: 2003
International Nuclear Information System (INIS)
Schwenk-Ferrero, A.
1986-11-01
GANTRAS is a system of codes for neutron transport calculations in which the anisotropy of elastic and inelastic (including (n,n'x)-reactions) scattering is fully taken into account. This is achieved by employing a rigorous method, so-called I * -method, to represent the scattering term of the transport equation and with the use of double-differential cross-sections for the description of the emission of secondary neutrons. The I * -method was incorporated into the conventional transport code ONETRAN. The ONETRAN subroutines were modified for the new purpose. An implementation of the updated version ANTRA1 was accomplished for plane and spherical geometry. ANTRA1 was included in GANTRAS and linked to another modules which prepare angle-dependent transfer matrices. The GANTRAS code consists of three modules: 1. The CROMIX code which calculates the macroscopic transfer matrices for mixtures on the base of microscopic nuclide-dependent data. 2. The ATP code which generates discretized angular transfer probabilities (i.e. discretizes the I * -function). 3. The ANTRA1 code to perform S N transport calculations in one-dimensional plane and spherical geometries. This structure of GANTRAS allows to accommodate the system to various transport problems. (orig.) [de
Concollato, Anna; Parisi, Giuliana; Masoero, Giorgio; Romvàri, Robert; Olsen, Rolf-Erik; Dalle Zotte, Antonella
2016-08-01
Methods of stunning used in salmon slaughter are still the subject of research. Fish quality can be influenced by pre-, ante- and post-mortem conditions, including handling before slaughter, slaughter methods and storage conditions. Carbon monoxide (CO) is known to improve colour stability in red muscle and to reduce microbial growth and lipid oxidation in live fish exposed to CO. Quality differences in Atlantic salmon, Salmo salar L., stunned by CO or percussion, were evaluated and compared by different techniques [near infrared reflectance spectroscopy (NIRS), electronic nose (EN), electronic tongue (ET)] and sensory analysis. Thawed samples, freeze-dried preparates and NIRS devices proved to be the most efficient combinations for discriminating the treatments applied to salmon, i.e. first the stunning methods adopted, then the back-prediction of the maximum time to reach rigor mortis and finally to correlate some sensory attributes. A trained panel found significant differences between control and CO-stunned salmon: reduced tactile crumbliness, reduced odour and aroma intensities, and reduced tenderness of CO-treated fillets. CO stunning reduced radiation absorbance in spectra of thawed and freeze-dried fillets, but not fillet samples stored in ethanol, where it may have interacted with myoglobin and myosin. The good results in a rapid discrimination of thawed samples detected by NIRS suggest suitable applications in the fish industry. CO treatment could mitigate sensory perception, but consumer tests are needed to confirm our findings. © 2015 Society of Chemical Industry. © 2015 Society of Chemical Industry.
Uchiyama, T
1974-01-01
Rigorous lower bounds are derived from axiomatic field theory, by invoking analyticity and unitarity of the S-matrix. The bounds are expressed in terms of the total cross section and the slope parameter, and are found to be compatible with CERN experimental pp scattering data. It is also shown that the calculated lower-bound values imply non-existence of zeros for -t
Directory of Open Access Journals (Sweden)
Wang Yong
2011-07-01
Full Text Available Abstract Background Systematic mutagenesis studies have shown that only a few interface residues termed hot spots contribute significantly to the binding free energy of protein-protein interactions. Therefore, hot spots prediction becomes increasingly important for well understanding the essence of proteins interactions and helping narrow down the search space for drug design. Currently many computational methods have been developed by proposing different features. However comparative assessment of these features and furthermore effective and accurate methods are still in pressing need. Results In this study, we first comprehensively collect the features to discriminate hot spots and non-hot spots and analyze their distributions. We find that hot spots have lower relASA and larger relative change in ASA, suggesting hot spots tend to be protected from bulk solvent. In addition, hot spots have more contacts including hydrogen bonds, salt bridges, and atomic contacts, which favor complexes formation. Interestingly, we find that conservation score and sequence entropy are not significantly different between hot spots and non-hot spots in Ab+ dataset (all complexes. While in Ab- dataset (antigen-antibody complexes are excluded, there are significant differences in two features between hot pots and non-hot spots. Secondly, we explore the predictive ability for each feature and the combinations of features by support vector machines (SVMs. The results indicate that sequence-based feature outperforms other combinations of features with reasonable accuracy, with a precision of 0.69, a recall of 0.68, an F1 score of 0.68, and an AUC of 0.68 on independent test set. Compared with other machine learning methods and two energy-based approaches, our approach achieves the best performance. Moreover, we demonstrate the applicability of our method to predict hot spots of two protein complexes. Conclusion Experimental results show that support vector machine
2011-01-01
Background Systematic mutagenesis studies have shown that only a few interface residues termed hot spots contribute significantly to the binding free energy of protein-protein interactions. Therefore, hot spots prediction becomes increasingly important for well understanding the essence of proteins interactions and helping narrow down the search space for drug design. Currently many computational methods have been developed by proposing different features. However comparative assessment of these features and furthermore effective and accurate methods are still in pressing need. Results In this study, we first comprehensively collect the features to discriminate hot spots and non-hot spots and analyze their distributions. We find that hot spots have lower relASA and larger relative change in ASA, suggesting hot spots tend to be protected from bulk solvent. In addition, hot spots have more contacts including hydrogen bonds, salt bridges, and atomic contacts, which favor complexes formation. Interestingly, we find that conservation score and sequence entropy are not significantly different between hot spots and non-hot spots in Ab+ dataset (all complexes). While in Ab- dataset (antigen-antibody complexes are excluded), there are significant differences in two features between hot pots and non-hot spots. Secondly, we explore the predictive ability for each feature and the combinations of features by support vector machines (SVMs). The results indicate that sequence-based feature outperforms other combinations of features with reasonable accuracy, with a precision of 0.69, a recall of 0.68, an F1 score of 0.68, and an AUC of 0.68 on independent test set. Compared with other machine learning methods and two energy-based approaches, our approach achieves the best performance. Moreover, we demonstrate the applicability of our method to predict hot spots of two protein complexes. Conclusion Experimental results show that support vector machine classifiers are quite
Vanderwee, M A; Humphrey, S M; Gavin, J B; Armiger, L C
1981-01-01
Transmural slices from the left anterior papillary muscle of dog hearts were maintained for 120 min in a moist atmosphere at 37 degrees C. At 15-min intervals tissue samples were taken for estimation of adenosine triphosphate (ATP) and glucose-6-phosphate (G6P) and for electron microscopic examination. At the same time the deformability under standard load of comparable regions of an adjacent slice of tissue was measured. ATP levels fell rapidly during the first 45 to 75 min after excision of the heart. During a subsequent further decline in ATP, the mean deformability of myocardium fell from 30 to 12% indicating the development of rigor mortis. Conversely, G6P levels increased during the first decline in adenosine triphosphate but remained relatively steady thereafter. Whereas many of the myocardial cells fixed after 5 min contracted on contact with glutaraldehyde, all cells examined after 15 to 40 min were relaxed. A progressive increase in the proportion of contracted cells was observed during the rapid increase in myocardial rigidity. During this late contraction the cells showed morphological evidence of irreversible injury. These findings suggest that ischaemic myocytes contract just before actin and myosin become strongly linked to maintain the state of rigor mortis.
International Nuclear Information System (INIS)
Kim, Chang-Bae; Krommes, J.A.
1988-08-01
The work of Krommes and Smith on rigorous upper bounds for the turbulent transport of a passively advected scalar [/ital Ann. Phys./ 177:246 (1987)] is extended in two directions: (1) For their ''reference model,'' improved upper bounds are obtained by utilizing more sophisticated two-time constraints which include the effects of cross-correlations up to fourth order. Numerical solutions of the model stochastic differential equation are also obtained; they show that the new bounds compare quite favorably with the exact results, even at large Reynolds and Kubo numbers. (2) The theory is extended to take account of a finite spatial autocorrelation length L/sub c/. As a reasonably generic example, the problem of particle transport due to statistically specified stochastic magnetic fields in a collisionless turbulent plasma is revisited. A bound is obtained which reduces for small L/sub c/ to the quasilinear limit and for large L/sub c/ to the strong turbulence limit, and which provides a reasonable and rigorous interpolation for intermediate values of L/sub c/. 18 refs., 6 figs
Cotar, Codina; Friesecke, Gero; Klüppelberg, Claudia
2018-06-01
We prove rigorously that the exact N-electron Hohenberg-Kohn density functional converges in the strongly interacting limit to the strictly correlated electrons (SCE) functional, and that the absolute value squared of the associated constrained search wavefunction tends weakly in the sense of probability measures to a minimizer of the multi-marginal optimal transport problem with Coulomb cost associated to the SCE functional. This extends our previous work for N = 2 ( Cotar etal. in Commun Pure Appl Math 66:548-599, 2013). The correct limit problem has been derived in the physics literature by Seidl (Phys Rev A 60 4387-4395, 1999) and Seidl, Gorigiorgi and Savin (Phys Rev A 75:042511 1-12, 2007); in these papers the lack of a rigorous proofwas pointed out.We also give amathematical counterexample to this type of result, by replacing the constraint of given one-body density—an infinite dimensional quadratic expression in the wavefunction—by an infinite-dimensional quadratic expression in the wavefunction and its gradient. Connections with the Lawrentiev phenomenon in the calculus of variations are indicated.
Virtue-based Approaches to Professional Ethics: a Plea for More Rigorous Use of Empirical Science
Georg Spielthenner
2017-01-01
Until recently, the method of professional ethics has been largely principle-based. But the failure of this approach to take into sufficient account the character of professionals has led to a revival of virtue ethics. The kind of professional virtue ethics that I am concerned with in this paper is teleological in that it relates the virtues of a profession to the ends of this profession. My aim is to show how empirical research can (in addition to philosophical inquiry) be used to develop vi...
DEFF Research Database (Denmark)
Frutiger, Jerome; Abildskov, Jens; Sin, Gürkan
) weighted-least-square regression. 3) Initialization of estimation by use of linear algebra providing a first guess. 4) Sequential parameter and simultaneous GC parameter by using of 4 different minimization algorithms. 5) Thorough uncertainty analysis: a) based on asymptotic approximation of parameter...... covariance matrix b) based on boot strap method. Providing 95%-confidence intervals of parameters and predicted property. 6) Performance statistics analysis and model application. The application of the methodology is shown for a new GC model built to predict lower flammability limit (LFL) for refrigerants...... their credibility and robustness in wider industrial and scientific applications....
Electromagnetic field computation by network methods
Felsen, Leopold B; Russer, Peter
2009-01-01
This monograph proposes a systematic and rigorous treatment of electromagnetic field representations in complex structures. The book presents new strong models by combining important computational methods. This is the last book of the late Leopold Felsen.
The Researchers' View of Scientific Rigor-Survey on the Conduct and Reporting of In Vivo Research.
Reichlin, Thomas S; Vogt, Lucile; Würbel, Hanno
2016-01-01
Reproducibility in animal research is alarmingly low, and a lack of scientific rigor has been proposed as a major cause. Systematic reviews found low reporting rates of measures against risks of bias (e.g., randomization, blinding), and a correlation between low reporting rates and overstated treatment effects. Reporting rates of measures against bias are thus used as a proxy measure for scientific rigor, and reporting guidelines (e.g., ARRIVE) have become a major weapon in the fight against risks of bias in animal research. Surprisingly, animal scientists have never been asked about their use of measures against risks of bias and how they report these in publications. Whether poor reporting reflects poor use of such measures, and whether reporting guidelines may effectively reduce risks of bias has therefore remained elusive. To address these questions, we asked in vivo researchers about their use and reporting of measures against risks of bias and examined how self-reports relate to reporting rates obtained through systematic reviews. An online survey was sent out to all registered in vivo researchers in Switzerland (N = 1891) and was complemented by personal interviews with five representative in vivo researchers to facilitate interpretation of the survey results. Return rate was 28% (N = 530), of which 302 participants (16%) returned fully completed questionnaires that were used for further analysis. According to the researchers' self-report, they use measures against risks of bias to a much greater extent than suggested by reporting rates obtained through systematic reviews. However, the researchers' self-reports are likely biased to some extent. Thus, although they claimed to be reporting measures against risks of bias at much lower rates than they claimed to be using these measures, the self-reported reporting rates were considerably higher than reporting rates found by systematic reviews. Furthermore, participants performed rather poorly when asked to
International Nuclear Information System (INIS)
Krommes, J.A.; Kim, C.
1990-01-01
The fundamental problem in the theory of turbulent transport is to find the flux Γ of a quantity such as heat. Methods based on statistical closures are mired in conceptual controversies and practical difficulties. However, it is possible to bound Γ by employing constraints derived rigorously from the equations of motion. Brief reviews of the general theory and its application to passive advection are given. Then, a detailed application is made to anomalous resistivity generated by self-consistent turbulence in a reversed-field pinch. A nonlinear variational principle for an upper bound on the turbulent electromotive force for fixed current is formulated from the magnetohydrodynamic equations in cylindrical geometry. Numerical solution of a case constrained solely by energy balance leads to a reasonable bound and nonlinear eigenfunctions that share intriguing features with experimental data: The dominant mode numbers appear to be correct, and field reversal is predicted at reasonable values of the pinch parameter. Although open questions remain, upon considering all bounding calculations to date it can be concluded, remarkably, that global energy balance constrains transport sufficiently so that bounds derived therefrom are not unreasonable and that bounding calculations are feasible even for involved practical problems. The potential of the method has hardly been tapped; it provides a fertile area for future research
International Nuclear Information System (INIS)
Krommes, J.A.; Kim, Chang-Bae
1990-06-01
The fundamental problem in the theory of turbulent transport is to find the flux Γ of a quantity such as heat. Methods based on statistical closures are mired in conceptual controversies and practical difficulties. However, it is possible to bound Γ by employing constraints derived rigorously from the equations of motion. Brief reviews of the general theory and its application to passive advection are given. Then, a detailed application is made to anomalous resistivity generated by self-consistent turbulence in a reversed-field pinch. A nonlinear variational principle for an upper bound on the turbulence electromotive force for fixed current is formulated from the magnetohydrodynamic equations in cylindrical geometry. Numerical solution of a case constrained solely by energy balance leads to a reasonable bound and nonlinear eigenfunctions that share intriguing features with experimental data: the dominant mode numbers appear to be correct, and field reversal is predicted at reasonable values of the pinch parameter. Although open questions remain upon considering all bounding calculations to date one can conclude, remarkably, that global energy balance constrains transport sufficiently so that bounds derived therefrom are not unreasonable and that bounding calculations are feasible even for involved practical problems. The potential of the method has hardly been tapped; it provides a fertile area for future research. 29 refs
Toward a Rigorous Definition of a Strength of Any Interaction Between Bader's Atomic Basins.
Ananyev, Ivan V; Karnoukhova, Valentina A; Dmitrienko, Artem O; Lyssenko, Konstantin A
2017-06-15
Strength of interaction between Bader's atomic basins, enclosed by zero-flux surfaces of electron distribution, was proposed to be a measure of elastic deformation of an interaction. The set containing 53 atomic aggregate and covering all range of interaction strength (from van der Waals interactions to triple covalent bonds) was calculated by DFT and perturbation theory methods. Further analysis was performed to seek correlations between various local quantities based on electron density and effective force constants of stretching diatomic vibrations. The linear trend between effective force constants and the potential energy density at the (3, -1) critical point of electron distribution was found. This correlation was improved by the integration of the potential energy density over an interbasin zero-flux surface of electron density. Simple mechanical explanation of established trends is presented. The correlations can be further used to at least semiquantitatively compare any pair of interactions between Bader's atomic basins.
Exactly solvable models: the way towards a rigorous treatment of phase transitions in finite systems
International Nuclear Information System (INIS)
Bugaev, K.A.
2007-01-01
The exact analytical solutions of a variety of statistical models recently obtained for finite systems are thoroughly discussed. Among them are a constrained version of the statistical multifragmentation model, the Bag Model of Gases and the Hills and Dales Model of surface partition. The finite volume analytical solutions of these models were obtained by a novel powerful mathematical method - the Laplace-Fourier transform. The Laplace-Fourier transform allows one to study the nuclear matter equation of state, the equation of state of hadronic and quark-gluon plasma and the surface entropy of large clusters on the same footing. A complete analysis of the isobaric partition singularities of these models is done for finite systems. The developed formalism allows one to exactly define the finite volume analogs of gaseous, liquid and mixed phases of these models from the first principles of statistical mechanics [ru
Cell culture medium improvement by rigorous shuffling of components using media blending.
Jordan, Martin; Voisard, Damien; Berthoud, Antoine; Tercier, Laetitia; Kleuser, Beate; Baer, Gianni; Broly, Hervé
2013-01-01
A novel high-throughput methodology for the simultaneous optimization of many cell culture media components is presented. The method is based on the media blending approach which has several advantages as it works with ready-to-use media. In particular it allows precise pH and osmolarity adjustments and eliminates the need of concentrated stock solutions, a frequent source of serious solubility issues. In addition, media blending easily generates a large number of new compositions providing a remarkable screening tool. However, media blending designs usually do not provide information on distinct factors or components that are causing the desired improvements. This paper addresses this last point by considering the concentration of individual medium components to fix the experimental design and for the interpretation of the results. The extended blending strategy was used to reshuffle the 20 amino acids in one round of experiments. A small set of 10 media was specifically designed to generate a large number of mixtures. 192 mixtures were then prepared by media blending and tested on a recombinant CHO cell line expressing a monoclonal antibody. A wide range of performances (titers and viable cell density) was achieved from the different mixtures with top titers significantly above our previous results seen with this cell line. In addition, information about major effects of key amino acids on cell densities and titers could be extracted from the experimental results. This demonstrates that the extended blending approach is a powerful experimental tool which allows systematic and simultaneous reshuffling of multiple medium components.
Directory of Open Access Journals (Sweden)
Alireza G. Kashani
2015-11-01
Full Text Available In addition to precise 3D coordinates, most light detection and ranging (LIDAR systems also record “intensity”, loosely defined as the strength of the backscattered echo for each measured point. To date, LIDAR intensity data have proven beneficial in a wide range of applications because they are related to surface parameters, such as reflectance. While numerous procedures have been introduced in the scientific literature, and even commercial software, to enhance the utility of intensity data through a variety of “normalization”, “correction”, or “calibration” techniques, the current situation is complicated by a lack of standardization, as well as confusing, inconsistent use of terminology. In this paper, we first provide an overview of basic principles of LIDAR intensity measurements and applications utilizing intensity information from terrestrial, airborne topographic, and airborne bathymetric LIDAR. Next, we review effective parameters on intensity measurements, basic theory, and current intensity processing methods. We define terminology adopted from the most commonly-used conventions based on a review of current literature. Finally, we identify topics in need of further research. Ultimately, the presented information helps lay the foundation for future standards and specifications for LIDAR radiometric calibration.
Virtue-based Approaches to Professional Ethics: a Plea for More Rigorous Use of Empirical Science
Directory of Open Access Journals (Sweden)
Georg Spielthenner
2017-08-01
Full Text Available Until recently, the method of professional ethics has been largely principle-based. But the failure of this approach to take into sufficient account the character of professionals has led to a revival of virtue ethics. The kind of professional virtue ethics that I am concerned with in this paper is teleological in that it relates the virtues of a profession to the ends of this profession. My aim is to show how empirical research can (in addition to philosophical inquiry be used to develop virtue-based accounts of professional ethics, and that such empirically well-informed approaches are more convincing than traditional kinds of professional virtue ethics. The paper is divided into four sections. In the first, I outline the structure of a teleological approach to virtue ethics. In Section 2, I show that empirical research can play an essential role in professional ethics by emphasizing the difference between conceptual and empirical matters. Section 3 demonstrates the relevance of virtues in professional life; and the last section is concerned with some meta-ethical issues that are raised by a teleological account of professional virtues.
Kashani, Alireza G; Olsen, Michael J; Parrish, Christopher E; Wilson, Nicholas
2015-11-06
In addition to precise 3D coordinates, most light detection and ranging (LIDAR) systems also record "intensity", loosely defined as the strength of the backscattered echo for each measured point. To date, LIDAR intensity data have proven beneficial in a wide range of applications because they are related to surface parameters, such as reflectance. While numerous procedures have been introduced in the scientific literature, and even commercial software, to enhance the utility of intensity data through a variety of "normalization", "correction", or "calibration" techniques, the current situation is complicated by a lack of standardization, as well as confusing, inconsistent use of terminology. In this paper, we first provide an overview of basic principles of LIDAR intensity measurements and applications utilizing intensity information from terrestrial, airborne topographic, and airborne bathymetric LIDAR. Next, we review effective parameters on intensity measurements, basic theory, and current intensity processing methods. We define terminology adopted from the most commonly-used conventions based on a review of current literature. Finally, we identify topics in need of further research. Ultimately, the presented information helps lay the foundation for future standards and specifications for LIDAR radiometric calibration.
Thermal Impact Assessment of Groundwater Heat Pumps (GWHPs: Rigorous vs. Simplified Models
Directory of Open Access Journals (Sweden)
Bruno Piga
2017-09-01
Full Text Available Groundwater Heat Pumps (GWHPs are increasingly adopted for air conditioning in urban areas, thus reducing CO2 emissions, and this growth needs to be managed to ensure the sustainability of the thermal alteration of aquifers. However, few studies have addressed the propagation of thermal plumes from open-loop geothermal systems from a long-term perspective. We provide a comprehensive sensitivity analysis, performed with numerical finite-element simulations, to assess how the size of the thermally affected zone is driven by hydrodynamic and thermal subsurface properties, the vadose zone and aquifer thickness, and plant setup. In particular, we focus the analysis on the length and width of thermal plumes, and on their time evolution. Numerical simulations are compared with two simplified methods, namely (i replacing the time-varying thermal load with its yearly average and (ii analytical formulae for advective heat transport in the aquifer. The former proves acceptable for the assessment of plume length, while the latter can be used to estimate the width of the thermally affected zone. The results highlight the strong influence of groundwater velocity on the plume size and, especially for its long-term evolution, of ground thermal properties and of subsurface geometrical parameters.
Rigorous home range estimation with movement data: a new autocorrelated kernel density estimator.
Fleming, C H; Fagan, W F; Mueller, T; Olson, K A; Leimgruber, P; Calabrese, J M
2015-05-01
Quantifying animals' home ranges is a key problem in ecology and has important conservation and wildlife management applications. Kernel density estimation (KDE) is a workhorse technique for range delineation problems that is both statistically efficient and nonparametric. KDE assumes that the data are independent and identically distributed (IID). However, animal tracking data, which are routinely used as inputs to KDEs, are inherently autocorrelated and violate this key assumption. As we demonstrate, using realistically autocorrelated data in conventional KDEs results in grossly underestimated home ranges. We further show that the performance of conventional KDEs actually degrades as data quality improves, because autocorrelation strength increases as movement paths become more finely resolved. To remedy these flaws with the traditional KDE method, we derive an autocorrelated KDE (AKDE) from first principles to use autocorrelated data, making it perfectly suited for movement data sets. We illustrate the vastly improved performance of AKDE using analytical arguments, relocation data from Mongolian gazelles, and simulations based upon the gazelle's observed movement process. By yielding better minimum area estimates for threatened wildlife populations, we believe that future widespread use of AKDE will have significant impact on ecology and conservation biology.
Olsson, U; Hertzman, C; Tornberg, E
1994-01-01
The course of rigor mortis, ageing and tenderness have been evaluated for two beef muscles, M. semimembranosus (SM) and M. longissimus dorsi (LD), when entering rigor at constant temperatures in the cold-shortening region (1, 4, 7 and 10°C). The influence of electrical stimulation (ES) was also examined. Post-mortem changes were registered by shortening and isometric tension and by following the decline of pH, ATP and creatine phosphate. The effect of ageing on tenderness was recorded by measuring shear-force (2, 8 and 15 days post mortem) and the sensory properties were assessed 15 days post mortem. It was found that shortening increased with decreasing temperature, resulting in decreased tenderness. Tenderness for LD, but not for SM, was improved by ES at 1 and 4°C, whereas ES did not give rise to any decrease in the degree of shortening during rigor mortis development. This suggests that ES influences tenderization more than it prevents cold-shortening. The samples with a pre-rigor mortis temperature of 1°C could not be tenderized, when stored up to 15 days, whereas this was the case for the muscles entering rigor mortis at the other higher temperatures. The results show that under the conditions used in this study, the course of rigor mortis is more important for the ultimate tenderness than the course of ageing. Copyright © 1994. Published by Elsevier Ltd.
Energy Technology Data Exchange (ETDEWEB)
Ngampitipan, Tritos, E-mail: tritos.ngampitipan@gmail.com [Faculty of Science, Chandrakasem Rajabhat University, Ratchadaphisek Road, Chatuchak, Bangkok 10900 (Thailand); Particle Physics Research Laboratory, Department of Physics, Faculty of Science, Chulalongkorn University, Phayathai Road, Patumwan, Bangkok 10330 (Thailand); Boonserm, Petarpa, E-mail: petarpa.boonserm@gmail.com [Department of Mathematics and Computer Science, Faculty of Science, Chulalongkorn University, Phayathai Road, Patumwan, Bangkok 10330 (Thailand); Chatrabhuti, Auttakit, E-mail: dma3ac2@gmail.com [Particle Physics Research Laboratory, Department of Physics, Faculty of Science, Chulalongkorn University, Phayathai Road, Patumwan, Bangkok 10330 (Thailand); Visser, Matt, E-mail: matt.visser@msor.vuw.ac.nz [School of Mathematics, Statistics, and Operations Research, Victoria University of Wellington, PO Box 600, Wellington 6140 (New Zealand)
2016-06-02
Hawking radiation is the evidence for the existence of black hole. What an observer can measure through Hawking radiation is the transmission probability. In the laboratory, miniature black holes can successfully be generated. The generated black holes are, most commonly, Myers-Perry black holes. In this paper, we will derive the rigorous bounds on the transmission probabilities for massless scalar fields of non-negative-angular-momentum modes emitted from a generated Myers-Perry black hole in six, seven, and eight dimensions. The results show that for low energy, the rigorous bounds increase with the increase in the energy of emitted particles. However, for high energy, the rigorous bounds decrease with the increase in the energy of emitted particles. When the black holes spin faster, the rigorous bounds decrease. For dimension dependence, the rigorous bounds also decrease with the increase in the number of extra dimensions. Furthermore, as comparison to the approximate transmission probability, the rigorous bound is proven to be useful.
International Nuclear Information System (INIS)
Ngampitipan, Tritos; Boonserm, Petarpa; Chatrabhuti, Auttakit; Visser, Matt
2016-01-01
Hawking radiation is the evidence for the existence of black hole. What an observer can measure through Hawking radiation is the transmission probability. In the laboratory, miniature black holes can successfully be generated. The generated black holes are, most commonly, Myers-Perry black holes. In this paper, we will derive the rigorous bounds on the transmission probabilities for massless scalar fields of non-negative-angular-momentum modes emitted from a generated Myers-Perry black hole in six, seven, and eight dimensions. The results show that for low energy, the rigorous bounds increase with the increase in the energy of emitted particles. However, for high energy, the rigorous bounds decrease with the increase in the energy of emitted particles. When the black holes spin faster, the rigorous bounds decrease. For dimension dependence, the rigorous bounds also decrease with the increase in the number of extra dimensions. Furthermore, as comparison to the approximate transmission probability, the rigorous bound is proven to be useful.
Al-Durgham, Kaleel; Lichti, Derek D.; Kuntze, Gregor; Ronsky, Janet
2017-06-01
High-speed biplanar videoradiography, or clinically referred to as dual fluoroscopy (DF), imaging systems are being used increasingly for skeletal kinematics analysis. Typically, a DF system comprises two X-ray sources, two image intensifiers and two high-speed video cameras. The combination of these elements provides time-series image pairs of articulating bones of a joint, which permits the measurement of bony rotation and translation in 3D at high temporal resolution (e.g., 120-250 Hz). Assessment of the accuracy of 3D measurements derived from DF imaging has been the subject of recent research efforts by several groups, however with methodological limitations. This paper presents a novel and simple accuracy assessment procedure based on using precise photogrammetric tools. We address the fundamental photogrammetry principles for the accuracy evaluation of an imaging system. Bundle adjustment with selfcalibration is used for the estimation of the system parameters. The bundle adjustment calibration uses an appropriate sensor model and applies free-network constraints and relative orientation stability constraints for a precise estimation of the system parameters. A photogrammetric intersection of time-series image pairs is used for the 3D reconstruction of a rotating planar object. A point-based registration method is used to combine the 3D coordinates from the intersection and independently surveyed coordinates. The final DF accuracy measure is reported as the distance between 3D coordinates from image intersection and the independently surveyed coordinates. The accuracy assessment procedure is designed to evaluate the accuracy over the full DF image format and a wide range of object rotation. Experiment of reconstruction of a rotating planar object reported an average positional error of 0.44 +/- 0.2 mm in the derived 3D coordinates (minimum 0.05 and maximum 1.2 mm).
McCaig, Chris; Begon, Mike; Norman, Rachel; Shankland, Carron
2011-03-01
Changing scale, for example, the ability to move seamlessly from an individual-based model to a population-based model, is an important problem in many fields. In this paper, we introduce process algebra as a novel solution to this problem in the context of models of infectious disease spread. Process algebra allows us to describe a system in terms of the stochastic behaviour of individuals, and is a technique from computer science. We review the use of process algebra in biological systems, and the variety of quantitative and qualitative analysis techniques available. The analysis illustrated here solves the changing scale problem: from the individual behaviour we can rigorously derive equations to describe the mean behaviour of the system at the level of the population. The biological problem investigated is the transmission of infection, and how this relates to individual interactions.
Nakayama, Y; Aoki, Y; Niitsu, H; Saigusa, K
2001-04-15
Forensic dentistry plays an essential role in personal identification procedures. An adequate interincisal space of cadavers with rigor mortis is required to obtain detailed dental findings. We have developed intraoral and two directional approaches, for myotomy of the temporal muscles. The intraoral approach, in which the temporalis was dissected with scissors inserted via an intraoral incision, was adopted for elderly cadavers, females and emaciated or exhausted bodies, and had a merit of no incision on the face. The two directional approach, in which myotomy was performed with thread-wire saw from behind and with scissors via the intraoral incision, was designed for male muscular youths. Both approaches were effective to obtain a desired degree of an interincisal opening without facial damage.
Energy Technology Data Exchange (ETDEWEB)
Clocquet, R. [DHV, Amersfoort (Netherlands); Koene, F. [ECN Efficiency and Infrastructure, Petten (Netherlands)
2010-05-15
How can existing terraced houses be renovated in such a way that their energy use decreases with 75 percent? The Rigorous project of the Energy research Centre of the Netherlands (ECN), TNO, Delft University of Technology and DHV, developed innovative renovation concepts that make such savings feasible by combining constructional measures with installation concepts. On top of that it is also essential that consumer behavior is addressed. [Dutch] Hoe kunnen bestaande rijtjeswoningen zo worden gerenoveerd dat het totale energiegebruik met 75 procent afneemt? In het Rigoureus-project hebben ECN, TNO, TU Delft en DHV innovatieve renovatieconcepten ontwikkeld die dat, door een combinatie van bouwkundige maatregelen en uitgeldende installatieconcepten, mogelijk maken. Daarbij blijkt het van essentieel belang ook het gebruikersgedrag aan te pakken.
International Nuclear Information System (INIS)
Adam, G.; Adam, S.
2007-01-01
The Green function (GF) equation of motion technique for solving the effective two-band Hubbard model of high-T c superconductivity in cuprates rests on the Hubbard operator (HO) algebra. We show that, if we take into account the invariance to translations and spin reversal, the HO algebra results in invariance properties of several specific correlation functions. The use of these properties allows rigorous derivation and simplification of the expressions of the frequency matrix (FM) and of the generalized mean-field approximation (GMFA) Green functions (GFs) of the model. For the normal singlet hopping and anomalous exchange pairing correlation functions which enter the FM and GMFA-GFs, the use of spectral representations allows the identification and elimination of exponentially small quantities. This procedure secures the reduction of the correlation order to the GMFA-GF expressions
Importance of All-in-one (MCNPX2.7.0+CINDER2008) Code for Rigorous Transmutation Study
Energy Technology Data Exchange (ETDEWEB)
Kim, Oyeon [Institute for Modeling and Simulation Convergence, Daegu (Korea, Republic of); Kim, Kwanghyun [RadTek Co. Ltd., Daejeon (Korea, Republic of)
2015-10-15
It can be utilized as a possible mechanism for reducing the volume and hazard of radioactive waste by transforming hazardous radioactive elements with long half-life into less hazardous elements with short halflife. Thus, the understanding of the transmutation mechanism and beneficial machinery design technologies are important and useful. Although the terminology transmutation was rooted back to alchemy which transforms the base metals into gold in the middle ages, Rutherford and Soddy were the first observers by discovering the natural transmutation as a part of radioactive decay of the alpha decay type in early 20th century. Along with the development of computing technology, analysis software, for example, CINDER was developed for rigorous atomic transmutation study. The code has a long history of development from the original work of T. England at Bettis Atomic Power Laboratory (BAPL) in the early 1960s. It has been used to calculate the inventory of nuclides in an irradiated material. CINDER'90 which is recently released involved an upgrade of the code to allow the spontaneous tracking of chains based upon the significant density or pass-by of a nuclide, where pass-by represents the density of a nuclide transforming to other nuclides. Nuclear transmutation process is governed by highly non-linear differential equation. Chaotic nature of the non-linear equation bespeaks the importance of the accurate input data (i.e. number of significant digits). Thus, reducing the human interrogation is very important for the rigorous transmutation study and 'allin- one' code structure is desired. Note that non-linear characteristic of the transmutation equation caused by the flux changes due to the number density change during a given time interval (intrinsic physical phenomena) is not considered in this study. In this study, we only emphasized the effects of human interrogation in the computing process solving nonlinear differential equations, as shown in
Directory of Open Access Journals (Sweden)
Matthew Hunt
Full Text Available Research conducted following natural disasters such as earthquakes, floods or hurricanes is crucial for improving relief interventions. Such research, however, poses ethical, methodological and logistical challenges for researchers. Oversight of disaster research also poses challenges for research ethics committees (RECs, in part due to the rapid turnaround needed to initiate research after a disaster. Currently, there is limited knowledge available about how RECs respond to and appraise disaster research. To address this knowledge gap, we investigated the experiences of REC members who had reviewed disaster research conducted in low- or middle-income countries.We used interpretive description methodology and conducted in-depth interviews with 15 respondents. Respondents were chairs, members, advisors, or coordinators from 13 RECs, including RECs affiliated with universities, governments, international organizations, a for-profit REC, and an ad hoc committee established during a disaster. Interviews were analyzed inductively using constant comparative techniques.Through this process, three elements were identified as characterizing effective and high-quality review: timeliness, responsiveness and rigorousness. To ensure timeliness, many RECs rely on adaptations of review procedures for urgent protocols. Respondents emphasized that responsive review requires awareness of and sensitivity to the particularities of disaster settings and disaster research. Rigorous review was linked with providing careful assessment of ethical considerations related to the research, as well as ensuring independence of the review process.Both the frequency of disasters and the conduct of disaster research are on the rise. Ensuring effective and high quality review of disaster research is crucial, yet challenges, including time pressures for urgent protocols, exist for achieving this goal. Adapting standard REC procedures may be necessary. However, steps should be
Araújo, Luciano V; Malkowski, Simon; Braghetto, Kelly R; Passos-Bueno, Maria R; Zatz, Mayana; Pu, Calton; Ferreira, João E
2011-12-22
Recent medical and biological technology advances have stimulated the development of new testing systems that have been providing huge, varied amounts of molecular and clinical data. Growing data volumes pose significant challenges for information processing systems in research centers. Additionally, the routines of genomics laboratory are typically characterized by high parallelism in testing and constant procedure changes. This paper describes a formal approach to address this challenge through the implementation of a genetic testing management system applied to human genome laboratory. We introduced the Human Genome Research Center Information System (CEGH) in Brazil, a system that is able to support constant changes in human genome testing and can provide patients updated results based on the most recent and validated genetic knowledge. Our approach uses a common repository for process planning to ensure reusability, specification, instantiation, monitoring, and execution of processes, which are defined using a relational database and rigorous control flow specifications based on process algebra (ACP). The main difference between our approach and related works is that we were able to join two important aspects: 1) process scalability achieved through relational database implementation, and 2) correctness of processes using process algebra. Furthermore, the software allows end users to define genetic testing without requiring any knowledge about business process notation or process algebra. This paper presents the CEGH information system that is a Laboratory Information Management System (LIMS) based on a formal framework to support genetic testing management for Mendelian disorder studies. We have proved the feasibility and showed usability benefits of a rigorous approach that is able to specify, validate, and perform genetic testing using easy end user interfaces.
Sams, A R; Dzuik, C S
1999-10-01
This study was conducted to evaluate the combined rigor-accelerating effects of postmortem electrical stimulation (ES) and argon-induced anoxia (Ar) of broiler chickens. One hundred broilers were processed in the following treatments: untreated controls, ES, Ar, or Ar with ES (Ar + ES). Breast fillets were harvested at 1 h postmortem for all treatments or at 1 and 6 h postmortem for the control carcasses. Fillets were sampled for pH and ratio of inosine to adenosine (R-value) and were then individually quick frozen (IQF) or aged on ice (AOI) until 24 h postmortem. Color was measured in the AOI fillets at 24 h postmortem. All fillets were then cooked and evaluated for Allo-Kramer shear value. The Ar treatment accelerated the normal pH decline, whereas the ES and AR + ES treatments yielded even lower pH values at 1 h postmortem. The Ar + ES treatment had a greater R-value than the ES treatment, which was greater than either the Ar or 1-h controls, which, in turn, were not different from each other. The ES treatment had the lowest L* value, and ES, Ar, and Ar + ES produced significantly higher a* values than the 1-h controls. For the IQF fillets, the ES and Ar + ES treatments were not different in shear value but were lower than Ar, which was lower than the 1-h controls. The same was true for the AOI fillets except that the ES and the Ar treatments were not different. These results indicated that although ES and Ar had rigor-accelerating and tenderizing effects, ES seemed to be more effective than Ar; there was little enhancement when Ar was added to the ES treatment and fillets were deboned at 1 h postmortem.
Qian, Ma; Ma, Jie
2009-06-07
Fletcher's spherical substrate model [J. Chem. Phys. 29, 572 (1958)] is a basic model for understanding the heterogeneous nucleation phenomena in nature. However, a rigorous thermodynamic formulation of the model has been missing due to the significant complexities involved. This has not only left the classical model deficient but also likely obscured its other important features, which would otherwise have helped to better understand and control heterogeneous nucleation on spherical substrates. This work presents a rigorous thermodynamic formulation of Fletcher's model using a novel analytical approach and discusses the new perspectives derived. In particular, it is shown that the use of an intermediate variable, a selected geometrical angle or pseudocontact angle between the embryo and spherical substrate, revealed extraordinary similarities between the first derivatives of the free energy change with respect to embryo radius for nucleation on spherical and flat substrates. Enlightened by the discovery, it was found that there exists a local maximum in the difference between the equivalent contact angles for nucleation on spherical and flat substrates due to the existence of a local maximum in the difference between the shape factors for nucleation on spherical and flat substrate surfaces. This helps to understand the complexity of the heterogeneous nucleation phenomena in a practical system. Also, it was found that the unfavorable size effect occurs primarily when R<5r( *) (R: radius of substrate and r( *): critical embryo radius) and diminishes rapidly with increasing value of R/r( *) beyond R/r( *)=5. This finding provides a baseline for controlling the size effects in heterogeneous nucleation.
Simic, Vladimir
2016-06-01
As the number of end-of-life vehicles (ELVs) is estimated to increase to 79.3 million units per year by 2020 (e.g., 40 million units were generated in 2010), there is strong motivation to effectively manage this fast-growing waste flow. Intensive work on management of ELVs is necessary in order to more successfully tackle this important environmental challenge. This paper proposes an interval-parameter chance-constraint programming model for end-of-life vehicles management under rigorous environmental regulations. The proposed model can incorporate various uncertainty information in the modeling process. The complex relationships between different ELV management sub-systems are successfully addressed. Particularly, the formulated model can help identify optimal patterns of procurement from multiple sources of ELV supply, production and inventory planning in multiple vehicle recycling factories, and allocation of sorted material flows to multiple final destinations under rigorous environmental regulations. A case study is conducted in order to demonstrate the potentials and applicability of the proposed model. Various constraint-violation probability levels are examined in detail. Influences of parameter uncertainty on model solutions are thoroughly investigated. Useful solutions for the management of ELVs are obtained under different probabilities of violating system constraints. The formulated model is able to tackle a hard, uncertainty existing ELV management problem. The presented model has advantages in providing bases for determining long-term ELV management plans with desired compromises between economic efficiency of vehicle recycling system and system-reliability considerations. The results are helpful for supporting generation and improvement of ELV management plans. Copyright © 2016 Elsevier Ltd. All rights reserved.
Directory of Open Access Journals (Sweden)
Grimshaw Jeremy M
2010-02-01
Full Text Available Abstract Background There is growing interest in the use of cognitive, behavioural, and organisational theories in implementation research. However, the extent of use of theory in implementation research is uncertain. Methods We conducted a systematic review of use of theory in 235 rigorous evaluations of guideline dissemination and implementation studies published between 1966 and 1998. Use of theory was classified according to type of use (explicitly theory based, some conceptual basis, and theoretical construct used and stage of use (choice/design of intervention, process/mediators/moderators, and post hoc/explanation. Results Fifty-three of 235 studies (22.5% were judged to have employed theories, including 14 studies that explicitly used theory. The majority of studies (n = 42 used only one theory; the maximum number of theories employed by any study was three. Twenty-five different theories were used. A small number of theories accounted for the majority of theory use including PRECEDE (Predisposing, Reinforcing, and Enabling Constructs in Educational Diagnosis and Evaluation, diffusion of innovations, information overload and social marketing (academic detailing. Conclusions There was poor justification of choice of intervention and use of theory in implementation research in the identified studies until at least 1998. Future research should explicitly identify the justification for the interventions. Greater use of explicit theory to understand barriers, design interventions, and explore mediating pathways and moderators is needed to advance the science of implementation research.
Jenkins, Michael G.; Salem, Jonathan A.
2016-01-01
Physical and mechanical properties and performance of advanced ceramics and glasses are difficult to measure correctly without the proper techniques. For over three decades, ASTM Committee C28 on Advanced Ceramics, has developed high-quality, technically-rigorous, full-consensus standards (e.g., test methods, practices, guides, terminology) to measure properties and performance of monolithic and composite ceramics that may be applied to glasses in some cases. These standards contain testing particulars for many mechanical, physical, thermal, properties and performance of these materials. As a result these standards are used to generate accurate, reliable, repeatable and complete data. Within Committee C28, users, producers, researchers, designers, academicians, etc. have written, continually updated, and validated through round-robin test programs, 50 standards since the Committee's founding in 1986. This paper provides a detailed retrospective of the 30 years of ASTM Committee C28 including a graphical pictogram listing of C28 standards along with examples of the tangible benefits of standards for advanced ceramics to demonstrate their practical applications.
Jenkins, Michael G.; Salem, Jonathan A.
2016-01-01
Physical and mechanical properties and performance of advanced ceramics and glasses are difficult to measure correctly without the proper techniques. For over three decades, ASTM Committee C28 on Advanced Ceramics, has developed high quality, rigorous, full-consensus standards (e.g., test methods, practices, guides, terminology) to measure properties and performance of monolithic and composite ceramics that may be applied to glasses in some cases. These standards testing particulars for many mechanical, physical, thermal, properties and performance of these materials. As a result these standards provide accurate, reliable, repeatable and complete data. Within Committee C28 users, producers, researchers, designers, academicians, etc. have written, continually updated, and validated through round-robin test programs, nearly 50 standards since the Committees founding in 1986. This paper provides a retrospective review of the 30 years of ASTM Committee C28 including a graphical pictogram listing of C28 standards along with examples of the tangible benefits of advanced ceramics standards to demonstrate their practical applications.
Benn, Emma K T; Alexis, Andrew; Mohamed, Nihal; Wang, Yan-Hong; Khan, Ikhlas A; Liu, Bian
2016-12-01
Skin-bleaching practices, such as using skin creams and soaps to achieve a lighter skin tone, are common throughout the world and are triggered by cosmetic reasons that oftentimes have deep historical, economic, sociocultural, and psychosocial roots. Exposure to chemicals in the bleaching products, notably, mercury (Hg), hydroquinone, and steroids, has been associated with a variety of adverse health effects, such as Hg poisoning and exogenous ochronosis. In New York City (NYC), skin care product use has been identified as an important route of Hg exposure, especially among Caribbean-born blacks and Dominicans. However, surprisingly sparse information is available on the epidemiology of the health impacts of skin-bleaching practices among these populations. We highlight the dearth of large-scale, comprehensive, community-based, clinical, and translational research in this area, especially the limited skin-bleaching-related research among non-White populations in the US. We offer five new research directions, including investigating the known and under-studied health consequences among populations for which the skin bleach practice is newly emerging at an alarming rate using innovative laboratory and statistical methods. We call for conducting methodologically rigorous, multidisciplinary, and culturally sensitive research in order to provide insights into the root and the epidemiological status of the practice and provide evidence of exposure-outcome associations, with an ultimate goal of developing potential intervention strategies to reduce the health burdens of skin-bleaching practice.
The Method of Recursive Counting: Can one go further?
International Nuclear Information System (INIS)
Creutz, M.; Horvath, I.
1993-12-01
After a short review of the Method of Recursive Counting we introduce a general algebraic description of recursive lattice building. This provides a rigorous framework for discussion of method's limitations
Alvarado, C Z; Sams, A R
2000-09-01
This study was conducted to evaluate the effects of electrical stimulation (ES) on rigor mortis development, calpastatin activity, and tenderness in anatomically similar avian muscles composed primarily of either red or white muscle fibers. A total of 72 broilers and 72 White Pekin ducks were either treated with postmortem (PM) ES (450 mA) at the neck in a 1% NaCl solution for 2 s on and 1 s off for a total of 15 s or were used as nonstimulated controls. Both pectoralis muscles were harvested from the carcasses after 0.25, 1.25, and 24 h PM and analyzed for pH, inosine:adenosine ratio (R-value), sarcomere length, gravimetric fragmentation index, calpastatin activity, shear value, and cook loss. All data were analyzed within species for the effects of ES. Electrically stimulated ducks had a lower muscle pH at 0.25 and 1.25 h PM and higher R-values at 0.25 h PM compared with controls. Electrically stimulated broilers had a lower muscle pH at 1.25 h and higher R-values at 0.25 and 1.25 h PM compared with controls. Muscles of electrically stimulated broilers exhibited increased myofibrillar fragmentation at 0.25 and 1.25 h PM, whereas there was no such difference over PM time in the duck muscle. Electrical stimulation did not affect calpastatin activity in either broilers or ducks; however, the calpastatin activity of the broilers did decrease over the aging time period, whereas that of the ducks did not. Electrical stimulation decreased shear values in broilers at 1.25 h PM compared with controls; however, there was no difference in shear values of duck muscle due to ES at any sampling time. Cook loss was lower for electrically stimulated broilers at 0.25 and 1.25 h PM compared with the controls, but had no effect in the ducks. These results suggest that the red fibers of the duck pectoralis have less potential for rigor mortis acceleration and tenderization due to ES than do the white fibers of the broiler pectoralis.
International Nuclear Information System (INIS)
Corti, D.S.; Debenedetti, P.G.
1998-01-01
The rigorous statistical mechanics of metastability requires the imposition of internal constraints that prevent access to regions of phase space corresponding to inhomogeneous states. We derive exactly the Helmholtz energy and equation of state of the one-dimensional hard rod fluid under the influence of an internal constraint that places an upper bound on the distance between nearest-neighbor rods. This type of constraint is relevant to the suppression of boiling in a superheated liquid. We determine the effects of this constraint upon the thermophysical properties and internal structure of the hard rod fluid. By adding an infinitely weak and infinitely long-ranged attractive potential to the hard core, the fluid exhibits a first-order vapor-liquid transition. We determine exactly the equation of state of the one-dimensional superheated liquid and show that it exhibits metastable phase equilibrium. We also derive statistical mechanical relations for the equation of state of a fluid under the action of arbitrary constraints, and show the connection between the statistical mechanics of constrained and unconstrained ensembles. copyright 1998 The American Physical Society
Changyong, Dou; Huadong, Guo; Chunming, Han; yuquan, Liu; Xijuan, Yue; Yinghui, Zhao
2014-03-01
Raw signal simulation is a useful tool for the system design, mission planning, processing algorithm testing, and inversion algorithm design of Synthetic Aperture Radar (SAR). Due to the wide and high frequent variation of aircraft's trajectory and attitude, and the low accuracy of the Position and Orientation System (POS)'s recording data, it's difficult to quantitatively study the sensitivity of the key parameters, i.e., the baseline length and inclination, absolute phase and the orientation of the antennas etc., of the airborne Interferometric SAR (InSAR) system, resulting in challenges for its applications. Furthermore, the imprecise estimation of the installation offset between the Global Positioning System (GPS), Inertial Measurement Unit (IMU) and the InSAR antennas compounds the issue. An airborne interferometric SAR (InSAR) simulation based on the rigorous geometric model and real navigation data is proposed in this paper, providing a way for quantitatively studying the key parameters and for evaluating the effect from the parameters on the applications of airborne InSAR, as photogrammetric mapping, high-resolution Digital Elevation Model (DEM) generation, and surface deformation by Differential InSAR technology, etc. The simulation can also provide reference for the optimal design of the InSAR system and the improvement of InSAR data processing technologies such as motion compensation, imaging, image co-registration, and application parameter retrieval, etc.
International Nuclear Information System (INIS)
Changyong, Dou; Huadong, Guo; Chunming, Han; Yuquan, Liu; Xijuan, Yue; Yinghui, Zhao
2014-01-01
Raw signal simulation is a useful tool for the system design, mission planning, processing algorithm testing, and inversion algorithm design of Synthetic Aperture Radar (SAR). Due to the wide and high frequent variation of aircraft's trajectory and attitude, and the low accuracy of the Position and Orientation System (POS)'s recording data, it's difficult to quantitatively study the sensitivity of the key parameters, i.e., the baseline length and inclination, absolute phase and the orientation of the antennas etc., of the airborne Interferometric SAR (InSAR) system, resulting in challenges for its applications. Furthermore, the imprecise estimation of the installation offset between the Global Positioning System (GPS), Inertial Measurement Unit (IMU) and the InSAR antennas compounds the issue. An airborne interferometric SAR (InSAR) simulation based on the rigorous geometric model and real navigation data is proposed in this paper, providing a way for quantitatively studying the key parameters and for evaluating the effect from the parameters on the applications of airborne InSAR, as photogrammetric mapping, high-resolution Digital Elevation Model (DEM) generation, and surface deformation by Differential InSAR technology, etc. The simulation can also provide reference for the optimal design of the InSAR system and the improvement of InSAR data processing technologies such as motion compensation, imaging, image co-registration, and application parameter retrieval, etc
International Nuclear Information System (INIS)
ZareNezhad, Bahman; Mottahedin, Mona
2012-01-01
Highlights: ► A mechanistic model for predicting gas hydrate formation kinetics is presented. ► A secondary nucleation rate model is proposed for the first time. ► Crystal–crystal collisions and crystal–impeller collisions are distinguished. ► Simultaneous determination of nucleation and growth kinetics are established. ► Important for design of gas hydrate based energy storage and CO 2 recovery systems. - Abstract: A rigorous mechanistic model for predicting gas hydrate formation crystallization kinetics is presented and the special case of CO 2 gas hydrate formation regarding CO 2 recovery and sequestration processes has been investigated by using the proposed model. A physical model for prediction of secondary nucleation rate is proposed for the first time and the formation rates of secondary nuclei by crystal–crystal collisions and crystal–impeller collisions are formulated. The objective functions for simultaneous determination of nucleation and growth kinetics are presented and a theoretical framework for predicting the dynamic behavior of gas hydrate formation is presented. Predicted time variations of CO 2 content, total number and surface area of produced hydrate crystals are in good agreement with the available experimental data. The proposed approach can have considerable application for design of gas hydrate converters regarding energy storage and CO 2 recovery processes.
Directory of Open Access Journals (Sweden)
Tatiana Pacheco Rodrigues
2004-08-01
Full Text Available Esta pesquisa utilizou 12 eqüinos de diferentes idades, abatidos em um matadouro-frigorífico (SIF 1803 em Araguari-MG, e estudou a temperatura, pH, comprimento de sarcômero em diferentes intervalos de tempo após abate (1h, 5h, 8h, 10h, 12h, 15h e 24h e força de cisalhamento (maciez dos músculos Longissimus dorsi e Semitendinosus, com intuito de caracterizar o desenvolvimento do processo de rigor mortis de eqüídeos durante o processamento industrial. A temperatura da câmara fria variou de 10,2°C a 4,0°C e a temperatura média inicial das carcaças foi de 35,32°C e a final de 4,15°C. O pH inicial médio do músculo Longissimus dorsi foi 6,49 e o final 5,63, e para o músculo Semitendinosus o pH inicial médio foi 6,44 e o final 5,70. A menor medida de sarcômero observada em ambos os músculos foi na 15ª hora após abate, ou seja, 1,44µm e 1,41µm, respectivamente. A carne dos eqüídeos adultos foi mais dura (pThis work studied 12 horses at different ages butchered in a slaughterhouse in Minas Gerais State, Brazil (SIF 1803 and evaluated temperature, pH, sarcomere length in different periods after slaughter (1h, 5h, 8h, 10h, 12h, 15h, and 24 hours as well as the shear force (meat tenderness of the Longissimus dorsi and Semitendinosus muscles, aiming at characterizing the rigor mortis onset in the meat during industrial processing. The chilly room temperature varied from 10.2°C to 4.0°C, and the mean initial carcass temperature was 35.32°C and the final one was 4.15°C. The mean initial pH of Longissimus dorsi was 6.49 and the final one was 5.63; the mean initial pH of Semitendinosus was 6.44 and the final one was 5.70. The smallest sarcomere size obtained in both muscles occurred at 15 hours postmortem, and the sarcomere lengths were 1.44 µm and 1.41 µm, respectively. The meat from adult horses was tougher than that from young ones (p<0.05, and the Semitendinosus muscle was tougher than Longissimus dorsi muscle.
International Nuclear Information System (INIS)
Verhaart, Rene F.; Rijnen, Zef; Verduijn, Gerda M.; Paulides, Margarethus M.; Fortunati, Valerio; Walsum, Theo van; Veenland, Jifke F.
2014-01-01
Hyperthermia treatment planning (HTP) is used in the head and neck region (H and N) for pretreatment optimization, decision making, and real-time HTP-guided adaptive application of hyperthermia. In current clinical practice, HTP is based on power-absorption predictions, but thermal dose-effect relationships advocate its extension to temperature predictions. Exploitation of temperature simulations requires region- and temperature-specific thermal tissue properties due to the strong thermoregulatory response of H and N tissues. The purpose of our work was to develop a technique for patient group-specific optimization of thermal tissue properties based on invasively measured temperatures, and to evaluate the accuracy achievable. Data from 17 treated patients were used to optimize the perfusion and thermal conductivity values for the Pennes bioheat equation-based thermal model. A leave-one-out approach was applied to accurately assess the difference between measured and simulated temperature (∇T). The improvement in ∇T for optimized thermal property values was assessed by comparison with the ∇T for values from the literature, i.e., baseline and under thermal stress. The optimized perfusion and conductivity values of tumor, muscle, and fat led to an improvement in simulation accuracy (∇T: 2.1 ± 1.2 C) compared with the accuracy for baseline (∇T: 12.7 ± 11.1 C) or thermal stress (∇T: 4.4 ± 3.5 C) property values. The presented technique leads to patient group-specific temperature property values that effectively improve simulation accuracy for the challenging H and N region, thereby making simulations an elegant addition to invasive measurements. The rigorous leave-one-out assessment indicates that improvements in accuracy are required to rely only on temperature-based HTP in the clinic. (orig.) [de
Azese, Martin Ndi
2018-02-01
This article presents a rigorous calculation involving velocity slip of Newtonian fluid where we analyze and solve the unsteady Navier-Stokes equation with emphasis on its rheological implication. The goal of which is to model a simple yet effective non-invasive way of quantifying and characterizing slippage. Indeed this contrasts with previous techniques that exhibit inherent limitations whereby injecting foreign objects usually alter the flow. This problem is built on the Couette rheological flow system such that μ-Newton force and μ-stress are captured and processed to obtain wall slip. Our model leads to a linear partial differential equation and upon enforcing linear-Navier slip boundary conditions (BC) yields inhomogeneous and unsteady "Robin-type" BC. A dimensional analysis reveals salient dimensionless parameters: Roshko, Strouhal, and Reynolds while highlighting slip-numbers from BC. We also solve the slip-free case to corroborate and validate our results. Several graphs are generated showing slip effects, particularly, studying how slip-numbers, a key input, differentiate themselves to the outputs. We also confirm this in a graphical fashion by presenting the flow profile across channel width, velocity, and stress at both walls. A perturbation scheme is introduced to calculate long-time behavior when the system seats for long. More importantly, in the end, we justify the existence of a reverse mechanism, where an inverse transformation like Fourier transform uses the output data to retrieve slip-numbers and slip law, thus quantifying and characterizing slip. Therefore, we not only substantiate our analysis, but we also justify our claim, measurement and characterization, and theorize realizability of our proposition.
Lowe, Donald R.
1989-01-01
Thin cherty sedimentary layers within the volcanic portions of the 3,500 to 3,300 Ma-old Onverwacht and Fig Tree Groups, Barberton Greenstone belt, South Africa, and Warrawoona Group, eastern Pilbara Block, Western Australia, contain an abundant record of early Archean life. Five principal types of organic and probably biogenic remains and or structures can be identifed: stromatolites, stromatolite detritus, carbonaceous laminite or flat stromalite, carbonaceous detrital particles, and microfossils. Early Archean stromatolites were reported from both the Barberton and eastern Pilbara greenstone belts. Systematic studies are lacking, but two main morphological types of stromatolites appear to be represented by these occurrences. Morphology of the stromalites is described. Preserved early Archean stromatolites and carbonaceous matter appear to reflect communities of photosynthetic cyanobacteria inhabiting shallow, probably marine environments developed over the surfaces of low-relief, rapidly subsiding, simatic volcanic platforms. The overall environmental and tectonic conditions were those that probably prevailed at Earth's surface since the simatic crust and oceans formed sometime before 3,800 Ma. Recent studies also suggest that these early Archean sequences contain layers of debris formed by large-body impacts on early Earth. If so, then these early bacterial communities had developed strategies for coping with the disruptive effects of possibly globe-encircling high-temperature impact vapor clouds, dust blankets, and impact-generated tsunamis. It is probable that these early Archean biogenic materials represent organic communities that evolved long before the beginning of the preserved geological record and were well adapted to the rigors of life on a young, volcanically active Earth during late bombardment. These conditions may have had parallels on Mars during its early evolution.
International Nuclear Information System (INIS)
Lin, Keh Ying
2006-01-01
We have derived rigorously the perimeter generating functions for the mean-squared radius of gyration of rectangular, Ferrers and pyramid polygons. These functions were found by Jensen recently. His nonrigorous results are based on the analysis of the long series expansions. (comment)
Conn, Katharine
2014-01-01
In the last three decades, there has been a large increase in the number of rigorous experimental and quasi-experimental evaluations of education programs in developing countries. These impact evaluations have taken place all over the globe, including a large number in Sub-Saharan Africa (SSA). The fact that the developing world is socially and…
Directory of Open Access Journals (Sweden)
Salma Noor-E Islami
2014-12-01
Full Text Available Rigor-index in market-size striped catfish (Pangasianodon hypophthalmus, locally called Thai-Pangas was determined to assess fillet yield for production of value-added products. In whole fish, rigor started within 1 hr after death under both iced and room temperature conditions while rigor-index reached a maximum of 72.23% within 8 hr and 85.5% within 5 hr at room temperature and iced condition, respectively, which was fully relaxed after 22 hr under both storage conditions. Post-mortem muscle pH decreased to 6.8 after 2 hr, 6.2 after 8 hr and sharp increase to 6.9 after 9 hr. There was a positive correlation between rigor progress and pH shift in fish fillets. Hand filleting was done post-rigor and fillet yield experiment showed 50.4±2.1% fillet, 8.0±0.2% viscera, 8.0±1.3% skin and 32.0±3.2% carcass could be obtained from Thai-Pangas. Proximate composition analysis of four regions of Thai-Pangas viz., head region, middle region, tail region and viscera revealed moisture 78.36%, 81.14%, 81.45% and 57.33%; protein 15.83%, 15.97%, 16.14% and 17.20%; lipid 4.61%, 1.82%, 1.32% and 24.31% and ash 1.09%, 0.96%, 0.95% and 0.86%, respectively indicating suitability of Thai-Pangas for production of value-added products such as fish fillets.
Gates, Julie; Lam, Geanette; Ortiz, José A; Losson, Régine; Thummel, Carl S
2004-01-01
Pulses of the steroid hormone ecdysone trigger the major developmental transitions in Drosophila, including molting and puparium formation. The ecdysone signal is transduced by the EcR/USP nuclear receptor heterodimer that binds to specific response elements in the genome and directly regulates target gene transcription. We describe a novel nuclear receptor interacting protein encoded by rigor mortis (rig) that is required for ecdysone responses during larval development. rig mutants display defects in molting, delayed larval development, larval lethality, duplicated mouth parts, and defects in puparium formation--phenotypes that resemble those seen in EcR, usp, E75A and betaFTZ-F1 mutants. Although the expression of these nuclear receptor genes is essentially normal in rig mutant larvae, the ecdysone-triggered switch in E74 isoform expression is defective. rig encodes a protein with multiple WD-40 repeats and an LXXLL motif, sequences that act as specific protein-protein interaction domains. Consistent with the presence of these elements and the lethal phenotypes of rig mutants, Rig protein interacts with several Drosophila nuclear receptors in GST pull-down experiments, including EcR, USP, DHR3, SVP and betaFTZ-F1. The ligand binding domain of betaFTZ-F1 is sufficient for this interaction, which can occur in an AF-2-independent manner. Antibody stains reveal that Rig protein is present in the brain and imaginal discs of second and third instar larvae, where it is restricted to the cytoplasm. In larval salivary gland and midgut cells, however, Rig shuttles between the cytoplasm and nucleus in a spatially and temporally regulated manner, at times that correlate with the major lethal phase of rig mutants and major switches in ecdysone-regulated gene expression. Taken together, these data indicate that rig exerts essential functions during larval development through gene-specific effects on ecdysone-regulated transcription, most likely as a cofactor for one or more
Parametric Studies on Artificial Morpho Butterfly Wing Scales for Optical Device Applications
Kim, Hyun Myung; Kim, Sang Hyeok; Lee, Gil Ju; Kim, Kyujung; Song, Young Min
2015-01-01
We calculated diffraction efficiencies of grating structures inspired by Morpho butterfly wings by using a rigorous coupled-wave analysis method. The geometrical effects, such as grating width, period, thickness, and material index, were investigated in order to obtain better optical performance. Closely packed grating structures with an optimized membrane thickness show vivid reflected colors and provide high sensitivity to surrounding media variations, which is applicable to vapor sensing o...
Directory of Open Access Journals (Sweden)
Joana Maia Mendes
2015-06-01
Full Text Available ResumoO presente trabalho avaliou a influência do estresse pré-abate e do método de abate sobre o rigor mortis do tambaqui durante armazenamento em gelo. Foram estudadas respostas fisiológicas do tambaqui ao estresse durante o pré-abate, que foi dividido em quatro etapas: despesca, transporte, recuperação por 24 h e por 48 h. Ao final de cada etapa, os peixes foram amostrados para caracterização do estresse pré-abate por meio de análises dos parâmetros plasmáticos de glicose, lactato e amônia e, em seguida, os peixes foram abatidos por hipotermia ou por asfixia com gás carbônico para o estudo do rigor mortis. Verificou-se que o estado fisiológico de estresse dos peixes foi mais agudo logo após o transporte, implicando numa entrada em rigor mortis mais rápida: 60 minutos para tambaquis abatidos por hipotermia e 120 minutos para tambaquis abatidos por asfixia com gás carbônico. Nos viveiros, os peixes abatidos logo após a despesca apresentaram estado de estresse intermediário, sem diferença no tempo de entrada em rigor mortis em relação ao método de abate (135 minutos. Os peixes que passaram por recuperação ao estresse causado pelo transporte em condições simuladas de indústria apresentaram entrada em rigor mortis mais tardia: 225 minutos (com 24 h de recuperação e 255 minutos (com 48 h de recuperação, igualmente sem diferença em relação aos métodos de abate testados. A resolução do rigor mortis foi mais rápida nos peixes abatidos após o transporte, que foi de 12 dias. Nos peixes abatidos logo após a despesca, a resolução ocorreu com 16 dias e, nos peixes abatidos após recuperação, com 20 dias para 24 h de recuperação ao estresse pré-abate e 24 dias para 48 h de recuperação, sem influência do método de abate na resolução do rigor mortis. Assim, é desejável que o abate do tambaqui destinado à indústria seja feito após período de recuperação ao estresse, com vistas a aumentar sua
Mixed Methods and Action Research: similar or different?
Wiśniewska, Danuta
2011-01-01
This article attempts to analyse and compare ELT studies grounded solely in mixed methods and ELT action research studies based on a mixed methods approach in order to identify to what degree action research studies combining different methods in a single study comply with the principles of rigorous mixed methods study.
Lu, Han; Liu, Xiaochang; Zhang, Yuemei; Wang, Hang; Luo, Yongkang
2015-12-01
To investigate the effects of chilling and partial freezing on rigor mortis changes in bighead carp (Aristichthys nobilis), pH, cathepsin B, cathepsin B+L activities, SDS-PAGE of sarcoplasmic and myofibrillar proteins, texture, and changes in microstructure of fillets at 4 °C and -3 °C were determined at 0, 2, 4, 8, 12, 24, 48, and 72 h after slaughter. The results indicated that pH of fillets (6.50 to 6.80) was appropriate for cathepsin function during the rigor mortis. For fillets that were chilled and partially frozen, the cathepsin activity in lysosome increased consistently during the first 12 h, followed by a decrease from the 12 to 24 h, which paralleled an increase in activity in heavy mitochondria, myofibrils and sarcoplasm. There was no significant difference in cathepsin activity in lysosomes between fillets at 4 °C and -3 °C (P > 0.05). Partially frozen fillets had greater cathepsin activity in heavy mitochondria than chilled samples from the 48 to 72 h. In addition, partially frozen fillets showed higher cathepsin activity in sarcoplasm and lower cathepsin activity in myofibrils compared with chilled fillets. Correspondingly, we observed degradation of α-actinin (105 kDa) by cathepsin L in chilled fillets and degradation of creatine kinase (41 kDa) by cathepsin B in partially frozen fillets during the rigor mortis. The decline of hardness for both fillets might be attributed to the accumulation of cathepsin in myofibrils from the 8 to 24 h. The lower cathepsin activity in myofibrils for fillets that were partially frozen might induce a more intact cytoskeletal structure than fillets that were chilled. © 2015 Institute of Food Technologists®
Pueyo Bellafont, Noèlia; Bagus, Paul S; Illas, Francesc
2015-06-07
A systematic study of the N(1s) core level binding energies (BE's) in a broad series of molecules is presented employing Hartree-Fock (HF) and the B3LYP, PBE0, and LC-BPBE density functional theory (DFT) based methods with a near HF basis set. The results show that all these methods give reasonably accurate BE's with B3LYP being slightly better than HF but with both PBE0 and LCBPBE being poorer than HF. A rigorous and general decomposition of core level binding energy values into initial and final state contributions to the BE's is proposed that can be used within either HF or DFT methods. The results show that Koopmans' theorem does not hold for the Kohn-Sham eigenvalues. Consequently, Kohn-Sham orbital energies of core orbitals do not provide estimates of the initial state contribution to core level BE's; hence, they cannot be used to decompose initial and final state contributions to BE's. However, when the initial state contribution to DFT BE's is properly defined, the decompositions of initial and final state contributions given by DFT, with several different functionals, are very similar to those obtained with HF. Furthermore, it is shown that the differences of Kohn-Sham orbital energies taken with respect to a common reference do follow the trend of the properly calculated initial state contributions. These conclusions are especially important for condensed phase systems where our results validate the use of band structure calculations to determine initial state contributions to BE shifts.
International Nuclear Information System (INIS)
Zhang Yongde.
1987-03-01
In this paper, the neutron Dirac-equation is presented. After decoupling it into two equations of the simple spinors, the rigorous solution of this equation is obtained in the case of slab-like uniform magnetic fields at perpendicular incidence. At non-relativistic approximation and first order approximation of weak field (NRWFA), our results have included all results that have been obtained in references for this case up to now. The corresponding transformations of the neutron's spin vectors are given. The single particle spectrum and its approximate expression are obtained. The characteristics of quantum statistics with the approximate expression of energy spectrum are studied. (author). 15 refs
DEFF Research Database (Denmark)
Coomber, S J; Bartels, E M; Elliott, G F
2011-01-01
contracts and breaks the microelectrode. Therefore the rigor state was studied. There is no reason to suppose a priori that a similar voltage switch does not occur during contraction, however. Calcium dependence is still apparent in muscles stretched beyond overlap (sarcomere length>3.8 μm) and is also seen...... in the gap filaments between the A- and I-band ends; further stretching abolishes the dependence. These experiments strongly suggest that calcium dependence is controlled initially by the titin component, and that this control is lost when titin filaments break. We suppose that that effect is mediated...
International Nuclear Information System (INIS)
Hara, T.; Tasaki, H.
1987-01-01
Continuing the analysis started in Part I of this work, they investigate critical phenomena in weakly coupled phi 4 spin systems in four dimensions. Concerning the critical behavior of the susceptibility and the correlation length (in the high-temperature phase), the existence of logarithmic corrections to their mean field type behavior is rigorously shown (i.e., they prove chi(t) ∼ t -1 absolute value 1n t/sup 1/3/, zeta(t) ∼ t/sup -1/2/ absolute value of ln t/sup 1/6/)
Analysis of a wavelength selectable cascaded DFB laser based on the transfer matrix method
International Nuclear Information System (INIS)
Xie Hongyun; Chen Liang; Shen Pei; Sun Botao; Wang Renqing; Xiao Ying; You Yunxia; Zhang Wanrong
2010-01-01
A novel cascaded DFB laser, which consists of two serial gratings to provide selectable wavelengths, is presented and analyzed by the transfer matrix method. In this method, efficient facet reflectivity is derived from the transfer matrix built for each serial section and is then used to simulate the performance of the novel cascaded DFB laser through self-consistently solving the gain equation, the coupled wave equation and the current continuity equations. The simulations prove the feasibility of this kind of wavelength selectable laser and a corresponding designed device with two selectable wavelengths of 1.51 μm and 1.53 μm is realized by experiments on InP-based multiple quantum well structure. (semiconductor devices)
Realizing Scientific Methods for Cyber Security
Energy Technology Data Exchange (ETDEWEB)
Carroll, Thomas E.; Manz, David O.; Edgar, Thomas W.; Greitzer, Frank L.
2012-07-18
There is little doubt among cyber security researchers about the lack of scientic rigor that underlies much of the liter-ature. The issues are manifold and are well documented. Further complicating the problem is insufficient scientic methods to address these issues. Cyber security melds man and machine: we inherit the challenges of computer science, sociology, psychology, and many other elds and create new ones where these elds interface. In this paper we detail a partial list of challenges imposed by rigorous science and survey how other sciences have tackled them, in the hope of applying a similar approach to cyber security science. This paper is by no means comprehensive: its purpose is to foster discussion in the community on how we can improve rigor in cyber security science.
DeGeest, David Scott; Schmidt, Frank
2015-01-01
Our objective was to apply the rigorous test developed by Browne (1992) to determine whether the circumplex model fits Big Five personality data. This test has yet to be applied to personality data. Another objective was to determine whether blended items explained correlations among the Big Five traits. We used two working adult samples, the Eugene-Springfield Community Sample and the Professional Worker Career Experience Survey. Fit to the circumplex was tested via Browne's (1992) procedure. Circumplexes were graphed to identify items with loadings on multiple traits (blended items), and to determine whether removing these items changed five-factor model (FFM) trait intercorrelations. In both samples, the circumplex structure fit the FFM traits well. Each sample had items with dual-factor loadings (8 items in the first sample, 21 in the second). Removing blended items had little effect on construct-level intercorrelations among FFM traits. We conclude that rigorous tests show that the fit of personality data to the circumplex model is good. This finding means the circumplex model is competitive with the factor model in understanding the organization of personality traits. The circumplex structure also provides a theoretically and empirically sound rationale for evaluating intercorrelations among FFM traits. Even after eliminating blended items, FFM personality traits remained correlated.
Hughes, Brianna H; Perkins, L Brian; Yang, Tom C; Skonberg, Denise I
2016-03-01
High pressure processing (HPP) of post-rigor abalone at 300MPa for 10min extended the refrigerated shelf-life to four times that of unprocessed controls. Shucked abalone meats were processed at 100 or 300MPa for 5 or 10min, and stored at 2°C for 35days. Treatments were analyzed for aerobic plate count (APC), total volatile base nitrogen (TVBN), K-value, biogenic amines, color, and texture. APC did not exceed 10(6) and TVBN levels remained below 35mg/100g for 35days for the 300MPa treatments. No biogenic amines were detected in the 300MPa treatments, but putrescine and cadaverine were detected in the control and 100MPa treatments. Color and texture were not affected by HPP or storage time. These results indicate that post-rigor processing at 300MPa for 10min can significantly increase refrigerated shelf-life of abalone without affecting chemical or physical quality characteristics important to consumers. Copyright © 2015 Elsevier Ltd. All rights reserved.
Kang, I S; Sams, A R
1999-01-01
In Experiment 1, 400 male broilers were stunned using a gradient of 40 to 60% CO2 over a period of 25 s or a 1% brine solution that was electrically charged (35 mA) for 7 s. Blood loss during bleeding was measured in 30-s intervals for a total of 120 s. After conventional processing and chilling, carcass damage was subjectively evaluated. Results indicated that the birds stunned with electricity bled faster than the CO2-stunned birds until 60 s. However, the cumulative blood loss was not different after 90 s. Carcass damage evaluation indicated that birds stunned with CO2 had a significantly lower percentage of broken clavicles, and had fewer hemorrhages on the surface of the Pectoralis. However, there was no difference between the two stunning methods in the frequency of damage at the shoulder. In Experiment 2, 256 broilers were stunned using the same conditions as in Experiment 1. Measurements of pH, R-value, sarcomere length (SL), and fragmentation index (FI) were evaluated from the left breast fillets harvested at 0, 1, 2, and 6 h postmortem. Shear values (SV) were determined using the right fillets harvested at the same four postmortem times and aged on ice until 24 h. No significant difference in breast muscle pH value was observed at 0, 2, and 6 h postmortem. However, CO2-stunned fillets had significantly higher pH values than the ES fillets at 1 h postmortem. Carbon dioxide produced greater R values than electricity at 2 and 6 h. Sarcomere length, FI, and SV were not significantly different at any time tested. These data suggest that CO2 stunning reduced carcass damage but did not reduce the need for aging before deboning when compared to the electrical stunning method used.
Püthe, Christoph; Manoj, Chandrasekharan; Kuvshinov, Alexey
2015-04-01
Electric fields induced in the conducting Earth during magnetic storms drive currents in power transmission grids, telecommunication lines or buried pipelines. These geomagnetically induced currents (GIC) can cause severe service disruptions. The prediction of GIC is thus of great importance for public and industry. A key step in the prediction of the hazard to technological systems during magnetic storms is the calculation of the geoelectric field. To address this issue for mid-latitude regions, we developed a method that involves 3-D modelling of induction processes in a heterogeneous Earth and the construction of a model of the magnetospheric source. The latter is described by low-degree spherical harmonics; its temporal evolution is derived from observatory magnetic data. Time series of the electric field can be computed for every location on Earth's surface. The actual electric field however is known to be perturbed by galvanic effects, arising from very local near-surface heterogeneities or topography, which cannot be included in the conductivity model. Galvanic effects are commonly accounted for with a real-valued time-independent distortion matrix, which linearly relates measured and computed electric fields. Using data of various magnetic storms that occurred between 2000 and 2003, we estimated distortion matrices for observatory sites onshore and on the ocean bottom. Strong correlations between modellings and measurements validate our method. The distortion matrix estimates prove to be reliable, as they are accurately reproduced for different magnetic storms. We further show that 3-D modelling is crucial for a correct separation of galvanic and inductive effects and a precise prediction of electric field time series during magnetic storms. Since the required computational resources are negligible, our approach is suitable for a real-time prediction of GIC. For this purpose, a reliable forecast of the source field, e.g. based on data from satellites
Methods for assessing Phytophthora ramorum chlamydospore germination
Joyce Eberhart; Elilzabeth Stamm; Jennifer Parke
2013-01-01
Germination of chlamydospores is difficult to accurately assess when chlamydospores are attached to remnants of supporting hyphae. We developed two approaches for closely observing and rigorously quantifying the frequency of chlamydospore germination in vitro. The plate marking and scanning method was useful for quantifying germination of large...
Most, Sebastian; Nowak, Wolfgang; Bijeljic, Branko
2015-04-01
Fickian transport in groundwater flow is the exception rather than the rule. Transport in porous media is frequently simulated via particle methods (i.e. particle tracking random walk (PTRW) or continuous time random walk (CTRW)). These methods formulate transport as a stochastic process of particle position increments. At the pore scale, geometry and micro-heterogeneities prohibit the commonly made assumption of independent and normally distributed increments to represent dispersion. Many recent particle methods seek to loosen this assumption. Hence, it is important to get a better understanding of the processes at pore scale. For our analysis we track the positions of 10.000 particles migrating through the pore space over time. The data we use come from micro CT scans of a homogeneous sandstone and encompass about 10 grain sizes. Based on those images we discretize the pore structure and simulate flow at the pore scale based on the Navier-Stokes equation. This flow field realistically describes flow inside the pore space and we do not need to add artificial dispersion during the transport simulation. Next, we use particle tracking random walk and simulate pore-scale transport. Finally, we use the obtained particle trajectories to do a multivariate statistical analysis of the particle motion at the pore scale. Our analysis is based on copulas. Every multivariate joint distribution is a combination of its univariate marginal distributions. The copula represents the dependence structure of those univariate marginals and is therefore useful to observe correlation and non-Gaussian interactions (i.e. non-Fickian transport). The first goal of this analysis is to better understand the validity regions of commonly made assumptions. We are investigating three different transport distances: 1) The distance where the statistical dependence between particle increments can be modelled as an order-one Markov process. This would be the Markovian distance for the process, where
Integrals of Frullani type and the method of brackets
Directory of Open Access Journals (Sweden)
Bravo Sergio
2017-01-01
Full Text Available The method of brackets is a collection of heuristic rules, some of which have being made rigorous, that provide a flexible, direct method for the evaluation of definite integrals. The present work uses this method to establish classical formulas due to Frullani which provide values of a specific family of integrals. Some generalizations are established.
Klahr, David
2013-08-20
Although the "science of science communication" usually refers to the flow of scientific knowledge from scientists to the public, scientists direct most of their communications not to the public, but instead to other scientists in their field. This paper presents a case study on this understudied type of communication: within a discipline, among its practitioners. I argue that many of the contentious disagreements that exist today in the field in which I conduct my research--early science education--derive from a lack of operational definitions, such that when competing claims are made for the efficacy of one type of science instruction vs. another, the arguments are hopelessly disjointed. The aim of the paper is not to resolve the current claims and counterclaims about the most effective pedagogies in science education, but rather to note that the assessment of one approach vs. the other is all too often defended on the basis of strongly held beliefs, rather than on the results of replicable experiments, designed around operational definitions of the teaching methods being investigated. A detailed example of operational definitions from my own research on elementary school science instruction is provided. In addition, the paper addresses the issue of how casual use of labels-both within the discipline and when communicating with the public-may inadvertently "undo" the benefits of operational definitions.
Directory of Open Access Journals (Sweden)
Osman Sevket Ipsiroglu
2015-03-01
Full Text Available Introduction Advanced video technology is available for sleep-laboratories. However, low-cost equipment for screening in the home setting has not been identified and tested, nor has a methodology for analysis of video recordings been suggested.MethodsWe investigated different combinations of hardware /software for home-videosomnography (HVS and established a process for qualitative and quantitative analysis of HVS-recordings. A case vignette (HVS analysis for a 5.5-year-old girl with major insomnia and several co-morbidities demonstrates how methodological considerations were addressed and how HVS added value to clinical assessment.ResultsWe suggest an ‘ideal set of hardware/software’ that is reliable, affordable (~$500 and portable (= 2.8kg to conduct non-invasive HVS, which allows time-lapse analyses. The equipment consists of a net-book, a camera with infra-red optics, and a video capture device. (1 We present an HVS-analysis protocol consisting of 3 steps of analysis at varying replay speeds: (a basic overview and classification at 16x normal speed; (b 2nd viewing and detailed descriptions at 4-8x normal speed, and (c viewing, listening, and in-depth descriptions at real-time speed. (2 We also present a custom software program that facilitates video analysis and note-taking (Annotator©, and Optical Flow software that automatically quantifies movement for internal quality control of the HVS-recording.The case vignette demonstrates how the HVS-recordings revealed the dimension of insomnia caused by Restless Legs Syndrome, and illustrated the cascade of symptoms, challenging behaviors and resulting medications.ConclusionThe strategy of using HVS, although requiring validation and reliability testing, opens the floor for a new ‘observational sleep medicine’, which has been useful in describing discomfort-related behavioral movement patterns in patients with communication difficulties presenting with challenging/disruptive sleep
Kobayashi, M; Takatori, T; Nakajima, M; Saka, K; Iwase, H; Nagao, M; Niijima, H; Matsuda, Y
1999-01-01
We examined the postmortem changes in the levels of ATP, glycogen and lactic acid in two masticatory muscles and three leg muscles of rats. The proportion of fibre types of the muscles was determined with NIH image software. The ATP levels in the white muscles did not decrease up to 1 h after death, and the ATP levels 1 and 2 h after death in the white muscles were higher than those in the red muscles with a single exception. The glycogen level at death and 1 h after death and the lactic acid level 1 h after death in masticatory muscles were lower than in the leg muscles. It is possible that the differences in the proportion of muscle fibre types and in glycogen level in muscles influences the postmortem change in ATP and lactic acid, which would accelerate or retard rigor mortis of the muscles.
Mendes, Joana Maia; Inoue, Luis Antonio Kioshi Aoki; Jesus, Rogério Souza
2015-01-01
ResumoO presente trabalho avaliou a influência do estresse pré-abate e do método de abate sobre o rigor mortis do tambaqui durante armazenamento em gelo. Foram estudadas respostas fisiológicas do tambaqui ao estresse durante o pré-abate, que foi dividido em quatro etapas: despesca, transporte, recuperação por 24 h e por 48 h. Ao final de cada etapa, os peixes foram amostrados para caracterização do estresse pré-abate por meio de análises dos parâmetros plasmáticos de glicose, lactato e amônia...
Imprecise Probability Methods for Weapons UQ
Energy Technology Data Exchange (ETDEWEB)
Picard, Richard Roy [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Vander Wiel, Scott Alan [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
2016-05-13
Building on recent work in uncertainty quanti cation, we examine the use of imprecise probability methods to better characterize expert knowledge and to improve on misleading aspects of Bayesian analysis with informative prior distributions. Quantitative approaches to incorporate uncertainties in weapons certi cation are subject to rigorous external peer review, and in this regard, certain imprecise probability methods are well established in the literature and attractive. These methods are illustrated using experimental data from LANL detonator impact testing.
Pençe, Halime Hanim; Pençe, Sadrettin; Kurtul, Naciye; Yilmaz, Necat; Kocoglu, Hasan; Bakan, Ebubekir
2003-01-01
In this study, adenosine triphosphate (ATP), adenosine diphosphate (ADP), adenosine monophosphate (AMP) and lactic acid in the muscles of masseter, triceps, and quadriceps obtained from right and left sides of Spraque-Dawley rats following death were investigated. The samples were taken immediately and 120 minutes after death occurred. The rats were killed either by cervical dislocation or drowning. ATP concentrations in the muscles of masseter, triceps, and quadriceps were lower in samples obtained 120 minutes after death than in those obtained immediately after death. ADP, AMP, and lactic acid concentrations in these muscles were higher in samples obtained 120 minutes after death than those obtained immediately after death. A positive linear correlation was determined between ATP and ADP concentrations in quadriceps muscles of the rats killed with cervical dislocation and in triceps muscles of the rats killed with drowning. When rats killed with cervical dislocation and with drowning were compared, ADP, AMP, and lactic acid concentrations were lower in the former than in the latter for both times (immediately and 120 minutes after death occurred). In the case of drowning, ATP is consumed faster because of hard exercise or severe physical activity, resulting in a faster rigor mortis. Higher lactic acid levels were determined in muscles of the rats killed with drowning than the other group. In the control and electric shock rats, ATP decreased in different levels in the three different muscle types mentioned above in control group, being much decline in masseter and then in quadriceps. This may be caused by lower mass and less glycogen storage of masseter. No different ATP levels were measured in drowning group with respect to the muscle type possibly because of the severe activity of triceps and quadriceps and because of smaller mass of masseter. One can conclude that the occurrence of rigor mortis is closely related to the mode of death.
Interpretation and method: Empirical research methods and the interpretive turn, 2nd ed.
Yanow, D.; Schwartz-Shea, P.
2014-01-01
This book demonstrates the relevance, rigor, and creativity of interpretive research methodologies for the social and human sciences. The book situates methods questions within the context of broader methodological questions--specifically, the character of social realities and their "know-ability."
The rigors of aligning performance
Hart, Andrew; Lucas, James
2015-01-01
Approved for public release; distribution is unlimited This Joint Applied Project addresses what can be done within the Naval Facilities Engineering Command Northwest community to better align its goals among competing interests from various stakeholders, while balancing the operational and regulatory constraints that often conflict with stakeholder goals and objectives. As a cross-functional organization, competing interests among the various business lines, support lines, and other stake...
The Rigors of Aligning Performance
2015-06-01
importance of the organization’s goals. To better align the commands goals with departmental goals, setting and continuously communicating goals and goal...which is vital to highlight the importance of the organization’s goals. To better align the commands goals with departmental goals, setting and...result of the 2004 organizational restructure, and as defined in the CONOPS, NAVFAC now operates as a matrix organization with integrated “vertical
Student Preferences for Instructional Methods in an Accounting Curriculum
Abeysekera, Indra
2015-01-01
Student preferences among instructional methods are largely unexplored across the accounting curriculum. The algorithmic rigor of courses and the societal culture can influence these preferences. This study explored students' preferences of instructional methods for learning in six courses of the accounting curriculum that differ in algorithmic…
Comprehensive methods for earlier detection and monitoring of forest decline
Jennifer Pontius; Richard Hallett
2014-01-01
Forested ecosystems are threatened by invasive pests, pathogens, and unusual climatic events brought about by climate change. Earlier detection of incipient forest health problems and a quantitatively rigorous assessment method is increasingly important. Here, we describe a method that is adaptable across tree species and stress agents and practical for use in the...
Darney, Blair G; Saavedra-Avendano, Biani; Lozano, Rafael
2017-01-01
A recent publication [Koch E, Chireau M, Pliego F, Stanford J, Haddad S, Calhoun B, Aracena P, Bravo M, Gatica S, Thorp J. Abortion legislation, maternal healthcare, fertility, female literacy, sanitation, violence against women and maternal deaths: a natural experiment in 32 Mexican states. BMJ Open 2015;5(2):e006013] claimed that Mexican states with more restrictive abortion laws had lower levels of maternal mortality. Our objectives are to replicate the analysis, reanalyze the data and offer a critique of the key flaws of the Koch study. We used corrected maternal mortality data (2006-2013), live births, and state-level indicators of poverty. We replicate the published analysis. We then reclassified state-level exposure to abortion on demand based on actual availability of abortion (Mexico City versus the other 31 states) and test the association of abortion access and the maternal mortality ratio (MMR) using descriptives over time, pooled chi-square tests and regression models. We included 256 state-year observations. We did not find significant differences in MMR between Mexico City (MMR=49.1) and the 31 states (MMR=44.6; p=.44). Using Koch's classification of states, we replicated published differences of higher MMR where abortion is more available. We found a significant, negative association between MMR and availability of abortion in the same multivariable models as Koch, but using our state classification (beta=-22.49, 95% CI=-38.9; -5.99). State-level poverty remains highly correlated with MMR. Koch makes errors in methodology and interpretation, making false causal claims about abortion law and MMR. MMR is falling most rapidly in Mexico City, but our main study limitation is an inability to draw causal inference about abortion law or access and maternal mortality. We need rigorous evidence about the health impacts of increasing access to safe abortion worldwide. Transparency and integrity in research is crucial, as well as perhaps even more in
O lugar de referência e o rigor do método no Jornalismo: algumas considerações
Directory of Open Access Journals (Sweden)
Alfredo Eurico Vizeu
2010-07-01
Full Text Available O objetivo deste artigo é propor algumas questões quanto ao jornalismo como um lugar de referência e a conseqüente necessidade do rigor no método de investigação. Procuramos analisar o jornalismo como um das instituições centrais na orientação do homem moderno, bem como problematizar a questão da investigação jornalística. Com base nos referenciais teóricos da cordialidade, da função pedagógica do Jornalismo e do método de apuração, seleção e produção da notícia, procuramos apontar algumas pistas que indicam as mudanças que vêm ocorrendo no campo jornalístico nesta já quase primeira década do século XXI.
Gomes, Carolina N; Passaglia, Jaqueline; Vilela, Felipe P; Pereira da Silva, Fátima M H S; Duque, Sheila S; Falcão, Juliana P
2018-08-01
Campylobacter spp. have been the most commonly reported gastrointestinal bacterial pathogen in many countries. Consumption of improperly prepared poultry meat has been the main transmission route of Campylobacter spp. Although Brazil is the largest exporter of poultry meat in the world, campylobacteriosis has been a neglected disease in the country. The aim of this study was to characterize 50 Campylobacter coli strains isolated from different sources in Brazil regarding the frequency of 16 virulence genes and their survival capability under five different stress conditions. All strains studied presented the cadF, flaA, and sodB genes that are considered essential for colonization. All strains grew at 4 °C and 37 °C after 24 h. High survival rates were observed when the strains were incubated in BHI with 7.5% NaCl and exposed to acid and oxidative stress. In conclusion, the pathogenic potential of the strains studied was reinforced by the presence of several important virulence genes and by the high growth and survival rates of the majority of those strains under different stress conditions. The results enabled a better understanding of strains circulating in Brazil and suggest that more rigorous control measures may be needed, given the importance of contaminated food as vehicles for Campylobacter coli. Copyright © 2018 Elsevier Ltd. All rights reserved.
Directory of Open Access Journals (Sweden)
Høyer Christian Bjerre
2016-06-01
Full Text Available A 63-year-old woman underwent a nephrectomy on the right side for renal cancer. Postoperatively she developed abdominal and lower back pain, which was treated with an injection of analgesics in an epidural catheter. The following morning it was discovered that the patient had cold legs with pallor and no palpable femoral pulse. Rigor mortis and livor mortis were diagnosed in both legs, even though the patient was still alive and awake. Doppler ultrasound examination revealed the absence of blood flow in the lower part of the abdominal aorta and distally. A cross disciplinary conference including specialists in urology, orthopaedics, vascular surgery, anaesthesiology, internal medicine, and intensive care concluded that no lifesaving treatment was possible, and the patient died the following day. A forensic autopsy revealed severe atherosclerosis with thrombosis and dissection of the abdominal aorta. This case clearly demonstrates that a vascular emergency should be considered when patients complain about pain in the lower back, abdomen or limbs. Clinicians should be especially aware of symptoms of tissue death that can be masked by epidural analgesia.
Hopkins, D L; Toohey, E S; Lamb, T A; Kerr, M J; van de Ven, R; Refshauge, G
2011-08-01
The temperature when the pH=6.0 (temp@pH6) impacts on the tenderness and eating quality of sheep meat. Due to the expense, sarcomere length is not routinely measured as a variable to explain variation in shear force, but whether measures such as temp@pH6 are as useful a parameter needs to be established. Measures of rigor onset in 261 carcases, including the temp@pH6, were evaluated in this study for their ability to explain some of the variation in shear force. The results show that for 1 day aged product combinations of the temp@pH6, the pH at 18 °C and the pH at 24 h provided a larger reduction (almost double) in total shear force variation than sarcomere length alone, with pH at 24 h being the single best measure. For 5 day aged product, pH at 18 °C was the single best measure. Inclusion of sarcomere length did represent some improvement, but the marginal increase would not be cost effective. Crown Copyright © 2011. Published by Elsevier Ltd. All rights reserved.
Wuolikainen, Anna; Jonsson, Pär; Ahnlund, Maria; Antti, Henrik; Marklund, Stefan L; Moritz, Thomas; Forsgren, Lars; Andersen, Peter M; Trupp, Miles
2016-04-01
Amyotrophic lateral sclerosis (ALS) and Parkinson's disease (PD) are protein-aggregation diseases that lack clear molecular etiologies. Biomarkers could aid in diagnosis, prognosis, planning of care, drug target identification and stratification of patients into clinical trials. We sought to characterize shared and unique metabolite perturbations between ALS and PD and matched controls selected from patients with other diagnoses, including differential diagnoses to ALS or PD that visited our clinic for a lumbar puncture. Cerebrospinal fluid (CSF) and plasma from rigorously age-, sex- and sampling-date matched patients were analyzed on multiple platforms using gas chromatography (GC) and liquid chromatography (LC)-mass spectrometry (MS). We applied constrained randomization of run orders and orthogonal partial least squares projection to latent structure-effect projections (OPLS-EP) to capitalize upon the study design. The combined platforms identified 144 CSF and 196 plasma metabolites with diverse molecular properties. Creatine was found to be increased and creatinine decreased in CSF of ALS patients compared to matched controls. Glucose was increased in CSF of ALS patients and α-hydroxybutyrate was increased in CSF and plasma of ALS patients compared to matched controls. Leucine, isoleucine and ketoleucine were increased in CSF of both ALS and PD. Together, these studies, in conjunction with earlier studies, suggest alterations in energy utilization pathways and have identified and further validated perturbed metabolites to be used in panels of biomarkers for the diagnosis of ALS and PD.
Davies, Philippa; Walker, Anne E; Grimshaw, Jeremy M
2010-02-09
There is growing interest in the use of cognitive, behavioural, and organisational theories in implementation research. However, the extent of use of theory in implementation research is uncertain. We conducted a systematic review of use of theory in 235 rigorous evaluations of guideline dissemination and implementation studies published between 1966 and 1998. Use of theory was classified according to type of use (explicitly theory based, some conceptual basis, and theoretical construct used) and stage of use (choice/design of intervention, process/mediators/moderators, and post hoc/explanation). Fifty-three of 235 studies (22.5%) were judged to have employed theories, including 14 studies that explicitly used theory. The majority of studies (n = 42) used only one theory; the maximum number of theories employed by any study was three. Twenty-five different theories were used. A small number of theories accounted for the majority of theory use including PRECEDE (Predisposing, Reinforcing, and Enabling Constructs in Educational Diagnosis and Evaluation), diffusion of innovations, information overload and social marketing (academic detailing). There was poor justification of choice of intervention and use of theory in implementation research in the identified studies until at least 1998. Future research should explicitly identify the justification for the interventions. Greater use of explicit theory to understand barriers, design interventions, and explore mediating pathways and moderators is needed to advance the science of implementation research.
Um exercício rigoroso de investigação clínica A rigorous exercise on clinical investigation
Directory of Open Access Journals (Sweden)
Angélica Bastos
2007-12-01
Full Text Available O artigo visa extrair da produção psicanalítica de Juan Carlos Cosentino as coordenadas de sua investigação. Com base no procedimento segundo o qual Freud concebeu a elaboração do saber em psicanálise, busca-se delimitar suas questões clínicas. Nas fobias, na angústia e nos sonhos, o psicanalista distingue duas ordens de laço na experiência analítica: por um lado, a fantasia e a neurose de transferência, e, por outro, a estrutura. Procura-se demonstrar o quanto sua releitura de Freud é movida por problemáticas originais, configurando uma pesquisa rigorosa impulsionada pela direção do tratamento e ditada pela função do desejo do analista.The article aims at drawing out the coordinates of investigation from the psychoanalytical production of Juan Carlos Cosentino. Based on the proceedings which Freud conceived knowledge elaboration in psychoanalysis, the clinical questions are delimited in his investigation. The psychoanalyst distinguishes two bond types in the analytical experience in relation to phobias, angst and dreams: on one hand, fantasy and transfer neurosis and on the other hand, the structure. We try to demonstrate how much his reading on Freud is powered by original problems, thus outlining a rigorous research triggered by the treatment course and ruled by the function of the therapist's desire.
Mixing quantitative with qualitative methods
DEFF Research Database (Denmark)
Morrison, Ann; Viller, Stephen; Heck, Tamara
2017-01-01
with or are considering, researching, or working with both quantitative and qualitative evaluation methods (in academia or industry), join us in this workshop. In particular, we look at adding quantitative to qualitative methods to build a whole picture of user experience. We see a need to discuss both quantitative...... and qualitative research because there is often a perceived lack of understanding of the rigor involved in each. The workshop will result in a White Paper on the latest developments in this field, within Australia and comparative with international work. We anticipate sharing submissions and workshop outcomes...
RCWA and FDTD modeling of light emission from internally structured OLEDs.
Callens, Michiel Koen; Marsman, Herman; Penninck, Lieven; Peeters, Patrick; de Groot, Harry; ter Meulen, Jan Matthijs; Neyts, Kristiaan
2014-05-05
We report on the fabrication and simulation of a green OLED with an Internal Light Extraction (ILE) layer. The optical behavior of these devices is simulated using both Rigorous Coupled Wave Analysis (RCWA) and Finite Difference Time-Domain (FDTD) methods. Results obtained using these two different techniques show excellent agreement and predict the experimental results with good precision. By verifying the validity of both simulation methods on the internal light extraction structure we pave the way to optimization of ILE layers using either of these methods.
Munch, F. D.; Grayver, A. V.; Kuvshinov, A.; Khan, A.
2018-01-01
In this paper we estimate and invert local electromagnetic (EM) sounding data for 1-D conductivity profiles in the presence of nonuniform oceans and continents to most rigorously account for the ocean induction effect that is known to strongly influence coastal observatories. We consider a new set of high-quality time series of geomagnetic observatory data, including hitherto unused data from island observatories installed over the last decade. The EM sounding data are inverted in the period range 3-85 days using stochastic optimization and model exploration techniques to provide estimates of model range and uncertainty. The inverted conductivity profiles are best constrained in the depth range 400-1,400 km and reveal significant lateral variations between 400 km and 1,000 km depth. To interpret the inverted conductivity anomalies in terms of water content and temperature, we combine laboratory-measured electrical conductivity of mantle minerals with phase equilibrium computations. Based on this procedure, relatively low temperatures (1200-1350°C) are observed in the transition zone (TZ) underneath stations located in Southern Australia, Southern Europe, Northern Africa, and North America. In contrast, higher temperatures (1400-1500°C) are inferred beneath observatories on islands, Northeast Asia, and central Australia. TZ water content beneath European and African stations is ˜0.05-0.1 wt %, whereas higher water contents (˜0.5-1 wt %) are inferred underneath North America, Asia, and Southern Australia. Comparison of the inverted water contents with laboratory-constrained water storage capacities suggests the presence of melt in or around the TZ underneath four geomagnetic observatories in North America and Northeast Asia.
Mixed Methods in Biomedical and Health Services Research
Curry, Leslie A.; Krumholz, Harlan M.; O’Cathain, Alicia; Plano Clark, Vicki L.; Cherlin, Emily; Bradley, Elizabeth H.
2013-01-01
Mixed methods studies, in which qualitative and quantitative methods are combined in a single program of inquiry, can be valuable in biomedical and health services research, where the complementary strengths of each approach can yield greater insight into complex phenomena than either approach alone. Although interest in mixed methods is growing among science funders and investigators, written guidance on how to conduct and assess rigorous mixed methods studies is not readily accessible to th...
Directory of Open Access Journals (Sweden)
Rafael dos Santos Costa
2011-01-01
Full Text Available O desenvolvimento do processo de rigor mortis nas carcaças dos animais de açougue influenciam diretamente a qualidade da carne. As características do processo de rigor mortis em carcaça de ovinos durante o processamento industrial para obtenção de carcaças resfriadas já foram estudadas em outros países e no Brasil em ovinos Santa Inês, mas ainda não estabelecidas em ovinos F1 Santa Inês x Dorper. Assim, objetivou-se neste trabalho caracterizar o processo de rigor mortis dos músculos Semitendinosus e Triceps brachii durante o resfriamento industrial e maciez da carne, em 10 carcaças ovinas. Foram escolhidos aleatoriamente 10 ovinos machos inteiros, sendo seis da raça Santa Inês e quatro F1 Santa Inês x Dorper, abatidos no Matadouro Frigorífico de Campos - Campos dos Goytacazes, Rio de Janeiro. Após a sangria, analisou-se temperatura, pH, comprimento de sarcômero em diferentes intervalos de tempo (4h; 6h; 8h; 10h; 12h; e 24h e força de cisalhamento ou maciez às 48h, do músculo Semitendinosus. Paralelamente, foi realizada a correlação entre a análise sensorial e a análise instrumental desse músculo. A temperatura da câmara fria variou de 12,2°C (4h a -0,5°C (24h e a temperatura média das carcaças foi de 26,80°C e -0,20°C, respectivamente. O pH médio inicial do músculo Semitendinosus foi de 6,62 e o final 5,64 enquanto no músculo T. brachii foi de 6,50 (4h e 5,68 (24h. A contração máxima do sarcômero do músculo Semitendinosus ocorreu na 12ªh(1,50mm após a sangria e no músculo Triceps brachii, no intervalo entre a 10ªh e 24ªh (1,53 a 1,57mm. No músculo Semitendinosus a força de cisalhamento ou maciez foi semelhante entre cordeiros da raça Santa Inês e F1 Santa Inês x Dorper, demonstrando que o grupo genético não influencia na maciez da carne. O painel sensorial confirmou os resultados obtidos na análise instrumental. Na correlação da análise instrumental (força de cisalhamento com a an
Application of statistical methods at copper wire manufacturing
Directory of Open Access Journals (Sweden)
Z. Hajduová
2009-01-01
Full Text Available Six Sigma is a method of management that strives for near perfection. The Six Sigma methodology uses data and rigorous statistical analysis to identify defects in a process or product, reduce variability and achieve as close to zero defects as possible. The paper presents the basic information on this methodology.
Building "Applied Linguistic Historiography": Rationale, Scope, and Methods
Smith, Richard
2016-01-01
In this article I argue for the establishment of "Applied Linguistic Historiography" (ALH), that is, a new domain of enquiry within applied linguistics involving a rigorous, scholarly, and self-reflexive approach to historical research. Considering issues of rationale, scope, and methods in turn, I provide reasons why ALH is needed and…
Spline methods for conversation equations
International Nuclear Information System (INIS)
Bottcher, C.; Strayer, M.R.
1991-01-01
The consider the numerical solution of physical theories, in particular hydrodynamics, which can be formulated as systems of conservation laws. To this end we briefly describe the Basis Spline and collocation methods, paying particular attention to representation theory, which provides discrete analogues of the continuum conservation and dispersion relations, and hence a rigorous understanding of errors and instabilities. On this foundation we propose an algorithm for hydrodynamic problems in which most linear and nonlinear instabilities are brought under control. Numerical examples are presented from one-dimensional relativistic hydrodynamics. 9 refs., 10 figs
Gold, A. U.; Ledley, T. S.; McCaffrey, M. S.; Buhr, S. M.; Manduca, C. A.; Niepold, F.; Fox, S.; Howell, C. D.; Lynds, S. E.
2010-12-01
The topic of climate change permeates all aspects of our society: the news, household debates, scientific conferences, etc. To provide students with accurate information about climate science and energy awareness, educators require scientifically and pedagogically robust teaching materials. To address this need, the NSF-funded Climate Literacy & Energy Awareness Network (CLEAN) Pathway has assembled a new peer-reviewed digital collection as part of the National Science Digital Library (NSDL) featuring teaching materials centered on climate and energy science for grades 6 through 16. The scope and framework of the collection is defined by the Essential Principles of Climate Science (CCSP 2009) and a set of energy awareness principles developed in the project. The collection provides trustworthy teaching materials on these socially relevant topics and prepares students to become responsible decision-makers. While a peer-review process is desirable for curriculum developer as well as collection builder to ensure quality, its implementation is non-trivial. We have designed a rigorous and transparent peer-review process for the CLEAN collection, and our experiences provide general guidelines that can be used to judge the quality of digital teaching materials across disciplines. Our multi-stage review process ensures that only resources with teaching goals relevant to developing climate literacy and energy awareness are considered. Each relevant resource is reviewed by two individuals to assess the i) scientific accuracy, ii) pedagogic effectiveness, and iii) usability/technical quality. A science review by an expert ensures the scientific quality and accuracy. Resources that pass all review steps are forwarded to a review panel of educators and scientists who make a final decision regarding inclusion of the materials in the CLEAN collection. Results from the first panel review show that about 20% (~100) of the resources that were initially considered for inclusion
Fusion, collapse, and stationary bound states of incoherently coupled waves in bulk cubic media
DEFF Research Database (Denmark)
Bang, Ole; Bergé, L.; Juul Rasmussen, Jens
1999-01-01
these sufficient conditions numerically and show that only when the equations and the initial conditions are symmetric are they also close to bring necessary conditions. Using Gaussian initial conditions we predict and confirm numerically the power dependent characteristic initial separations that divide the phase...
An operational coupled wave-current forecasting system for the northern Adriatic Sea
Russo, A.; Coluccelli, A.; Deserti, M.; Valentini, A.; Benetazzo, A.; Carniel, S.
2012-04-01
Since 2005 an Adriatic implementation of the Regional Ocean Modeling System (AdriaROMS) is being producing operational short-term forecasts (72 hours) of some hydrodynamic properties (currents, sea level, temperature, salinity) of the Adriatic Sea at 2 km horizontal resolution and 20 vertical s-levels, on a daily basis. The main objective of AdriaROMS, which is managed by the Hydro-Meteo-Clima Service (SIMC) of ARPA Emilia Romagna, is to provide useful products for civil protection purposes (sea level forecasts, outputs to run other forecasting models as for saline wedge, oil spills and coastal erosion). In order to improve the forecasts in the coastal area, where most of the attention is focused, a higher resolution model (0.5 km, again with 20 vertical s-levels) has been implemented for the northern Adriatic domain. The new implementation is based on the Coupled-Ocean-Atmosphere-Wave-Sediment Transport Modeling System (COAWST)and adopts ROMS for the hydrodynamic and Simulating WAve Nearshore (SWAN) for the wave module, respectively. Air-sea fluxes are computed using forecasts produced by the COSMO-I7 operational atmospheric model. At the open boundary of the high resolution model, temperature, salinity and velocity fields are provided by AdriaROMS while the wave characteristics are provided by an operational SWAN implementation (also managed by SIMC). Main tidal components are imposed as well, derived from a tidal model. Work in progress is oriented now on the validation of model results by means of extensive comparisons with acquired hydrographic measurements (such as CTDs or XBTs from sea-truth campaigns), currents and waves acquired at observational sites (including those of SIMC, CNR-ISMAR network and its oceanographic tower, located off the Venice littoral) and satellite-derived wave-heights data. Preliminary results on the forecast waves denote how, especially during intense storms, the effect of coupling can lead to significant variations in the wave heights. Part of the activity has been funded by the EU FP VII program (project "MICORE", contract n. 202798) and by the Regione Veneto regional law 15/2007 (Progetto "MARINA").
Energy-flux characterization of conical and space-time coupled wave packets
International Nuclear Information System (INIS)
Lotti, A.; Couairon, A.; Faccio, D.; Trapani, P. Di
2010-01-01
We introduce the concept of energy density flux as a characterization tool for the propagation of ultrashort laser pulses with spatiotemporal coupling. In contrast with calculations for the Poynting vector, those for energy density flux are derived in the local frame moving at the velocity of the envelope of the wave packet under examination and do not need knowledge of the magnetic field. We show that the energy flux defined from a paraxial propagation equation follows specific geometrical connections with the phase front of the optical wave packet, which demonstrates that the knowledge of the phase fronts amounts to the measurement of the energy flux. We perform a detailed numerical study of the energy density flux in the particular case of conical waves, with special attention paid to stationary-envelope conical waves (X or O waves). A full characterization of linear conical waves is given in terms of their energy flux. We extend the definition of this concept to the case of nonlinear propagation in Kerr media with nonlinear losses.
Coupled Wave Energy and Erosion Dynamics along a Salt Marsh Boundary, Hog Island Bay, Virginia, USA
Directory of Open Access Journals (Sweden)
Anthony M. Priestas
2015-09-01
Full Text Available The relationship between lateral erosion of salt marshes and wind waves is studied in Hog Island Bay, Virginia USA, with high-resolution field measurements and aerial photographs. Marsh retreat is compared to wave climate calculated in the bay using the spectral wave-model Simulating Waves Nearshore (SWAN. We confirm the existence of a linear relationship between long-term salt marsh erosion and wave energy, and show that wave power can serve as a good proxy for average salt-marsh erosion rates. At each site, erosion rates are consistent across several temporal scales, ranging from months to decades, and are strongly related to wave power. On the contrary, erosion rates vary in space and weakly depend on the spatial distribution of wave energy. We ascribe this variability to spatial variations in geotechnical, biological, and morphological marsh attributes. Our detailed field measurements indicate that at a small spatial scale (tens of meters, a positive feedback between salt marsh geometry and wave action causes erosion rates to increase with boundary sinuosity. However, at the scale of the entire marsh boundary (hundreds of meters, this relationship is reversed: those sites that are more rapidly eroding have a marsh boundary which is significantly smoother than the marsh boundary of sheltered and slowly eroding marshes.
Gasiewski, A. J.; Jackson, D. M.
1992-01-01
W-band measurements of the bistatic scattering function of some common microwave absorbing structures, including periodic wedge-type and pyramid-type iron-epoxy calibration loads and flat carbon-foam 'Echosorb' samples, were made using a network analyzer interface to a focused-lens scattering range. Swept frequency measurements over the 75-100 GHz band revealed specular and Bragg reflection characteristics in the measured data.
Ganju, Neil K.; Sherwood, Christopher R.
2010-01-01
A variety of algorithms are available for parameterizing the hydrodynamic bottom roughness associated with grain size, saltation, bedforms, and wave–current interaction in coastal ocean models. These parameterizations give rise to spatially and temporally variable bottom-drag coefficients that ostensibly provide better representations of physical processes than uniform and constant coefficients. However, few studies have been performed to determine whether improved representation of these variable bottom roughness components translates into measurable improvements in model skill. We test the hypothesis that improved representation of variable bottom roughness improves performance with respect to near-bed circulation, bottom stresses, or turbulence dissipation. The inner shelf south of Martha’s Vineyard, Massachusetts, is the site of sorted grain-size features which exhibit sharp alongshore variations in grain size and ripple geometry over gentle bathymetric relief; this area provides a suitable testing ground for roughness parameterizations. We first establish the skill of a nested regional model for currents, waves, stresses, and turbulent quantities using a uniform and constant roughness; we then gauge model skill with various parameterization of roughness, which account for the influence of the wave-boundary layer, grain size, saltation, and rippled bedforms. We find that commonly used representations of ripple-induced roughness, when combined with a wave–current interaction routine, do not significantly improve skill for circulation, and significantly decrease skill with respect to stresses and turbulence dissipation. Ripple orientation with respect to dominant currents and ripple shape may be responsible for complicating a straightforward estimate of the roughness contribution from ripples. In addition, sediment-induced stratification may be responsible for lower stresses than predicted by the wave–current interaction model.
Determination of resonance parameters in QCD by functional analysis methods
International Nuclear Information System (INIS)
Ciulli, S.; Geniet, F.; Papadopoulos, N.A.; Schilcher, K.
1988-01-01
A mathematically rigorous method based on functional analysis is used to determine resonance parameters of an amplitude from its given asymptotic expression in the space-like region. This method is checked on a model amplitude where both the asymptotic expression and the exact function are known. This method is then applied to the determination of the mass and the width of the ρ-meson from the corresponding space-like asymptotic QCD expression. (orig.)
Extension of the analytic nodal method to four energy groups
International Nuclear Information System (INIS)
Parsons, D.K.; Nigg, D.W.
1985-01-01
The Analytic Nodal Method is one of several recently-developed coarse mesh numerical methods for efficiently and accurately solving the multidimensional static and transient neutron diffusion equations. This summary describes a mathematically rigorous extension of the Analytic Nodal Method to the frequently more physically realistic four-group case. A few general theoretical considerations are discussed, followed by some calculated results for a typical steady-state two-dimensional PWR quarter core application. 8 refs
Ivankova, Nataliya V.
2014-01-01
In spite of recent methodological developments related to quality assurance in mixed methods research, practical examples of how to implement quality criteria in designing and conducting sequential QUAN [right arrow] QUAL mixed methods studies to ensure the process is systematic and rigorous remain scarce. This article discusses a three-step…
Bloom, Guillaume; Larat, Christian; Lallier, Eric; Lee-Bouhours, Mane-Si Laure; Loiseaux, Brigitte; Huignard, Jean-Pierre
2011-02-10
We have designed a high-efficiency array generator composed of subwavelength grooves etched in a GaAs substrate for operation at 4.5 μm. The method used combines rigorous coupled wave analysis with an optimization algorithm. The optimized beam splitter has both a high efficiency (∼96%) and a good intensity uniformity (∼0.2%). The fabrication error tolerances are numerically calculated, and it is shown that this subwavelength array generator could be fabricated with current electron beam writers and inductively coupled plasma etching. Finally, we studied the effect of a simple and realistic antireflection coating on the performance of the beam splitter.
Three-port beam splitter of a binary fused-silica grating.
Feng, Jijun; Zhou, Changhe; Wang, Bo; Zheng, Jiangjun; Jia, Wei; Cao, Hongchao; Lv, Peng
2008-12-10
A deep-etched polarization-independent binary fused-silica phase grating as a three-port beam splitter is designed and manufactured. The grating profile is optimized by use of the rigorous coupled-wave analysis around the 785 nm wavelength. The physical explanation of the grating is illustrated by the modal method. Simple analytical expressions of the diffraction efficiencies and modal guidelines for the three-port beam splitter grating design are given. Holographic recording technology and inductively coupled plasma etching are used to manufacture the fused-silica grating. Experimental results are in good agreement with the theoretical values.
Absolute analytical prediction of photonic crystal guided mode resonance wavelengths
DEFF Research Database (Denmark)
Hermannsson, Pétur Gordon; Vannahme, Christoph; Smith, Cameron
2014-01-01
numerically with methods such as rigorous coupled wave analysis. Here it is demonstrated how the absolute resonance wavelengths of such structures can be predicted by analytically modeling them as slab waveguides in which the propagation constant is determined by a phase matching condition. The model...... is experimentally verified to be capable of predicting the absolute resonance wavelengths to an accuracy of within 0.75 nm, as well as resonance wavelength shifts due to changes in cladding index within an accuracy of 0.45 nm across the visible wavelength regime in the case where material dispersion is taken...
Effects of water-absorption and thermal drift on a polymeric photonic crystal slab sensor
DEFF Research Database (Denmark)
Sørensen, Kristian Tølbøl; Ingvorsen, Charlotte Bonde; Nielsen, Line Hagner
2018-01-01
of resonant reflection during absorption, by monitoring the release of water using ellipsometry, and by rigorous coupled-wave analysis (RCWA). The approach presented here enables monitoring of water uptake and thermal fluctuations, for drift-free, high-performance operation of a polymeric PCS sensor....... with additional challenges, besides those relating to temperature-variations, which must be considered in any refractive index based method: The polymeric waveguide core was found to swell by ?0.3% as water absorbed into the waveguide core over ?1.5 h. This was investigated by monitoring the wavelength...
Parametric Studies on Artificial Morpho Butterfly Wing Scales for Optical Device Applications
Directory of Open Access Journals (Sweden)
Hyun Myung Kim
2015-01-01
Full Text Available We calculated diffraction efficiencies of grating structures inspired by Morpho butterfly wings by using a rigorous coupled-wave analysis method. The geometrical effects, such as grating width, period, thickness, and material index, were investigated in order to obtain better optical performance. Closely packed grating structures with an optimized membrane thickness show vivid reflected colors and provide high sensitivity to surrounding media variations, which is applicable to vapor sensing or healthcare indicators. Morpho structures with high index materials such as zinc sulfide or gallium phosphide generate white color caused by broadband reflection that can be used as reflected light sources for display applications.
Nonlinear optical properties of interconnected gold nanoparticles on silicon
Lesuffleur, Antoine; Gogol, Philippe; Beauvillain, Pierre; Guizal, B.; Van Labeke, D.; Georges, P.
2008-12-01
We report second harmonic generation (SHG) measurements in reflectivity from chains of gold nanoparticles interconnected with metallic bridges. We measured more than 30 times a SHG enhancement when a surface plasmon resonance was excited in the chains of nanoparticles, which was influenced by coupling due to the electrical connectivity of the bridges. This enhancement was confirmed by rigorous coupled wave method calculations and came from high localization of the electric field at the bridge. The introduction of 10% random defects into the chains of nanoparticles dropped the SHG by a factor of 2 and was shown to be very sensitive to the fundamental wavelength.
Formal Methods for Life-Critical Software
Butler, Ricky W.; Johnson, Sally C.
1993-01-01
The use of computer software in life-critical applications, such as for civil air transports, demands the use of rigorous formal mathematical verification procedures. This paper demonstrates how to apply formal methods to the development and verification of software by leading the reader step-by-step through requirements analysis, design, implementation, and verification of an electronic phone book application. The current maturity and limitations of formal methods tools and techniques are then discussed, and a number of examples of the successful use of formal methods by industry are cited.
Measuring Change of Capitalist Varieties: Reflections on Method, Illustrations from the BRICs
Becker, U.
2013-01-01
How to indicate institutional diversity and gradual change in capitalist political economies? What is the appropriate typological method, how could a suitable typology look? In the Weberian tradition, this article pleads for a method that rigorously distinguishes ideal types from empirical cases.
Pochhammer symbol with negative indices. A new rule for the method of brackets
Directory of Open Access Journals (Sweden)
Gonzalez Ivan
2016-01-01
Full Text Available The method of brackets is a method of integration based upon a small number of heuristic rules. Some of these have been made rigorous. An example of an integral involving the Bessel function is used to motivate a new evaluation rule.
A Review of Mixed Methods Research on Bullying and Peer Victimization in School
Hong, Jun Sung; Espelage, Dorothy L.
2012-01-01
Recognizing the negative outcomes associated with experiences in bullying at school, educational researchers, school officials, and policy-makers have called for more rigorous research on bullying in school. Research on bullying behavior in school has primarily been examined using quantitative methods. Mixed methods research in the field of…
Alternative methods in criticality
International Nuclear Information System (INIS)
Pedicini, J.M.
1982-01-01
In this thesis two new methods of calculating the criticality of a nuclear system are introduced and verified. Most methods of determining the criticality of a nuclear system depend implicitly upon knowledge of the angular flux, net currents, or moments of the angular flux, on the system surface in order to know the leakage. For small systems, leakage is the predominant element in criticality calculations. Unfortunately, in these methods the least accurate fluxes, currents, or moments are those occurring near system surfaces or interfaces. This is due to a mathematical inability to satisfy rigorously with a finite order angular polynomial expansion or angular difference technique the physical boundary conditions which occur on these surfaces. Consequently, one must accept large computational effort or less precise criticality calculations. The methods introduced in this thesis, including a direct leakage operator and an indirect multiple scattering leakage operator, obviate the need to know angular fluxes accurately at system boundaries. Instead, the system wide scalar flux, an integral quantity which is substantially easier to obtain with good precision is sufficient to obtain production, absorption, scattering, and leakage rates
Alternative methods in criticality
International Nuclear Information System (INIS)
Pedicini, J.M.
1982-01-01
Two new methods of calculating the criticality of a nuclear system are introduced and verified. Most methods of determining the criticality of a nuclear system depend implicitly upon knowledge of the angular flux, net currents, or moments of the angular flux, on the system surface in order to know the leakage. For small systems, leakage is the predominant element in criticality calculations. Unfortunately, in these methods the least accurate fluxes, currents, or moments are those occuring near system surfaces or interfaces. This is due to a mathematical inability to satisfy rigorously with a finite order angular polynomial expansion or angular difference technique the physical boundary conditions which occur on these surfaces. Consequently, one must accept large computational effort or less precise criticality calculations. The methods introduced in this thesis, including a direct leakage operator and an indirect multiple scattering leakage operator, obviate the need to know angular fluxes accurately at system boundaries. Instead, the system wide scalar flux, an integral quantity which is substantially easier to obtain with good precision, is sufficient to obtain production, absorption, scattering, and leakage rates
Robust statistical methods with R
Jureckova, Jana
2005-01-01
Robust statistical methods were developed to supplement the classical procedures when the data violate classical assumptions. They are ideally suited to applied research across a broad spectrum of study, yet most books on the subject are narrowly focused, overly theoretical, or simply outdated. Robust Statistical Methods with R provides a systematic treatment of robust procedures with an emphasis on practical application.The authors work from underlying mathematical tools to implementation, paying special attention to the computational aspects. They cover the whole range of robust methods, including differentiable statistical functions, distance of measures, influence functions, and asymptotic distributions, in a rigorous yet approachable manner. Highlighting hands-on problem solving, many examples and computational algorithms using the R software supplement the discussion. The book examines the characteristics of robustness, estimators of real parameter, large sample properties, and goodness-of-fit tests. It...
Directory of Open Access Journals (Sweden)
Juliana Paulino Vieira
2012-03-01
Full Text Available Este trabalho utilizou seis carcaças de jacaré-do-pantanal (Caiman crocodilus yacare com o objetivo de caracterizar o processo de rigor mortis do músculo Ílio-ischiocaudalis durante o resfriamento industrial e avaliar a maciez dessa carne. Os jacarés foram escolhidos aleatoriamente e abatidos na Cooperativa de Criadores do Jacaré do Pantanal (COOCRIJAPAN, Cáceres, Mato Grosso. Após a sangria, aferiu-se as variações das temperaturas da câmara de resfriamento, das carcaças e o pH. Foram colhidas amostras para determinação do comprimento de sarcômero, da força de cisalhamento e perdas por cozimento em diferentes intervalos de tempo (0,5, 3, 5, 7, 10, 12, 15, 24 e 36h. A temperatura da câmara de resfriamento variou de 2,6°C (0,5h a 0,9°C (36h e a temperatura média das carcaças variou de 21,0°C a 4,2°C, respectivamente. O pH médio inicial do músculo foi de 6,7 e o final 5,6 e a contração máxima do sarcômero do músculo Ilio-ischiocaudalis ocorreu na 15ª hora após a sangria (1,5µm. Essa carne apresentou força de cisalhamento menor que 6,0kg.This paper studied six pantanal alligators (Caiman crocodilus yacare carcass with goal of rigor mortis process characterization of Ilio-ischiocaudalis muscle during industrial cooling and meat tenderness. The alligators were randomly assembled and slaughtered at Cooperativa de Criadores do Jacaré do Pantanal (COOCRIJAPAN - Cáceres- Mato Grosso After exsanguination, were mensured temperature of chilling room and carcasses, pH and samples were collected for determination the sarcomere length, shear force and cooking loss at different times (0.5, 3, 5, 7, 10, 12, 15, 24 and 36 hours. The temperature of chilling room varied from 2.6°C (0.5h to 0.9°C (36h and the mean carcass temperature from 21.0°C to 4.2°C, respectively. The mean initial pH of the muscle was 6.7 and the final was 5.6. The smallest sarcomere size ocurred at 15 hours after exsanguination (1.5µm. This meat presents
Gas Analysis and Control Methods for Thermal Batteries
2013-09-01
when using highly efficient microporous thermal insulation packages. An easily implemented method of H2 gas removal from vendor thermal batteries is... microporous thermal insulation packages (1, 4, 5) or reduce volume requirements significantly. More rigorous gas control methods combined with...measured from the DCM pressures and known internal volumes of the 3 GHS that were measured using the ideal gas law with a 10-cc internal volume SS
Scattering theory in quantum mechanics. Physical principles and mathematical methods
International Nuclear Information System (INIS)
Amrein, W.O.; Jauch, J.M.; Sinha, K.B.
1977-01-01
A contemporary approach is given to the classical topics of physics. The purpose is to explain the basic physical concepts of quantum scattering theory, to develop the necessary mathematical tools for their description, to display the interrelation between the three methods (the Schroedinger equation solutions, stationary scattering theory, and time dependence) to derive the properties of various quantities of physical interest with mathematically rigorous methods
Improved radioanalytical methods
International Nuclear Information System (INIS)
Erickson, M.D.; Aldstadt, J.H.; Alvarado, J.S.; Crain, J.S.; Orlandini, K.A.; Smith, L.L.
1995-01-01
Methods for the chemical characterization of the environment are being developed under a multitask project for the Analytical Services Division (EM-263) within the US Department of Energy (DOE) Office of Environmental Management. This project focuses on improvement of radioanalytical methods with an emphasis on faster and cheaper routine methods. We have developed improved methods, for separation of environmental levels of technetium-99 and strontium-89/90, radium, and actinides from soil and water; and for separation of actinides from soil and water matrix interferences. Among the novel separation techniques being used are element- and class-specific resins and membranes. (The 3M Corporation is commercializing Empore trademark membranes under a cooperative research and development agreement [CRADA] initiated under this project). We have also developed methods for simultaneous detection of multiple isotopes using inductively coupled plasma-mass spectrometry (ICP-MS). The ICP-MS method requires less rigorous chemical separations than traditional radiochemical analyses because of its mass-selective mode of detection. Actinides and their progeny have been isolated and concentrated from a variety of natural water matrices by using automated batch separation incorporating selective resins prior to ICP-MS analyses. In addition, improvements in detection limits, sample volume, and time of analysis were obtained by using other sample introduction techniques, such as ultrasonic nebulization and electrothermal vaporization. Integration and automation of the separation methods with the ICP-MS methodology by using flow injection analysis is underway, with an objective of automating methods to achieve more reproducible results, reduce labor costs, cut analysis time, and minimize secondary waste generation through miniaturization of the process
Lectures on the nearest neighbor method
Biau, Gérard
2015-01-01
This text presents a wide-ranging and rigorous overview of nearest neighbor methods, one of the most important paradigms in machine learning. Now in one self-contained volume, this book systematically covers key statistical, probabilistic, combinatorial and geometric ideas for understanding, analyzing and developing nearest neighbor methods. Gérard Biau is a professor at Université Pierre et Marie Curie (Paris). Luc Devroye is a professor at the School of Computer Science at McGill University (Montreal). .
Methods of radon measurement and devices
International Nuclear Information System (INIS)
Miles, J.
2004-01-01
The following topics and instrumentation are discussed: The quantity to be measured; Active measurement methods (scintillation cells, ionisation chambers, electrostatic collection of decay products); Passive measurement methods (charcoal detectors; electret ion chambers; etched track detectors); and Detector considerations for large-scale surveys ('always on' or 'switchable' detectors?; response to radon-220; avoidance of electrostatic effects; quality assurance for passive radon detectors; quality control within the laboratory; external quality assurance; detectors need to be easily deliverable). It is concluded that the ideal detector for large scale surveys of radon in houses is a small, closed detector in a conducting holder which excludes radon-220, supported by rigorous quality assurance procedures. (P.A.)
Particle methods: An introduction with applications
Directory of Open Access Journals (Sweden)
Moral Piere Del
2014-01-01
Full Text Available Interacting particle methods are increasingly used to sample from complex high-dimensional distributions. They have found a wide range of applications in applied probability, Bayesian statistics and information engineering. Understanding rigorously these new Monte Carlo simulation tools leads to fascinating mathematics related to Feynman-Kac path integral theory and their interacting particle interpretations. In these lecture notes, we provide a pedagogical introduction to the stochastic modeling and the theoretical analysis of these particle algorithms. We also illustrate these methods through several applications including random walk confinements, particle absorption models, nonlinear filtering, stochastic optimization, combinatorial counting and directed polymer models.
Allen, Kathleen P.
2012-01-01
This mixed methods study explores text messaging in a suburban US high school. Survey questions were answered by students (mean age 16.0; SD = 1.23) regarding the prevalence of bullying and victimization via text messaging. Students and staff members responded to a survey item regarding perceptions of hostile text messaging. Both students and…
Numerical methods for metamaterial design
2013-01-01
This book describes a relatively new approach for the design of electromagnetic metamaterials. Numerical optimization routines are combined with electromagnetic simulations to tailor the broadband optical properties of a metamaterial to have predetermined responses at predetermined wavelengths. After a review of both the major efforts within the field of metamaterials and the field of mathematical optimization, chapters covering both gradient-based and derivative-free design methods are considered. Selected topics including surrogate-base optimization, adaptive mesh search, and genetic algorithms are shown to be effective, gradient-free optimization strategies. Additionally, new techniques for representing dielectric distributions in two dimensions, including level sets, are demonstrated as effective methods for gradient-based optimization. Each chapter begins with a rigorous review of the optimization strategy used, and is followed by numerous examples that combine the strategy with either electromag...
Akse, L; Birkeland, S; Tobiassen, T; Joensen, S; Larsen, R
2008-10-01
Processing of fish is generally conducted postrigor, but prerigor processing is associated with some potential advantages. The aim of this study was to study how 5 processing regimes of cold-smoked cod and salmon conducted at different stages of rigor influenced yield, fillet shrinkage, and gaping. Farmed cod and salmon was filleted, salted by brine injection of 25% NaCl, and smoked for 2 h at different stages of rigor. Filleting and salting prerigor resulted in increased fillet shrinkage and less increase in weight during brine injection, which in turn was correlated to the salt content of the fillet. These effects were more pronounced in cod fillets when compared to salmon. Early processing reduced fillet gaping and fillets were evaluated as having a firmer texture. In a follow-up trial with cod, shrinkage and weight gain during injection was studied as an effect of processing time postmortem. No changes in weight gain were observed for fillets salted the first 24 h postmortem; however, by delaying the processing 12 h postmortem, the high and rapid shrinking of cod fillets during brine injection was halved.
Permutation statistical methods an integrated approach
Berry, Kenneth J; Johnston, Janis E
2016-01-01
This research monograph provides a synthesis of a number of statistical tests and measures, which, at first consideration, appear disjoint and unrelated. Numerous comparisons of permutation and classical statistical methods are presented, and the two methods are compared via probability values and, where appropriate, measures of effect size. Permutation statistical methods, compared to classical statistical methods, do not rely on theoretical distributions, avoid the usual assumptions of normality and homogeneity of variance, and depend only on the data at hand. This text takes a unique approach to explaining statistics by integrating a large variety of statistical methods, and establishing the rigor of a topic that to many may seem to be a nascent field in statistics. This topic is new in that it took modern computing power to make permutation methods available to people working in the mainstream of research. This research monograph addresses a statistically-informed audience, and can also easily serve as a ...
Mathematical methods of many-body quantum field theory
Lehmann, Detlef
2004-01-01
Mathematical Methods of Many-Body Quantum Field Theory offers a comprehensive, mathematically rigorous treatment of many-body physics. It develops the mathematical tools for describing quantum many-body systems and applies them to the many-electron system. These tools include the formalism of second quantization, field theoretical perturbation theory, functional integral methods, bosonic and fermionic, and estimation and summation techniques for Feynman diagrams. Among the physical effects discussed in this context are BCS superconductivity, s-wave and higher l-wave, and the fractional quantum Hall effect. While the presentation is mathematically rigorous, the author does not focus solely on precise definitions and proofs, but also shows how to actually perform the computations.Presenting many recent advances and clarifying difficult concepts, this book provides the background, results, and detail needed to further explore the issue of when the standard approximation schemes in this field actually work and wh...
Abeysekera, Indra
2015-01-01
The role of work-integrated learning in student preferences of instructional methods is largely unexplored across the accounting curriculum. This study conducted six experiments to explore student preferences of instructional methods for learning, in six courses of the accounting curriculum that differed in algorithmic rigor, in the context of a…
A modified 4D ROOSTER method using the Chambolle-Pock algorithm
Mory, Cyril; Jacques, Laurent; The Third International Conference on Image Formation in X-Ray Computed Tomography
2014-01-01
The 4D RecOnstructiOn using Spatial and TEmpo- ral Regularization method is a recent 4D cone beam computed tomography algorithm. 4D ROOSTER has not been rigorously proved to converge. This paper aims to reformulate it using the Chambolle & Pock primal-dual optimization scheme. The convergence of this reformulated 4D ROOSTER is therefore guaranteed.
ANALYSIS OF IMPACT ON COMPOSITE STRUCTURES WITH THE METHOD OF DIMENSIONALITY REDUCTION
Directory of Open Access Journals (Sweden)
Valentin L. Popov
2015-04-01
Full Text Available In the present paper, we discuss the impact of rigid profiles on continua with non-local criteria for plastic yield. For the important case of media whose hardness is inversely proportional to the indentation radius, we suggest a rigorous treatment based on the method of dimensionality reduction (MDR and study the example of indentation by a conical profile.
Slayton, Julie; Samkian, Artineh
2017-01-01
We present our approach to a qualitative research methods course to prepare practitioner-scholars for their dissertation and independent research. We explain how an instructor's guide provides consistency and rigor, and in-class activities to scaffold learning, and helps faculty connect the content to students' out-of-school lives. We explain how…
Mixing Methods in Instructional Research. Forum: The Future of Instructional Communication
Meluch, Andrea L.
2017-01-01
"Communication Education's" commitment to publishing rigorous instructional communication research with a diversity of methods is clear in its publication of quantitative, qualitative, and critical articles (Witt, 2012). Here, Andrea Meluch argues that the trends in the most recent research published within this journal indicate a lack…
Riazi, A. Mehdi
2016-01-01
Mixed-methods research (MMR), as an inter-discourse (quantitative and qualitative) methodology, can provide applied linguistics researchers the opportunity to draw on and integrate the strengths of the two research methodological approaches in favour of making more rigorous inferences about research problems. In this article, the argument is made…
Fiber facet gratings for high power fiber lasers
Vanek, Martin; Vanis, Jan; Baravets, Yauhen; Todorov, Filip; Ctyroky, Jiri; Honzatko, Pavel
2017-12-01
We numerically investigated the properties of diffraction gratings designated for fabrication on the facet of an optical fiber. The gratings are intended to be used in high-power fiber lasers as mirrors either with a low or high reflectivity. The modal reflectance of low reflectivity polarizing grating has a value close to 3% for TE mode while it is significantly suppressed for TM mode. Such a grating can be fabricated on laser output fiber facet. The polarizing grating with high modal reflectance is designed as a leaky-mode resonant diffraction grating. The grating can be etched in a thin layer of high index dielectric which is sputtered on fiber facet. We used refractive index of Ta2O5 for such a layer. We found that modal reflectance can be close to 0.95 for TE polarization and polarization extinction ratio achieves 18 dB. Rigorous coupled wave analysis was used for fast optimization of grating parameters while aperiodic rigorous coupled wave analysis, Fourier modal method and finite difference time domain method were compared and used to compute modal reflectance of designed gratings.
Qualitative Methods in Mental Health Services Research
Palinkas, Lawrence A.
2014-01-01
Qualitative and mixed methods play a prominent role in mental health services research. However, the standards for their use are not always evident, especially for those not trained in such methods. This paper reviews the rationale and common approaches to using qualitative and mixed methods in mental health services and implementation research based on a review of the papers included in this special series along with representative examples from the literature. Qualitative methods are used to provide a “thick description” or depth of understanding to complement breadth of understanding afforded by quantitative methods, elicit the perspective of those being studied, explore issues that have not been well studied, develop conceptual theories or test hypotheses, or evaluate the process of a phenomenon or intervention. Qualitative methods adhere to many of the same principles of scientific rigor as quantitative methods, but often differ with respect to study design, data collection and data analysis strategies. For instance, participants for qualitative studies are usually sampled purposefully rather than at random and the design usually reflects an iterative process alternating between data collection and analysis. The most common techniques for data collection are individual semi-structured interviews, focus groups, document reviews, and participant observation. Strategies for analysis are usually inductive, based on principles of grounded theory or phenomenology. Qualitative methods are also used in combination with quantitative methods in mixed method designs for convergence, complementarity, expansion, development, and sampling. Rigorously applied qualitative methods offer great potential in contributing to the scientific foundation of mental health services research. PMID:25350675
A METHOD FOR EXERGY ANALYSIS OF SUGARCANE BAGASSE BOILERS
Directory of Open Access Journals (Sweden)
CORTEZ L.A.B.
1998-01-01
Full Text Available This work presents a method to conduct a thermodynamic analysis of sugarcane bagasse boilers. The method is based on the standard and actual reactions which allows the calculation of the enthalpies of each process subequation and the exergies of each of the main flowrates participating in the combustion. The method is presented using an example with real data from a sugarcane bagasse boiler. A summary of the results obtained is also presented together based on the 1st Law of Thermodynamics analysis, the exergetic efficiencies, and the irreversibility rates. The method presented is very rigorous with respect to data consistency, particularly for the flue gas composition.
Mathematical methods for cancer evolution
Suzuki, Takashi
2017-01-01
The purpose of this monograph is to describe recent developments in mathematical modeling and mathematical analysis of certain problems arising from cell biology. Cancer cells and their growth via several stages are of particular interest. To describe these events, multi-scale models are applied, involving continuously distributed environment variables and several components related to particles. Hybrid simulations are also carried out, using discretization of environment variables and the Monte Carlo method for the principal particle variables. Rigorous mathematical foundations are the bases of these tools. The monograph is composed of four chapters. The first three chapters are concerned with modeling, while the last one is devoted to mathematical analysis. The first chapter deals with molecular dynamics occurring at the early stage of cancer invasion. A pathway network model based on a biological scenario is constructed, and then its mathematical structures are determined. In the second chapter mathematica...
Discrete calculus methods for counting
Mariconda, Carlo
2016-01-01
This book provides an introduction to combinatorics, finite calculus, formal series, recurrences, and approximations of sums. Readers will find not only coverage of the basic elements of the subjects but also deep insights into a range of less common topics rarely considered within a single book, such as counting with occupancy constraints, a clear distinction between algebraic and analytical properties of formal power series, an introduction to discrete dynamical systems with a thorough description of Sarkovskii’s theorem, symbolic calculus, and a complete description of the Euler-Maclaurin formulas and their applications. Although several books touch on one or more of these aspects, precious few cover all of them. The authors, both pure mathematicians, have attempted to develop methods that will allow the student to formulate a given problem in a precise mathematical framework. The aim is to equip readers with a sound strategy for classifying and solving problems by pursuing a mathematically rigorous yet ...
Bartholomew, Theodore T; Lockard, Allison J
2018-06-13
Mixed methods can foster depth and breadth in psychological research. However, its use remains in development in psychotherapy research. Our purpose was to review the use of mixed methods in psychotherapy research. Thirty-one studies were identified via the PRISMA systematic review method. Using Creswell & Plano Clark's typologies to identify design characteristics, we assessed each study for rigor and how each used mixed methods. Key features of mixed methods designs and these common patterns were identified: (a) integration of clients' perceptions via mixing; (b) understanding group psychotherapy; (c) integrating methods with cases and small samples; (d) analyzing clinical data as qualitative data; and (e) exploring cultural identities in psychotherapy through mixed methods. The review is discussed with respect to the value of integrating multiple data in single studies to enhance psychotherapy research. © 2018 Wiley Periodicals, Inc.
Henke, Luke
2010-01-01
The ICARE method is a flexible, widely applicable method for systems engineers to solve problems and resolve issues in a complete and comprehensive manner. The method can be tailored by diverse users for direct application to their function (e.g. system integrators, design engineers, technical discipline leads, analysts, etc.). The clever acronym, ICARE, instills the attitude of accountability, safety, technical rigor and engagement in the problem resolution: Identify, Communicate, Assess, Report, Execute (ICARE). This method was developed through observation of Space Shuttle Propulsion Systems Engineering and Integration (PSE&I) office personnel approach in an attempt to succinctly describe the actions of an effective systems engineer. Additionally it evolved from an effort to make a broadly-defined checklist for a PSE&I worker to perform their responsibilities in an iterative and recursive manner. The National Aeronautics and Space Administration (NASA) Systems Engineering Handbook states, engineering of NASA systems requires a systematic and disciplined set of processes that are applied recursively and iteratively for the design, development, operation, maintenance, and closeout of systems throughout the life cycle of the programs and projects. ICARE is a method that can be applied within the boundaries and requirements of NASA s systems engineering set of processes to provide an elevated sense of duty and responsibility to crew and vehicle safety. The importance of a disciplined set of processes and a safety-conscious mindset increases with the complexity of the system. Moreover, the larger the system and the larger the workforce, the more important it is to encourage the usage of the ICARE method as widely as possible. According to the NASA Systems Engineering Handbook, elements of a system can include people, hardware, software, facilities, policies and documents; all things required to produce system-level results, qualities, properties, characteristics
Handbook of mathematical methods in imaging
2015-01-01
The Handbook of Mathematical Methods in Imaging provides a comprehensive treatment of the mathematical techniques used in imaging science. The material is grouped into two central themes, namely, Inverse Problems (Algorithmic Reconstruction) and Signal and Image Processing. Each section within the themes covers applications (modeling), mathematics, numerical methods (using a case example) and open questions. Written by experts in the area, the presentation is mathematically rigorous. This expanded and revised second edition contains updates to existing chapters and 16 additional entries on important mathematical methods such as graph cuts, morphology, discrete geometry, PDEs, conformal methods, to name a few. The entries are cross-referenced for easy navigation through connected topics. Available in both print and electronic forms, the handbook is enhanced by more than 200 illustrations and an extended bibliography. It will benefit students, scientists and researchers in applied mathematics. Engineers and com...
Solving hyperbolic equations with finite volume methods
Vázquez-Cendón, M Elena
2015-01-01
Finite volume methods are used in numerous applications and by a broad multidisciplinary scientific community. The book communicates this important tool to students, researchers in training and academics involved in the training of students in different science and technology fields. The selection of content is based on the author’s experience giving PhD and master courses in different universities. In the book the introduction of new concepts and numerical methods go together with simple exercises, examples and applications that contribute to reinforce them. In addition, some of them involve the execution of MATLAB codes. The author promotes an understanding of common terminology with a balance between mathematical rigor and physical intuition that characterizes the origin of the methods. This book aims to be a first contact with finite volume methods. Once readers have studied it, they will be able to follow more specific bibliographical references and use commercial programs or open source software withi...
Large Deviations and Asymptotic Methods in Finance
Gatheral, Jim; Gulisashvili, Archil; Jacquier, Antoine; Teichmann, Josef
2015-01-01
Topics covered in this volume (large deviations, differential geometry, asymptotic expansions, central limit theorems) give a full picture of the current advances in the application of asymptotic methods in mathematical finance, and thereby provide rigorous solutions to important mathematical and financial issues, such as implied volatility asymptotics, local volatility extrapolation, systemic risk and volatility estimation. This volume gathers together ground-breaking results in this field by some of its leading experts. Over the past decade, asymptotic methods have played an increasingly important role in the study of the behaviour of (financial) models. These methods provide a useful alternative to numerical methods in settings where the latter may lose accuracy (in extremes such as small and large strikes, and small maturities), and lead to a clearer understanding of the behaviour of models, and of the influence of parameters on this behaviour. Graduate students, researchers and practitioners will find th...
Power Mobility Training Methods for Children: A Systematic Review.
Kenyon, Lisa K; Hostnik, Lisa; McElroy, Rachel; Peterson, Courtney; Farris, John P
2018-01-01
To summarize and critically appraise the existing evidence related to power mobility training methods used in research studies conducted with children 21 years or younger. A systematic review was conducted using 16 electronic databases to identify primary source quantitative studies published in peer-reviewed journals. Data extraction, determination of level of evidence, evaluation of methodological rigor, and assessment of the risk of bias were completed. The Evidence Alert Traffic Light Grading System (EATLS) was used. Twenty-seven studies were included in the review. Levels of evidence were II to V; scientific rigor scores were 2 to 7. An overall Yellow EATLS level of evidence was found indicating that therapists should use caution when providing power mobility training interventions and measure outcomes related to established goals in areas such as development, functional skills, or use of a power mobility device.
Tomorrow's Research Library: Vigor or Rigor Mortis?
Hacken, Richard D.
1988-01-01
Compares, contrasts, and critiques predictions that have been made about the future of research libraries, focusing on the impact of technology on the library's role and users' needs. The discussion includes models for the adaptation of new technologies that may assist in library planning and change. (38 references) (CLB)
Thermodynamic limit and decoherence: rigorous results
Energy Technology Data Exchange (ETDEWEB)
Frasca, Marco [Via Erasmo Gattamelata 3, 00176 Rome (Italy)
2007-05-15
Time evolution operator in quantum mechanics can be changed into a statistical operator by a Wick rotation. This strict relation between statistical mechanics and quantum evolution can reveal deep results when the thermodynamic limit is considered. These results translate in a set of theorems proving that these effects can be effectively at work producing an emerging classical world without recurring to any external entity that in some cases cannot be properly defined. For a many-body system it has been recently shown that Gaussian decay of the coherence is the rule with a duration of recurrence more and more small as the number of particles increases. This effect has been observed experimentally. More generally, a theorem about coherence of bulk matter can be proved. All this takes us to the conclusion that a well defined boundary for the quantum to classical world does exist and that can be drawn by the thermodynamic limit, extending in this way the deep link between statistical mechanics and quantum evolution to a high degree.
Rigorous meaning of McLennan ensembles
Czech Academy of Sciences Publication Activity Database
Maes, C.; Netočný, Karel
2010-01-01
Roč. 51, č. 1 (2010), 015219/1-015219/16 ISSN 0022-2488 R&D Projects: GA ČR GA202/07/0404 Institutional research plan: CEZ:AV0Z10100520 Keywords : nonequilibrium ensembes * entropy production * driven diffusion Subject RIV: BE - Theoretical Physics Impact factor: 1.291, year: 2010 http://scitation.aip.org/getabs/servlet/GetabsServlet?prog=normal&id=JMAPAQ000051000001015219000001&idtype=cvips&gifs=yes
Promoting Rigorous Validation Practice: An Applied Perspective
Mattern, Krista D.; Kobrin, Jennifer L.; Camara, Wayne J.
2012-01-01
As researchers at a testing organization concerned with the appropriate uses and validity evidence for our assessments, we provide an applied perspective related to the issues raised in the focus article. Newton's proposal for elaborating the consensus definition of validity is offered with the intention to reduce the risks of inadequate…
The origins of Cauchy's rigorous calculus
Grabiner, Judith V
2005-01-01
This text examines the reinterpretation of calculus by Augustin-Louis Cauchy and his peers in the 19th century. These intellectuals created a collection of well-defined theorems about limits, continuity, series, derivatives, and integrals. 1981 edition.
Rigorous analysis of non-magnetic cloaks
DEFF Research Database (Denmark)
Zhang, Jingjing; Luo, Yu; Mortensen, Asger
2010-01-01
Nonmagnetic cloak offers a feasible way to achieve invisibility at optical frequencies using materials with only electric responses. In this letter, we suggest an approximation of the ideal nonmagnetic cloak and quantitatively study its electromagnetic characteristics using a full-wave scattering...... to the surrounding material at the outer boundary. Our analysis also provides the flexibility of reducing the scattering in an arbitrary direction....
CSIR Research Space (South Africa)
Theron, B
2013-04-01
Full Text Available A key imaging quality metric for an optical system (e.g. a camera or surveillance device) is its resolution. Methods to measure this, range from visual observation methods on resolution target charts, through more rigorous characterization, e...
Cutting-edge statistical methods for a life-course approach.
Bub, Kristen L; Ferretti, Larissa K
2014-01-01
Advances in research methods, data collection and record keeping, and statistical software have substantially increased our ability to conduct rigorous research across the lifespan. In this article, we review a set of cutting-edge statistical methods that life-course researchers can use to rigorously address their research questions. For each technique, we describe the method, highlight the benefits and unique attributes of the strategy, offer a step-by-step guide on how to conduct the analysis, and illustrate the technique using data from the National Institute of Child Health and Human Development Study of Early Child Care and Youth Development. In addition, we recommend a set of technical and empirical readings for each technique. Our goal was not to address a substantive question of interest but instead to provide life-course researchers with a useful reference guide to cutting-edge statistical methods.
An analytical method to identify and quantify trace levels of C5 to C12 perfluorocarboxylic acids (PFCAs) in articles of commerce (AOC) is developed and rigorously validated. Solid samples were extracted in methanol, and liquid samples were diluted with a solvent consisting of 60...
Tools for Reactive Distillation Column Design: Graphical and Stage-to-Stage Computation Methods
DEFF Research Database (Denmark)
Sanchez Daza, O.; Cisneros, Eduardo Salvador P.; Hostrup, Martin
2001-01-01
Based on the element mass balance concept, a graphical design method and a stage-to-stage multicomponent design method for reactive distillation columns have been developed. For distillation columns comprising reactive and non-reactive stages, a simple design strategy based on reactive and non......-reactive bubble point calculations is proposed. This strategy tracks the conversion and temperature between the feed and the end stages of the column. An illustrative example highlights the verification of the design strategy through rigorous simulation....
A perturbation method for dark solitons based on a complete set of the squared Jost solutions
International Nuclear Information System (INIS)
Ao Shengmei; Yan Jiaren
2005-01-01
A perturbation method for dark solitons is developed, which is based on the construction and the rigorous proof of the complete set of squared Jost solutions. The general procedure solving the adiabatic solution of perturbed nonlinear Schroedinger + equation, the time-evolution equation of dark soliton parameters and a formula for calculating the first-order correction are given. The method can also overcome the difficulties resulting from the non-vanishing boundary condition
Two Novel Methods and Multi-Mode Periodic Solutions for the Fermi-Pasta-Ulam Model
Arioli, Gianni; Koch, Hans; Terracini, Susanna
2005-04-01
We introduce two novel methods for studying periodic solutions of the FPU β-model, both numerically and rigorously. One is a variational approach, based on the dual formulation of the problem, and the other involves computer-assisted proofs. These methods are used e.g. to construct a new type of solutions, whose energy is spread among several modes, associated with closely spaced resonances.
Gandomani, Taghi Javdani; Zulzalil, Hazura; Ghani, Abdul Azim Abdul; Sultan, Abu Bakar Md
2013-01-01
Agile software development methods (ASD) and open source software development methods (OSSD) are two different approaches which were introduced in last decade and both of them have their fanatical advocators. Yet, it seems that relation and interface between ASD and OSSD is a fertile area and few rigorous studies have been done in this matter. Major goal of this study was assessment of the relation and integration of ASD and OSSD. Analyzing of collected data shows that ASD and OSSD are able t...
Computational methods for planning and evaluating geothermal energy projects
International Nuclear Information System (INIS)
Goumas, M.G.; Lygerou, V.A.; Papayannakis, L.E.
1999-01-01
In planning, designing and evaluating a geothermal energy project, a number of technical, economic, social and environmental parameters should be considered. The use of computational methods provides a rigorous analysis improving the decision-making process. This article demonstrates the application of decision-making methods developed in operational research for the optimum exploitation of geothermal resources. Two characteristic problems are considered: (1) the economic evaluation of a geothermal energy project under uncertain conditions using a stochastic analysis approach and (2) the evaluation of alternative exploitation schemes for optimum development of a low enthalpy geothermal field using a multicriteria decision-making procedure. (Author)
Conventional laboratory methods for cyanotoxins.
Lawton, Linda A; Edwards, C
2008-01-01
It is clear from the literature that numerous methods are available for most cyanotoxins, although many publications on monitoring data indicate that the favored approach is the use proven, robust methods for individual toxins. The most effective approach is the utilization of a robust rapid screen, where positive samples are followed up by qualitative and quantitative analysis to provide the essential decision making data needed for successful management strategies (Fig. 2). Currently, rapid screens are available for microcystins, saxitoxins and anatoxin-a(s), whilst optimisation and validation is needed, many publications report good correlation with the mouse bioassay and HPLC. There is an urgent need for rapid, simple, and inexpensive assays for cylindrospermopsins, anatoxin-a and BMAA. Although methods exist for analysis of BMAA, the fact that a recent study showed 95% of cyanobacteria producing this, some at levels > 6,000 microg g(-1) dry wt, is of concern and rapid screening followed by robust analysis needed. An ideal approach would be a single method capable of extracting and detecting all cyanotoxins. Several publications describe such approaches using LC-MS, but as expected from a group of compounds with diverse chemistry, there are obvious limitations in recoveries during sample processing, chromatographic performance and sensitivity (Dahlmann et al. 2003, Dell'Aversano et al. 2004, Pietsch et al. 2001). Selection of methods must be based on the application requirements, equipment available and cost. For many organisations it may be more cost effective to out-source the occasional analysis. However, as the incidence of blooms appears to be increasing, the need for more rigorous monitoring is needed, sensible investment is needed to meet recommended guidelines. Most of the methods discussed in this paper are suitable for achieving this goal, although clean-up and concentration is usually necessary for physicochemical methods.
Influence of Four-Wave Mixing and Walk-Off on the Self-Focusing of Coupled Waves
DEFF Research Database (Denmark)
Bergé, L.; Bang, Ole; Krolikowski, W.
2000-01-01
Four-wave mixing and walk-off between two optical beams are! investigated For focusing Kerr media. It is shown that four-wave mixing reinforces the self-focusing of mutually trapped waves by lowering their power threshold for collapse, only when their phase mismatch is small. On the contrary, walk......-off inhibits the collapse by detrapping the beams, whose partial centroids experience nonlinear oscillations....
Coupled wave and surge modelling for the eastern Irish Sea and implications for model wind-stress
Brown, Jennifer M.; Wolf, Judith
2009-05-01
We revisit the surge of November 1977, a storm event which caused damage on the Sefton coast in NW England. A hindcast has been made with a coupled surge-tide-wave model, to investigate whether a wave-dependent surface drag is necessary for accurate surge prediction, and also if this can be represented by an optimised Charnock parameter. The Proudman Oceanographic Laboratory Coastal Modelling System-Wave Model (POLCOMS-WAM) has been used to model combined tides, surges, waves and wave-current interaction in the Irish Sea on a 1.85 km grid. This period has been previously thoroughly studied, e.g. Jones and Davies [Jones, J.E., Davies, A.M., 1998. Storm surge computations for the Irish Sea using a three-dimensional numerical model including wave-current interaction. Continental Shelf Research 18(2), 201-251] and we build upon this previous work to validate the POLCOMS-WAM model to test the accuracy of surge elevation predictions in the study area. A one-way nested approach has been set up from larger scale models to the Irish Sea model. It was demonstrated that (as expected) swell from the North Atlantic does not have a significant impact in the eastern Irish Sea. To capture the external surge generated outside of the Irish Sea a (1/9° by 1/6°) model extending beyond the continental shelf edge was run using the POLCOMS model for tide and surge. The model results were compared with tide gauge observations around the eastern Irish Sea. The model was tested with different wind-stress formulations including Smith and Banke [Smith, S.D., Banke, E.G., 1975. Variation of the surface drag coefficient with wind speed. Quarterly Journal of the Royal Meteorology Society, 101(429), 665-673] and Charnock [Charnock, H., 1955. Wind-stress on a water surface. Quarterly Journal of the Royal Meteorological Society, 81(350), 639-640]. In order to get a single parameterisation that works with wave-coupling, the wave-derived surface roughness length has been imposed in the surge model. One of the largest surge events that occurred at Liverpool in the last 10 years, in January 2007, has also been simulated for comparison to demonstrate the robust application of the coupled model.
Belkhatir, Zehor
2016-08-05
This paper deals with joint parameters and input estimation for coupled PDE-ODE system. The system consists of a damped wave equation and an infinite dimensional ODE. This model describes the spatiotemporal hemodynamic response in the brain and the objective is to characterize brain regions using functional Magnetic Resonance Imaging (fMRI) data. For this reason, we propose an adaptive estimator and prove the asymptotic convergence of the state, the unknown input and the unknown parameters. The proof is based on a Lyapunov approach combined with a priori identifiability assumptions. The performance of the proposed observer is illustrated through some simulation results.
Belkhatir, Zehor; Mechhoud, Sarra; Laleg-Kirati, Taous-Meriem
2016-01-01
This paper deals with joint parameters and input estimation for coupled PDE-ODE system. The system consists of a damped wave equation and an infinite dimensional ODE. This model describes the spatiotemporal hemodynamic response in the brain
Asymptotic methods in analysis
Bruijn, N G de
2010-01-01
An original, effective approach teaches by explaining worked examples in detail. ""Every step in the mathematical process is explained, its purpose and necessity made clear . . . the reader not only has no difficulty in following the rigorous proofs, but even turns to them with eager expectation."" - Nuclear Physics. 1981 edition.
Analysing the Logic and Rigor in the Process of Verification of HACCP Plan%论验证HACCP计划过程中的逻辑性和严密性
Institute of Scientific and Technical Information of China (English)
秦红
2013-01-01
HACCP体系是一种系统性的食品安全预防控制体系，现已经越来越受到出口食品加工企业的重视，在许多企业中广泛运用并且质量提升效果明显。但是随着HACCP的不断发展，无论国内外官方有无要求，正有越来越多的企业申请HACCP验证或认证。现在企业制订HACCP计划，基本是采用美国国家水产品HACCP培训和教育联盟编写的“HACCP教程”给出的模式，但这种模式的要求是非常严格的。笔者将从逻辑性和严密性两个方面论证此模式下的HACCP计划的验证。旨在帮助指导审核人员完善HACCP计划验证过程，确保HACCP计划的有效实施。%HACCP system is a systematic preventive food safety control system, nowadays,the export food processing enterprises pay more and more attention to it , which has been widely used in many enterprises and the quality improvement effect is obvious. But with the development of HACCP , more and more enterprises apply for HACCP verification or certification. Basically , the enterprise use the model of the"HACCP Guidance"writing by the SHA, to make its HACCP plan, however. This model's requirement is very strict. I will analyse the logic and rigor in the process of verification of HACCP plan. In order to help guiding the personnel perfecting the HACCP plan Verification process, ensure the effective implementation of the HACCP plan.
Methods for determining time of death.
Madea, Burkhard
2016-12-01
Medicolegal death time estimation must estimate the time since death reliably. Reliability can only be provided empirically by statistical analysis of errors in field studies. Determining the time since death requires the calculation of measurable data along a time-dependent curve back to the starting point. Various methods are used to estimate the time since death. The current gold standard for death time estimation is a previously established nomogram method based on the two-exponential model of body cooling. Great experimental and practical achievements have been realized using this nomogram method. To reduce the margin of error of the nomogram method, a compound method was developed based on electrical and mechanical excitability of skeletal muscle, pharmacological excitability of the iris, rigor mortis, and postmortem lividity. Further increasing the accuracy of death time estimation involves the development of conditional probability distributions for death time estimation based on the compound method. Although many studies have evaluated chemical methods of death time estimation, such methods play a marginal role in daily forensic practice. However, increased precision of death time estimation has recently been achieved by considering various influencing factors (i.e., preexisting diseases, duration of terminal episode, and ambient temperature). Putrefactive changes may be used for death time estimation in water-immersed bodies. Furthermore, recently developed technologies, such as H magnetic resonance spectroscopy, can be used to quantitatively study decompositional changes. This review addresses the gold standard method of death time estimation in forensic practice and promising technological and scientific developments in the field.
Experimental methods in behavioral teratology
Energy Technology Data Exchange (ETDEWEB)
Zbinden, G.
1981-09-01
Efforts are made to develop toxicological techniques with which new behavioral teratogens can be recognized. The review describes the most important experimental methods which are presently explored, and which are based on a rich body of knowledge accumulated by experimental psychologists. Most of the tests were developed with small animals, mostly with rats. They range from the rather straightforward determination of various reflexes to complex behavioral situations involving mechanical devices, operant conditioning techniques and procedures evaluating social behavior. In applying these methods in routine toxicology, it is important to remember, that many behavioral effects determined in newborn and adult animals are subtle. Moreover, they are influenced by a large variety of environmental factors affecting the health and the behavior of the mothers and of the offspring in the early and later phases of development. Therefore, the experiments must be conducted under highly standardized conditions and must be controlled rigorously. It is concluded that the best experimental strategy for the evaluation of potential behavioral teratogens is not yet established. Therefore, it would be premature to decide on a fixed protocol to be included in routine animal safety experiments for drugs and other chemical substances.
Designing a mixed methods study in primary care.
Creswell, John W; Fetters, Michael D; Ivankova, Nataliya V
2004-01-01
Mixed methods or multimethod research holds potential for rigorous, methodologically sound investigations in primary care. The objective of this study was to use criteria from the literature to evaluate 5 mixed methods studies in primary care and to advance 3 models useful for designing such investigations. We first identified criteria from the social and behavioral sciences to analyze mixed methods studies in primary care research. We then used the criteria to evaluate 5 mixed methods investigations published in primary care research journals. Of the 5 studies analyzed, 3 included a rationale for mixing based on the need to develop a quantitative instrument from qualitative data or to converge information to best understand the research topic. Quantitative data collection involved structured interviews, observational checklists, and chart audits that were analyzed using descriptive and inferential statistical procedures. Qualitative data consisted of semistructured interviews and field observations that were analyzed using coding to develop themes and categories. The studies showed diverse forms of priority: equal priority, qualitative priority, and quantitative priority. Data collection involved quantitative and qualitative data gathered both concurrently and sequentially. The integration of the quantitative and qualitative data in these studies occurred between data analysis from one phase and data collection from a subsequent phase, while analyzing the data, and when reporting the results. We recommend instrument-building, triangulation, and data transformation models for mixed methods designs as useful frameworks to add rigor to investigations in primary care. We also discuss the limitations of our study and the need for future research.
Mixed methods research in music therapy research.
Bradt, Joke; Burns, Debra S; Creswell, John W
2013-01-01
Music therapists have an ethical and professional responsibility to provide the highest quality care possible to their patients. Much of the time, high quality care is guided by evidence-based practice standards that integrate the most current, available research in making decisions. Accordingly, music therapists need research that integrates multiple ways of knowing and forms of evidence. Mixed methods research holds great promise for facilitating such integration. At this time, there have not been any methodological articles published on mixed methods research in music therapy. The purpose of this article is to introduce mixed methods research as an approach to address research questions relevant to music therapy practice. This article describes the core characteristics of mixed methods research, considers paradigmatic issues related to this research approach, articulates major challenges in conducting mixed methods research, illustrates four basic designs, and provides criteria for evaluating the quality of mixed methods articles using examples of mixed methods research from the music therapy literature. Mixed methods research offers unique opportunities for strengthening the evidence base in music therapy. Recommendations are provided to ensure rigorous implementation of this research approach.
Recent improvement of the resonance analysis methods
International Nuclear Information System (INIS)
Sirakov, I.; Lukyanov, A.
2000-01-01
By the use of a two-step method called Combined, the R-matrix Wigner-Eisenbud representation in the resonance reaction theory has been converted into other equivalent representations (parameterizations) of the collision matrix with Poles in E domain. Two of them called Capture Elimination (CE) and Reaction Elimination (RE) representation respectively, have energy independent parameters and are both rigorous and applicable. The CE representation is essentially a generalization of the Reich-Moore (RM) formalism. The RE representation, in turn, offers some distinct advantages when analyzing fissile nuclei. The latter does not require any approximation for the capture channels and does not need any assumption about the number of fission channels in contrast to the RM representation. Unlike the RM parameters the RE ones are uniquely determined for applications in the resonance analysis. When given in the RE representation, neutron cross sections of fissile nuclei in the resolved resonance region are presented through simple scalar expressions without the need of matrix inversion. Various computer codes have been developed to demonstrate the viability of the new method. The RM parameters of the fissile nuclei have been converted into equivalent RE parameters implying the RM assumptions (REFINE code). Conversely, the RE parameters have been converted into corresponding RM parameters when one fission channel is present and the RM parameter set is unique, e.g. Pu-239, J =1 (REVERSE code). To further enhance the flexibility of the proposed method the obtained RE parameters have been converted into equivalent Generalized Pole parameters (REFILE code), which are parameters of the rigorous pole expansion of the collision matrix in √E domain. equi valent sets of RM, RE and GP parameters of 239 Pu are given as an example. It has been pointed out that all the advantages of the newly proposed representation can be implemented through an independent evaluation of the RE resonance
Nayfeh, Ali H
2008-01-01
1. Introduction 1 2. Straightforward Expansions and Sources of Nonuniformity 23 3. The Method of Strained Coordinates 56 4. The Methods of Matched and Composite Asymptotic Expansions 110 5. Variation of Parameters and Methods of Averaging 159 6. The Method of Multiple Scales 228 7. Asymptotic Solutions of Linear Equations 308 References and Author Index 387 Subject Index 417
International Nuclear Information System (INIS)
Konecny, C.
1975-01-01
Two main methods of separation using the distillation method are given and evaluated, namely evaporation and distillation in carrier gas flow. Two basic apparatus are described for illustrating the methods used. The use of the distillation method in radiochemistry is documented by a number of examples of the separation of elements in elemental state, volatile halogenides and oxides. Tables give a survey of distillation methods used for the separation of the individual elements and give conditions under which this separation takes place. The suitability of the use of distillation methods in radiochemistry is discussed with regard to other separation methods. (L.K.)
Seismoelectric Effects based on Spectral-Element Method for Subsurface Fluid Characterization
Morency, C.
2017-12-01
Present approaches for subsurface imaging rely predominantly on seismic techniques, which alone do not capture fluid properties and related mechanisms. On the other hand, electromagnetic (EM) measurements add constraints on the fluid phase through electrical conductivity and permeability, but EM signals alone do not offer information of the solid structural properties. In the recent years, there have been many efforts to combine both seismic and EM data for exploration geophysics. The most popular approach is based on joint inversion of seismic and EM data, as decoupled phenomena, missing out the coupled nature of seismic and EM phenomena such as seismoeletric effects. Seismoelectric effects are related to pore fluid movements with respect to the solid grains. By analyzing coupled poroelastic seismic and EM signals, one can capture a pore scale behavior and access both structural and fluid properties.Here, we model the seismoelectric response by solving the governing equations derived by Pride and Garambois (1994), which correspond to Biot's poroelastic wave equations and Maxwell's electromagnetic wave equations coupled electrokinetically. We will show that these coupled wave equations can be numerically implemented by taking advantage of viscoelastic-electromagnetic mathematical equivalences. These equations will be solved using a spectral-element method (SEM). The SEM, in contrast to finite-element methods (FEM) uses high degree Lagrange polynomials. Not only does this allow the technique to handle complex geometries similarly to FEM, but it also retains exponential convergence and accuracy due to the use of high degree polynomials. Finally, we will discuss how this is a first step toward full coupled seismic-EM inversion to improve subsurface fluid characterization. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Simulations of Scatterometry Down to 22 nm Structure Sizes and Beyond with Special Emphasis on LER
Osten, W.; Ferreras Paz, V.; Frenner, K.; Schuster, T.; Bloess, H.
2009-09-01
In recent years, scatterometry has become one of the most commonly used methods for CD metrology. With decreasing structure size for future technology nodes, the search for optimized scatterometry measurement configurations gets more important to exploit maximum sensitivity. As widespread industrial scatterometry tools mainly still use a pre-set measurement configuration, there are still free parameters to improve sensitivity. Our current work uses a simulation based approach to predict and optimize sensitivity of future technology nodes. Since line edge roughness is getting important for such small structures, these imperfections of the periodic continuation cannot be neglected. Using fourier methods like e.g. rigorous coupled wave approach (RCWA) for diffraction calculus, nonperiodic features are hard to reach. We show that in this field certain types of fieldstitching methods show nice numerical behaviour and lead to useful results.
Finite element methods for viscous incompressible flows a guide to theory, practice, and algorithms
Gunzburger, Max D
2012-01-01
In this book, the author examines mathematical aspects of finite element methods for the approximate solution of incompressible flow problems. The principal goal is to present some of the important mathematical results that are relevant to practical computations. In so doing, useful algorithms are also discussed. Although rigorous results are stated, no detailed proofs are supplied; rather, the intention is to present these results so that they can serve as a guide for the selection and, in certain respects, the implementation of algorithms.
Knezevic, David; Patera, Anthony T.; Huynh, Dinh Bao Phuong
2010-01-01
We present a certified reduced basis (RB) method for the heat equation and wave equation. The critical ingredients are certified RB approximation of the Laplace transform; the inverse Laplace transform to develop the time-domain RB output approximation and rigorous error bound; a (Butterworth) filter in time to effect the necessary “modal” truncation; RB eigenfunction decomposition and contour integration for Offline–Online decomposition. We present numerical results to demonstrate the accura...
Academic Rigor or Academic Rigor Mortis? Supervising Dissertations Is Serious Business
Wright, Robin Redmon
2017-01-01
This reflection considers the importance of and responsibility to graduate research supervision through an examination of a published dissertation that has had significant influence on the country's current immigration debate. The author exhorts both graduate students and adult education faculty to insist on clearly stated theoretical and…