WorldWideScience

Sample records for regularly spaced intervals

  1. Interval matrices: Regularity generates singularity

    Czech Academy of Sciences Publication Activity Database

    Rohn, Jiří; Shary, S.P.

    2018-01-01

    Roč. 540, 1 March (2018), s. 149-159 ISSN 0024-3795 Institutional support: RVO:67985807 Keywords : interval matrix * regularity * singularity * P-matrix * absolute value equation * diagonally singilarizable matrix Subject RIV: BA - General Mathematics Impact factor: 0.973, year: 2016

  2. Diagrammatic methods in phase-space regularization

    International Nuclear Information System (INIS)

    Bern, Z.; Halpern, M.B.; California Univ., Berkeley

    1987-11-01

    Using the scalar prototype and gauge theory as the simplest possible examples, diagrammatic methods are developed for the recently proposed phase-space form of continuum regularization. A number of one-loop and all-order applications are given, including general diagrammatic discussions of the nogrowth theorem and the uniqueness of the phase-space stochastic calculus. The approach also generates an alternate derivation of the equivalence of the large-β phase-space regularization to the more conventional coordinate-space regularization. (orig.)

  3. Probability Distribution for Flowing Interval Spacing

    International Nuclear Information System (INIS)

    Kuzio, S.

    2001-01-01

    The purpose of this analysis is to develop a probability distribution for flowing interval spacing. A flowing interval is defined as a fractured zone that transmits flow in the Saturated Zone (SZ), as identified through borehole flow meter surveys (Figure 1). This analysis uses the term ''flowing interval spacing'' as opposed to fractured spacing, which is typically used in the literature. The term fracture spacing was not used in this analysis because the data used identify a zone (or a flowing interval) that contains fluid-conducting fractures but does not distinguish how many or which fractures comprise the flowing interval. The flowing interval spacing is measured between the midpoints of each flowing interval. Fracture spacing within the SZ is defined as the spacing between fractures, with no regard to which fractures are carrying flow. The Development Plan associated with this analysis is entitled, ''Probability Distribution for Flowing Interval Spacing'', (CRWMS M and O 2000a). The parameter from this analysis may be used in the TSPA SR/LA Saturated Zone Flow and Transport Work Direction and Planning Documents: (1) ''Abstraction of Matrix Diffusion for SZ Flow and Transport Analyses'' (CRWMS M and O 1999a) and (2) ''Incorporation of Heterogeneity in SZ Flow and Transport Analyses'', (CRWMS M and O 1999b). A limitation of this analysis is that the probability distribution of flowing interval spacing may underestimate the effect of incorporating matrix diffusion processes in the SZ transport model because of the possible overestimation of the flowing interval spacing. Larger flowing interval spacing results in a decrease in the matrix diffusion processes. This analysis may overestimate the flowing interval spacing because the number of fractures that contribute to a flowing interval cannot be determined from the data. Because each flowing interval probably has more than one fracture contributing to a flowing interval, the true flowing interval spacing could be

  4. Dimensional regularization in configuration space

    International Nuclear Information System (INIS)

    Bollini, C.G.; Giambiagi, J.J.

    1995-09-01

    Dimensional regularization is introduced in configuration space by Fourier transforming in D-dimensions the perturbative momentum space Green functions. For this transformation, Bochner theorem is used, no extra parameters, such as those of Feynman or Bogoliubov-Shirkov are needed for convolutions. The regularized causal functions in x-space have ν-dependent moderated singularities at the origin. They can be multiplied together and Fourier transformed (Bochner) without divergence problems. The usual ultraviolet divergences appear as poles of the resultant functions of ν. Several example are discussed. (author). 9 refs

  5. Time-Homogeneous Parabolic Wick-Anderson Model in One Space Dimension: Regularity of Solution

    OpenAIRE

    Kim, Hyun-Jung; Lototsky, Sergey V

    2017-01-01

    Even though the heat equation with random potential is a well-studied object, the particular case of time-independent Gaussian white noise in one space dimension has yet to receive the attention it deserves. The paper investigates the stochastic heat equation with space-only Gaussian white noise on a bounded interval. The main result is that the space-time regularity of the solution is the same for additive noise and for multiplicative noise in the Wick-It\\^o-Skorokhod interpretation.

  6. A Characterization of Strong Regularity of Interval Matrices

    Czech Academy of Sciences Publication Activity Database

    Rohn, Jiří

    2010-01-01

    Roč. 20, - (2010), s. 717-722 E-ISSN 1081-3810 R&D Projects: GA ČR GA201/09/1957; GA ČR GC201/08/J020 Institutional research plan: CEZ:AV0Z10300504 Keywords : interval matrix * strong regularity * spectral radius * matrix inequality * solvability Subject RIV: BA - General Mathematics Impact factor: 0.808, year: 2010 http://www.math.technion.ac.il/iic/ ela / ela -articles/articles/vol20_pp717-722.pdf

  7. Probability Distribution for Flowing Interval Spacing

    International Nuclear Information System (INIS)

    S. Kuzio

    2004-01-01

    Fracture spacing is a key hydrologic parameter in analyses of matrix diffusion. Although the individual fractures that transmit flow in the saturated zone (SZ) cannot be identified directly, it is possible to determine the fractured zones that transmit flow from flow meter survey observations. The fractured zones that transmit flow as identified through borehole flow meter surveys have been defined in this report as flowing intervals. The flowing interval spacing is measured between the midpoints of each flowing interval. The determination of flowing interval spacing is important because the flowing interval spacing parameter is a key hydrologic parameter in SZ transport modeling, which impacts the extent of matrix diffusion in the SZ volcanic matrix. The output of this report is input to the ''Saturated Zone Flow and Transport Model Abstraction'' (BSC 2004 [DIRS 170042]). Specifically, the analysis of data and development of a data distribution reported herein is used to develop the uncertainty distribution for the flowing interval spacing parameter for the SZ transport abstraction model. Figure 1-1 shows the relationship of this report to other model reports that also pertain to flow and transport in the SZ. Figure 1-1 also shows the flow of key information among the SZ reports. It should be noted that Figure 1-1 does not contain a complete representation of the data and parameter inputs and outputs of all SZ reports, nor does it show inputs external to this suite of SZ reports. Use of the developed flowing interval spacing probability distribution is subject to the limitations of the assumptions discussed in Sections 5 and 6 of this analysis report. The number of fractures in a flowing interval is not known. Therefore, the flowing intervals are assumed to be composed of one flowing zone in the transport simulations. This analysis may overestimate the flowing interval spacing because the number of fractures that contribute to a flowing interval cannot be

  8. Semisupervised Support Vector Machines With Tangent Space Intrinsic Manifold Regularization.

    Science.gov (United States)

    Sun, Shiliang; Xie, Xijiong

    2016-09-01

    Semisupervised learning has been an active research topic in machine learning and data mining. One main reason is that labeling examples is expensive and time-consuming, while there are large numbers of unlabeled examples available in many practical problems. So far, Laplacian regularization has been widely used in semisupervised learning. In this paper, we propose a new regularization method called tangent space intrinsic manifold regularization. It is intrinsic to data manifold and favors linear functions on the manifold. Fundamental elements involved in the formulation of the regularization are local tangent space representations, which are estimated by local principal component analysis, and the connections that relate adjacent tangent spaces. Simultaneously, we explore its application to semisupervised classification and propose two new learning algorithms called tangent space intrinsic manifold regularized support vector machines (TiSVMs) and tangent space intrinsic manifold regularized twin SVMs (TiTSVMs). They effectively integrate the tangent space intrinsic manifold regularization consideration. The optimization of TiSVMs can be solved by a standard quadratic programming, while the optimization of TiTSVMs can be solved by a pair of standard quadratic programmings. The experimental results of semisupervised classification problems show the effectiveness of the proposed semisupervised learning algorithms.

  9. Regularization methods in Banach spaces

    CERN Document Server

    Schuster, Thomas; Hofmann, Bernd; Kazimierski, Kamil S

    2012-01-01

    Regularization methods aimed at finding stable approximate solutions are a necessary tool to tackle inverse and ill-posed problems. Usually the mathematical model of an inverse problem consists of an operator equation of the first kind and often the associated forward operator acts between Hilbert spaces. However, for numerous problems the reasons for using a Hilbert space setting seem to be based rather on conventions than on an approprimate and realistic model choice, so often a Banach space setting would be closer to reality. Furthermore, sparsity constraints using general Lp-norms or the B

  10. Least square regularized regression in sum space.

    Science.gov (United States)

    Xu, Yong-Li; Chen, Di-Rong; Li, Han-Xiong; Liu, Lu

    2013-04-01

    This paper proposes a least square regularized regression algorithm in sum space of reproducing kernel Hilbert spaces (RKHSs) for nonflat function approximation, and obtains the solution of the algorithm by solving a system of linear equations. This algorithm can approximate the low- and high-frequency component of the target function with large and small scale kernels, respectively. The convergence and learning rate are analyzed. We measure the complexity of the sum space by its covering number and demonstrate that the covering number can be bounded by the product of the covering numbers of basic RKHSs. For sum space of RKHSs with Gaussian kernels, by choosing appropriate parameters, we tradeoff the sample error and regularization error, and obtain a polynomial learning rate, which is better than that in any single RKHS. The utility of this method is illustrated with two simulated data sets and five real-life databases.

  11. q-Space Upsampling Using x-q Space Regularization.

    Science.gov (United States)

    Chen, Geng; Dong, Bin; Zhang, Yong; Shen, Dinggang; Yap, Pew-Thian

    2017-09-01

    Acquisition time in diffusion MRI increases with the number of diffusion-weighted images that need to be acquired. Particularly in clinical settings, scan time is limited and only a sparse coverage of the vast q -space is possible. In this paper, we show how non-local self-similar information in the x - q space of diffusion MRI data can be harnessed for q -space upsampling. More specifically, we establish the relationships between signal measurements in x - q space using a patch matching mechanism that caters to unstructured data. We then encode these relationships in a graph and use it to regularize an inverse problem associated with recovering a high q -space resolution dataset from its low-resolution counterpart. Experimental results indicate that the high-resolution datasets reconstructed using the proposed method exhibit greater quality, both quantitatively and qualitatively, than those obtained using conventional methods, such as interpolation using spherical radial basis functions (SRBFs).

  12. Regular perturbations in a vector space with indefinite metric

    International Nuclear Information System (INIS)

    Chiang, C.C.

    1975-08-01

    The Klein space is discussed in connection with practical applications. Some lemmas are presented which are to be used for the discussion of regular self-adjoint operators. The criteria for the regularity of perturbed operators are given. (U.S.)

  13. Regular Generalized Star Star closed sets in Bitopological Spaces

    OpenAIRE

    K. Kannan; D. Narasimhan; K. Chandrasekhara Rao; R. Ravikumar

    2011-01-01

    The aim of this paper is to introduce the concepts of τ1τ2-regular generalized star star closed sets , τ1τ2-regular generalized star star open sets and study their basic properties in bitopological spaces.

  14. New Inner Product Quasilinear Spaces on Interval Numbers

    Directory of Open Access Journals (Sweden)

    Hacer Bozkurt

    2016-01-01

    Full Text Available Primarily we examine the new example of quasilinear spaces, namely, “IRn interval space.” We obtain some new theorems and results related to this new quasilinear space. After giving some new notions of quasilinear dependence-independence and basis on quasilinear functional analysis, we obtain some results on IRn interval space related to these concepts. Secondly, we present Is,Ic0,Il∞, and Il2 quasilinear spaces and we research some algebraic properties of these spaces. We obtain some new results and provide an important contribution to the improvement of quasilinear functional analysis.

  15. Optimal Embeddings of Distance Regular Graphs into Euclidean Spaces

    NARCIS (Netherlands)

    F. Vallentin (Frank)

    2008-01-01

    htmlabstractIn this paper we give a lower bound for the least distortion embedding of a distance regular graph into Euclidean space. We use the lower bound for finding the least distortion for Hamming graphs, Johnson graphs, and all strongly regular graphs. Our technique involves semidefinite

  16. Forty Necessary and Sufficient Conditions for Regularity of Interval Matrices: A survey

    Czech Academy of Sciences Publication Activity Database

    Rohn, Jiří

    2009-01-01

    Roč. 18, - (2009), s. 500-512 E-ISSN 1081-3810 R&D Projects: GA ČR GA201/09/1957; GA ČR GC201/08/J020 Institutional research plan: CEZ:AV0Z10300504 Keywords : interval matrix * regularity * singularity * necessary and sufficient condition * algorithm Subject RIV: BA - General Mathematics Impact factor: 0.892, year: 2009 http://www.math.technion.ac.il/iic/ ela / ela -articles/articles/vol18_pp500-512.pdf

  17. Some double sequence spaces of interval numbers defined by Orlicz function

    Directory of Open Access Journals (Sweden)

    Ayhan Esi

    2014-10-01

    Full Text Available In this paper we introduce some interval valued double sequence spaces defined by Orlicz function and study different properties of these spaces like inclusion relations, solidity, etc. We establish some inclusion relations among them. Also we introduce the concept of double statistical convergence for interval number sequences and give an inclusion relation between interval valued double sequence spaces.

  18. Restrictive metric regularity and generalized differential calculus in Banach spaces

    Directory of Open Access Journals (Sweden)

    Bingwu Wang

    2004-10-01

    Full Text Available We consider nonlinear mappings f:X→Y between Banach spaces and study the notion of restrictive metric regularity of f around some point x¯, that is, metric regularity of f from X into the metric space E=f(X. Some sufficient as well as necessary and sufficient conditions for restrictive metric regularity are obtained, which particularly include an extension of the classical Lyusternik-Graves theorem in the case when f is strictly differentiable at x¯ but its strict derivative ∇f(x¯ is not surjective. We develop applications of the results obtained and some other techniques in variational analysis to generalized differential calculus involving normal cones to nonsmooth and nonconvex sets, coderivatives of set-valued mappings, as well as first-order and second-order subdifferentials of extended real-valued functions.

  19. Total variation regularization in measurement and image space for PET reconstruction

    KAUST Repository

    Burger, M

    2014-09-18

    © 2014 IOP Publishing Ltd. The aim of this paper is to test and analyse a novel technique for image reconstruction in positron emission tomography, which is based on (total variation) regularization on both the image space and the projection space. We formulate our variational problem considering both total variation penalty terms on the image and on an idealized sinogram to be reconstructed from a given Poisson distributed noisy sinogram. We prove existence, uniqueness and stability results for the proposed model and provide some analytical insight into the structures favoured by joint regularization. For the numerical solution of the corresponding discretized problem we employ the split Bregman algorithm and extensively test the approach in comparison to standard total variation regularization on the image. The numerical results show that an additional penalty on the sinogram performs better on reconstructing images with thin structures.

  20. Regularity and predictability of human mobility in personal space.

    Directory of Open Access Journals (Sweden)

    Daniel Austin

    Full Text Available Fundamental laws governing human mobility have many important applications such as forecasting and controlling epidemics or optimizing transportation systems. These mobility patterns, studied in the context of out of home activity during travel or social interactions with observations recorded from cell phone use or diffusion of money, suggest that in extra-personal space humans follow a high degree of temporal and spatial regularity - most often in the form of time-independent universal scaling laws. Here we show that mobility patterns of older individuals in their home also show a high degree of predictability and regularity, although in a different way than has been reported for out-of-home mobility. Studying a data set of almost 15 million observations from 19 adults spanning up to 5 years of unobtrusive longitudinal home activity monitoring, we find that in-home mobility is not well represented by a universal scaling law, but that significant structure (predictability and regularity is uncovered when explicitly accounting for contextual data in a model of in-home mobility. These results suggest that human mobility in personal space is highly stereotyped, and that monitoring discontinuities in routine room-level mobility patterns may provide an opportunity to predict individual human health and functional status or detect adverse events and trends.

  1. Manifold-splitting regularization, self-linking, twisting, writhing numbers of space-time ribbons

    International Nuclear Information System (INIS)

    Tze, C.H.

    1988-01-01

    The authors present an alternative formulation of Polyakov's regularization of Gauss' integral formula for a single closed Feynman path. A key element in his proof of the D = 3 fermi-bose transmutations induced by topological gauge fields, this regularization is linked here with the existence and properties of a nontrivial topological invariant for a closed space ribbon. This self-linking coefficient, an integer, is the sum of two differential characteristics of the ribbon, its twisting and writhing numbers. These invariants form the basis for a physical interpretation of our regularization. Their connection to Polyakov's spinorization is discussed. The authors further generalize their construction to the self-linking, twisting and writhing of higher dimensional d = eta(odd) submanifolds in D = (2eta + 1) space-time

  2. Fast regularizing sequential subspace optimization in Banach spaces

    International Nuclear Information System (INIS)

    Schöpfer, F; Schuster, T

    2009-01-01

    We are concerned with fast computations of regularized solutions of linear operator equations in Banach spaces in case only noisy data are available. To this end we modify recently developed sequential subspace optimization methods in such a way that the therein employed Bregman projections onto hyperplanes are replaced by Bregman projections onto stripes whose width is in the order of the noise level

  3. Coordinate-invariant regularization

    International Nuclear Information System (INIS)

    Halpern, M.B.

    1987-01-01

    A general phase-space framework for coordinate-invariant regularization is given. The development is geometric, with all regularization contained in regularized DeWitt Superstructures on field deformations. Parallel development of invariant coordinate-space regularization is obtained by regularized functional integration of the momenta. As representative examples of the general formulation, the regularized general non-linear sigma model and regularized quantum gravity are discussed. copyright 1987 Academic Press, Inc

  4. Interval Management: Development and Implementation of an Airborne Spacing Concept

    Science.gov (United States)

    Barmore, Bryan E.; Penhallegon, William J.; Weitz, Lesley A.; Bone, Randall S.; Levitt, Ian; Flores Kriegsfeld, Julia A.; Arbuckle, Doug; Johnson, William C.

    2016-01-01

    Interval Management is a suite of ADS-B-enabled applications that allows the air traffic controller to instruct a flight crew to achieve and maintain a desired spacing relative to another aircraft. The flight crew, assisted by automation, manages the speed of their aircraft to deliver more precise inter-aircraft spacing than is otherwise possible, which increases traffic throughput at the same or higher levels of safety. Interval Management has evolved from a long history of research and is now seen as a core NextGen capability. With avionics standards recently published, completion of an Investment Analysis Readiness Decision by the FAA, and multiple flight tests planned, Interval Management will soon be part of everyday use in the National Airspace System. Second generation, Advanced Interval Management capabilities are being planned to provide a wider range of operations and improved performance and benefits. This paper briefly reviews the evolution of Interval Management and describes current development and deployment plans. It also reviews concepts under development as the next generation of applications.

  5. The space of colored interval exchange transformations with flips

    International Nuclear Information System (INIS)

    Zaw, Myint

    2002-04-01

    We study the space Cr(2h, c) of c-colored exchange transformations with flips on 2h-intervals. We describe its relation to the moduli space M g,c *c of non-orientable Riemann surfaces of genus g≥0 with one boundary curve and c≥0 extra points where g=h-c-1. (author)

  6. Approximation by Müntz spaces on positive intervals

    KAUST Repository

    Ait-Haddou, Rachid

    2013-11-01

    The so-called Bernstein operators were introduced by S.N. Bernstein in 1912 to give a constructive proof of Weierstrass\\' theorem. We show how to extend his result to Müntz spaces on positive intervals. © 2013 Académie des sciences.

  7. Approximation by Müntz spaces on positive intervals

    KAUST Repository

    Ait-Haddou, Rachid; Mazure, Marie Laurence

    2013-01-01

    The so-called Bernstein operators were introduced by S.N. Bernstein in 1912 to give a constructive proof of Weierstrass' theorem. We show how to extend his result to Müntz spaces on positive intervals. © 2013 Académie des sciences.

  8. Dimensional regularization in position space and a forest formula for regularized Epstein-Glaser renormalization

    Energy Technology Data Exchange (ETDEWEB)

    Keller, Kai Johannes

    2010-04-15

    The present work contains a consistent formulation of the methods of dimensional regularization (DimReg) and minimal subtraction (MS) in Minkowski position space. The methods are implemented into the framework of perturbative Algebraic Quantum Field Theory (pAQFT). The developed methods are used to solve the Epstein-Glaser recursion for the construction of time-ordered products in all orders of causal perturbation theory. A solution is given in terms of a forest formula in the sense of Zimmermann. A relation to the alternative approach to renormalization theory using Hopf algebras is established. (orig.)

  9. Dimensional regularization in position space and a forest formula for regularized Epstein-Glaser renormalization

    International Nuclear Information System (INIS)

    Keller, Kai Johannes

    2010-04-01

    The present work contains a consistent formulation of the methods of dimensional regularization (DimReg) and minimal subtraction (MS) in Minkowski position space. The methods are implemented into the framework of perturbative Algebraic Quantum Field Theory (pAQFT). The developed methods are used to solve the Epstein-Glaser recursion for the construction of time-ordered products in all orders of causal perturbation theory. A solution is given in terms of a forest formula in the sense of Zimmermann. A relation to the alternative approach to renormalization theory using Hopf algebras is established. (orig.)

  10. On rarely generalized regular fuzzy continuous functions in fuzzy topological spaces

    Directory of Open Access Journals (Sweden)

    Appachi Vadivel

    2016-11-01

    Full Text Available In this paper, we introduce the concept of rarely generalized regular fuzzy continuous functions in the sense of A.P. Sostak's and Ramadan is introduced. Some interesting properties and characterizations of them are investigated. Also, some applications to fuzzy compact spaces are established.

  11. On the necessary conditions of the regular minimum of the scale factor of the co-moving space

    International Nuclear Information System (INIS)

    Agakov, V.G.

    1980-01-01

    In the framework of homogeneous cosmologic model studied is the behaviour of the comoving space element volume filled with barotropous medium, deprived of energy fluxes. Presented are the necessary conditions at which a regular final minimum of the scale factor of the co-mowing space may take place. It is found that to carry out the above minimum at values of cosmological constant Λ <= 0 the presence of two from three anisotropy factors is necessary. Anisotropy of space deformation should be one of these factors. In case of Λ <= 0 the regular minimum is also possible if all three factors of anisotropy are equal to zero. However if none of the factors of Fsub(i), Asub(ik) anisotropy is equal to zero, the presence of deformation space anisotropy is necessary for final regular minimum appearance

  12. Neutrino stress tensor regularization in two-dimensional space-time

    International Nuclear Information System (INIS)

    Davies, P.C.W.; Unruh, W.G.

    1977-01-01

    The method of covariant point-splitting is used to regularize the stress tensor for a massless spin 1/2 (neutrino) quantum field in an arbitrary two-dimensional space-time. A thermodynamic argument is used as a consistency check. The result shows that the physical part of the stress tensor is identical with that of the massless scalar field (in the absence of Casimir-type terms) even though the formally divergent expression is equal to the negative of the scalar case. (author)

  13. Regularization in Hilbert space under unbounded operators and general source conditions

    International Nuclear Information System (INIS)

    Hofmann, Bernd; Mathé, Peter; Von Weizsäcker, Heinrich

    2009-01-01

    The authors study ill-posed equations with unbounded operators in Hilbert space. This setup has important applications, but only a few theoretical studies are available. First, the question is addressed and answered whether every element satisfies some general source condition with respect to a given self-adjoint unbounded operator. This generalizes a previous result from Mathé and Hofmann (2008 Inverse Problems 24 015009). The analysis then proceeds to error bounds for regularization, emphasizing some specific points for regularization under unbounded operators. The study finally reviews two examples within the light of the present study, as these are fractional differentiation and some Cauchy problems for the Helmholtz equation, both studied previously and in more detail by U Tautenhahn and co-authors

  14. Application of Interval Predictor Models to Space Radiation Shielding

    Science.gov (United States)

    Crespo, Luis G.; Kenny, Sean P.; Giesy,Daniel P.; Norman, Ryan B.; Blattnig, Steve R.

    2016-01-01

    This paper develops techniques for predicting the uncertainty range of an output variable given input-output data. These models are called Interval Predictor Models (IPM) because they yield an interval valued function of the input. This paper develops IPMs having a radial basis structure. This structure enables the formal description of (i) the uncertainty in the models parameters, (ii) the predicted output interval, and (iii) the probability that a future observation would fall in such an interval. In contrast to other metamodeling techniques, this probabilistic certi cate of correctness does not require making any assumptions on the structure of the mechanism from which data are drawn. Optimization-based strategies for calculating IPMs having minimal spread while containing all the data are developed. Constraints for bounding the minimum interval spread over the continuum of inputs, regulating the IPMs variation/oscillation, and centering its spread about a target point, are used to prevent data over tting. Furthermore, we develop an approach for using expert opinion during extrapolation. This metamodeling technique is illustrated using a radiation shielding application for space exploration. In this application, we use IPMs to describe the error incurred in predicting the ux of particles resulting from the interaction between a high-energy incident beam and a target.

  15. A remark on partial linear spaces of girth 5 with an application to strongly regular graphs

    NARCIS (Netherlands)

    Brouwer, A.E.; Neumaier, A.

    1988-01-01

    We derive a lower bound on the number of points of a partial linear space of girth 5. As an application, certain strongly regular graphs with=2 are ruled out by observing that the first subconstituents are partial linear spaces.

  16. A function space framework for structural total variation regularization with applications in inverse problems

    Science.gov (United States)

    Hintermüller, Michael; Holler, Martin; Papafitsoros, Kostas

    2018-06-01

    In this work, we introduce a function space setting for a wide class of structural/weighted total variation (TV) regularization methods motivated by their applications in inverse problems. In particular, we consider a regularizer that is the appropriate lower semi-continuous envelope (relaxation) of a suitable TV type functional initially defined for sufficiently smooth functions. We study examples where this relaxation can be expressed explicitly, and we also provide refinements for weighted TV for a wide range of weights. Since an integral characterization of the relaxation in function space is, in general, not always available, we show that, for a rather general linear inverse problems setting, instead of the classical Tikhonov regularization problem, one can equivalently solve a saddle-point problem where no a priori knowledge of an explicit formulation of the structural TV functional is needed. In particular, motivated by concrete applications, we deduce corresponding results for linear inverse problems with norm and Poisson log-likelihood data discrepancy terms. Finally, we provide proof-of-concept numerical examples where we solve the saddle-point problem for weighted TV denoising as well as for MR guided PET image reconstruction.

  17. On Landweber–Kaczmarz methods for regularizing systems of ill-posed equations in Banach spaces

    International Nuclear Information System (INIS)

    Leitão, A; Alves, M Marques

    2012-01-01

    In this paper, iterative regularization methods of Landweber–Kaczmarz type are considered for solving systems of ill-posed equations modeled (finitely many) by operators acting between Banach spaces. Using assumptions of uniform convexity and smoothness on the parameter space, we are able to prove a monotony result for the proposed method, as well as to establish convergence (for exact data) and stability results (in the noisy data case). (paper)

  18. Salt-body Inversion with Minimum Gradient Support and Sobolev Space Norm Regularizations

    KAUST Repository

    Kazei, Vladimir

    2017-05-26

    Full-waveform inversion (FWI) is a technique which solves the ill-posed seismic inversion problem of fitting our model data to the measured ones from the field. FWI is capable of providing high-resolution estimates of the model, and of handling wave propagation of arbitrary complexity (visco-elastic, anisotropic); yet, it often fails to retrieve high-contrast geological structures, such as salt. One of the reasons for the FWI failure is that the updates at earlier iterations are too smooth to capture the sharp edges of the salt boundary. We compare several regularization approaches, which promote sharpness of the edges. Minimum gradient support (MGS) regularization focuses the inversion on blocky models, even more than the total variation (TV) does. However, both approaches try to invert undesirable high wavenumbers in the model too early for a model of complex structure. Therefore, we apply the Sobolev space norm as a regularizing term in order to maintain a balance between sharp and smooth updates in FWI. We demonstrate the application of these regularizations on a Marmousi model, enriched by a chunk of salt. The model turns out to be too complex in some parts to retrieve its full velocity distribution, yet the salt shape and contrast are retrieved.

  19. Critical phenomena of regular black holes in anti-de Sitter space-time

    Energy Technology Data Exchange (ETDEWEB)

    Fan, Zhong-Ying [Peking University, Center for High Energy Physics, Beijing (China)

    2017-04-15

    In General Relativity, addressing coupling to a non-linear electromagnetic field, together with a negative cosmological constant, we obtain the general static spherical symmetric black hole solution with magnetic charges, which is asymptotic to anti-de Sitter (AdS) space-times. In particular, for a degenerate case the solution becomes a Hayward-AdS black hole, which is regular everywhere in the full space-time. The existence of such a regular black hole solution preserves the weak energy condition, while the strong energy condition is violated. We then derive the first law and the Smarr formula of the black hole solution. We further discuss its thermodynamic properties and study the critical phenomena in the extended phase space where the cosmological constant is treated as a thermodynamic variable as well as the parameter associated with the non-linear electrodynamics. We obtain many interesting results such as: the Maxwell equal area law in the P-V (or S-T) diagram is violated and consequently the critical point (T{sub *},P{sub *}) of the first order small-large black hole transition does not coincide with the inflection point (T{sub c},P{sub c}) of the isotherms; the Clapeyron equation describing the coexistence curve of the Van der Waals (vdW) fluid is no longer valid; the heat capacity at constant pressure is finite at the critical point; the various exponents near the critical point are also different from those of the vdW fluid. (orig.)

  20. Parameter choice in Banach space regularization under variational inequalities

    International Nuclear Information System (INIS)

    Hofmann, Bernd; Mathé, Peter

    2012-01-01

    The authors study parameter choice strategies for the Tikhonov regularization of nonlinear ill-posed problems in Banach spaces. The effectiveness of any parameter choice for obtaining convergence rates depends on the interplay of the solution smoothness and the nonlinearity structure, and it can be expressed concisely in terms of variational inequalities. Such inequalities are link conditions between the penalty term, the norm misfit and the corresponding error measure. The parameter choices under consideration include an a priori choice, the discrepancy principle as well as the Lepskii principle. For the convenience of the reader, the authors review in an appendix a few instances where the validity of a variational inequality can be established. (paper)

  1. Supersymmetric dimensional regularization

    International Nuclear Information System (INIS)

    Siegel, W.; Townsend, P.K.; van Nieuwenhuizen, P.

    1980-01-01

    There is a simple modification of dimension regularization which preserves supersymmetry: dimensional reduction to real D < 4, followed by analytic continuation to complex D. In terms of component fields, this means fixing the ranges of all indices on the fields (and therefore the numbers of Fermi and Bose components). For superfields, it means continuing in the dimensionality of x-space while fixing the dimensionality of theta-space. This regularization procedure allows the simple manipulation of spinor derivatives in supergraph calculations. The resulting rules are: (1) First do all algebra exactly as in D = 4; (2) Then do the momentum integrals as in ordinary dimensional regularization. This regularization procedure needs extra rules before one can say that it is consistent. Such extra rules needed for superconformal anomalies are discussed. Problems associated with renormalizability and higher order loops are also discussed

  2. Zeta-function regularization approach to finite temperature effects in Kaluza-Klein space-times

    International Nuclear Information System (INIS)

    Bytsenko, A.A.; Vanzo, L.; Zerbini, S.

    1992-01-01

    In the framework of heat-kernel approach to zeta-function regularization, in this paper the one-loop effective potential at finite temperature for scalar and spinor fields on Kaluza-Klein space-time of the form M p x M c n , where M p is p-dimensional Minkowski space-time is evaluated. In particular, when the compact manifold is M c n = H n /Γ, the Selberg tracer formula associated with discrete torsion-free group Γ of the n-dimensional Lobachevsky space H n is used. An explicit representation for the thermodynamic potential valid for arbitrary temperature is found. As a result a complete high temperature expansion is presented and the roles of zero modes and topological contributions is discussed

  3. Restricted Interval Valued Neutrosophic Sets and Restricted Interval Valued Neutrosophic Topological Spaces

    Directory of Open Access Journals (Sweden)

    Anjan Mukherjee

    2016-08-01

    Full Text Available In this paper we introduce the concept of restricted interval valued neutrosophic sets (RIVNS in short. Some basic operations and properties of RIVNS are discussed. The concept of restricted interval valued neutrosophic topology is also introduced together with restricted interval valued neutrosophic finer and restricted interval valued neutrosophic coarser topology. We also define restricted interval valued neutrosophic interior and closer of a restricted interval valued neutrosophic set. Some theorems and examples are cites. Restricted interval valued neutrosophic subspace topology is also studied.

  4. Regularization and renormalization of quantum field theory in curved space-time

    International Nuclear Information System (INIS)

    Bernard, C.; Duncan, A.

    1977-01-01

    It is proposed that field theories quantized in a curved space-time manifold can be conveniently regularized and renormalized with the aid of Pauli-Villars regulator fields. The method avoids the conceptual difficulties of covariant point-separation approaches, by starting always from a manifestly generally covariant action, and the technical limitations of the dimensional reqularization approach, which requires solution of the theory in arbitrary dimension in order to go beyond a weak-field expansion. An action is constructed which renormalizes the weak-field perturbation theory of a massive scalar field in two space-time dimensions--it is shown that the trace anomaly previously found in dimensional regularization and some point-separation calculations also arises in perturbation theory when the theory is Pauli-Villars regulated. One then studies a specific solvable two-dimensional model of a massive scalar field in a Robertson-Walker asymptotically flat universe. It is shown that the action previously considered leads, in this model, to a well defined finite expectation value for the stress-energy tensor. The particle production (less than 0 in/vertical bar/theta/sup mu nu/(x,t)/vertical bar/0 in greater than for t → + infinity) is computed explicitly. Finally, the validity of weak-field perturbation theory (in the appropriate range of parameters) is checked directly in the solvable model, and the trace anomaly computed in the asymptotic regions t→ +- infinity independently of any weak field approximation. The extension of the model to higher dimensions and the renormalization of interacting (scalar) field theories are briefly discussed

  5. The Accuracy of Remapping Irregularly Spaced Velocity Data onto a Regular Grid and the Computation of Vorticity

    National Research Council Canada - National Science Library

    Cohn, R

    1998-01-01

    .... This technique may be viewed as the molecular counterpart of PIV. To take advantage of standard data processing techniques, the MTV data need to be remapped onto a regular grid with a uniform spacing...

  6. The Accuracy of Remapping Irregularly Spaced Velocity Data onto a Regular Grid and the Computation of Vorticity

    National Research Council Canada - National Science Library

    Cohn, Richard

    1999-01-01

    .... This technique may be viewed as the molecular counterpart of PIV. To take advantage of standard data processing techniques, the MTV data need to be remapped onto a regular grid with a uniform spacing...

  7. Regularity of difference equations on Banach spaces

    CERN Document Server

    Agarwal, Ravi P; Lizama, Carlos

    2014-01-01

    This work introduces readers to the topic of maximal regularity for difference equations. The authors systematically present the method of maximal regularity, outlining basic linear difference equations along with relevant results. They address recent advances in the field, as well as basic semigroup and cosine operator theories in the discrete setting. The authors also identify some open problems that readers may wish to take up for further research. This book is intended for graduate students and researchers in the area of difference equations, particularly those with advance knowledge of and interest in functional analysis.

  8. On RC-spaces

    OpenAIRE

    Bielas, Wojciech; Plewik, Szymon

    2018-01-01

    Following Frink's characterization of completely regular spaces, we say that a regular T_1-space is an RC-space whenever the family of all regular open sets constitutes a regular normal base. Normal spaces are RC-spaces and there exist completely regular spaces which are not RC-spaces. So the question arises, which of the known examples of completely regular and not normal spaces are RC-spaces. We show that the Niemytzki plane and the Sorgenfrey plane are RC-spaces.

  9. Metric regularity and subdifferential calculus

    International Nuclear Information System (INIS)

    Ioffe, A D

    2000-01-01

    The theory of metric regularity is an extension of two classical results: the Lyusternik tangent space theorem and the Graves surjection theorem. Developments in non-smooth analysis in the 1980s and 1990s paved the way for a number of far-reaching extensions of these results. It was also well understood that the phenomena behind the results are of metric origin, not connected with any linear structure. At the same time it became clear that some basic hypotheses of the subdifferential calculus are closely connected with the metric regularity of certain set-valued maps. The survey is devoted to the metric theory of metric regularity and its connection with subdifferential calculus in Banach spaces

  10. Regularity criterion for solutions to the Navier Stokes equations in the whole 3D space based on two vorticity components

    Czech Academy of Sciences Publication Activity Database

    Guo, Z.; Kučera, P.; Skalák, Zdeněk

    2018-01-01

    Roč. 458, č. 1 (2018), s. 755-766 ISSN 0022-247X R&D Projects: GA ČR GA13-00522S Institutional support: RVO:67985874 Keywords : Navier Stokes equations * conditional regularity * regularity criteria * vorticity * Besov spaces * bony decomposition Subject RIV: BA - General Mathematics OBOR OECD: Fluids and plasma physics (including surface physics) Impact factor: 1.064, year: 2016

  11. Optokinetic stimulation modulates neglect for the number space: Evidence from mental number interval bisection

    Directory of Open Access Journals (Sweden)

    Konstantinos ePriftis

    2012-02-01

    Full Text Available Behavioral, neuropsychological, and neuroimaging data support the idea that numbers are represented along a mental number line (MNL, an analogical, visuo-spatial representation of number magnitude. The MNL is left-to-right oriented, with small numbers on the left and larger numbers on the right. Left neglect patients are impaired in processing the left side of the MNL and show a rightward deviation in the mental bisection of numerical intervals. In the present study we investigated the effects of optokinetic stimulation (OKS -a technique inducing spatial attention shifts by means of activation of the optokinetic nystagmus- on mental number interval bisection. One patient with left neglect following right hemisphere stroke (BG and four control patients with right hemisphere damage, but without neglect, performed the mental number interval bisection task in three experimental conditions of OKS: static, leftward, and rightward. In the static condition, BG misbisected to the right of the true midpoint. BG misbisected to the left following leftward OKS, but again to the right of the midpoint following rightward OKS. In contrast, the performance of controls was not significantly affected by the direction of OKS. We argue that shifts of visuospatial attention, induced by OKS, may affect the mental number interval bisection, suggesting the presence of an interaction between the processing of number magnitude and the processing of the perceptual space, in patients with neglect for the mental number space.

  12. Continuum-regularized quantum gravity

    International Nuclear Information System (INIS)

    Chan Huesum; Halpern, M.B.

    1987-01-01

    The recent continuum regularization of d-dimensional Euclidean gravity is generalized to arbitrary power-law measure and studied in some detail as a representative example of coordinate-invariant regularization. The weak-coupling expansion of the theory illustrates a generic geometrization of regularized Schwinger-Dyson rules, generalizing previous rules in flat space and flat superspace. The rules are applied in a non-trivial explicit check of Einstein invariance at one loop: the cosmological counterterm is computed and its contribution is included in a verification that the graviton mass is zero. (orig.)

  13. Optokinetic Stimulation Modulates Neglect for the Number Space: Evidence from Mental Number Interval Bisection

    Science.gov (United States)

    Priftis, Konstantinos; Pitteri, Marco; Meneghello, Francesca; Umiltà, Carlo; Zorzi, Marco

    2012-01-01

    Behavioral, neuropsychological, and neuroimaging data support the idea that numbers are represented along a mental number line (MNL), an analogical, visuospatial representation of number magnitude. The MNL is left-to-right oriented in Western cultures, with small numbers on the left and larger numbers on the right. Left neglect patients are impaired in the mental bisection of numerical intervals, with a bias toward larger numbers that are relatively to the right on the MNL. In the present study we investigated the effects of optokinetic stimulation (OKS) – a technique inducing visuospatial attention shifts by means of activation of the optokinetic nystagmus – on number interval bisection. One patient with left neglect following right-hemisphere stroke (BG) and four control patients with right-hemisphere damage, but without neglect, performed the number interval bisection task in three conditions of OKS: static, leftward, and rightward. In the static condition, BG misbisected to the right of the true midpoint. BG misbisected to the left following leftward OKS, and again to the right of the midpoint following rightward OKS. Moreover, the variability of BG’s performance was smaller following both leftward and rightward OKS, suggesting that the attentional bias induced by OKS reduced the “indifference zone” that is thought to underlie the length effect reported in bisection tasks. We argue that shifts of visuospatial attention, induced by OKS, may affect number interval bisection, thereby revealing an interaction between the processing of the perceptual space and the processing of the number space. PMID:22363280

  14. Self-calibration for lab-μCT using space-time regularized projection-based DVC and model reduction

    Science.gov (United States)

    Jailin, C.; Buljac, A.; Bouterf, A.; Poncelet, M.; Hild, F.; Roux, S.

    2018-02-01

    An online calibration procedure for x-ray lab-CT is developed using projection-based digital volume correlation. An initial reconstruction of the sample is positioned in the 3D space for every angle so that its projection matches the initial one. This procedure allows a space-time displacement field to be estimated for the scanned sample, which is regularized with (i) rigid body motions in space and (ii) modal time shape functions computed using model reduction techniques (i.e. proper generalized decomposition). The result is an accurate identification of the position of the sample adapted for each angle, which may deviate from the desired perfect rotation required for standard reconstructions. An application of this procedure to a 4D in situ mechanical test is shown. The proposed correction leads to a much improved tomographic reconstruction quality.

  15. Genus Ranges of 4-Regular Rigid Vertex Graphs.

    Science.gov (United States)

    Buck, Dorothy; Dolzhenko, Egor; Jonoska, Nataša; Saito, Masahico; Valencia, Karin

    2015-01-01

    A rigid vertex of a graph is one that has a prescribed cyclic order of its incident edges. We study orientable genus ranges of 4-regular rigid vertex graphs. The (orientable) genus range is a set of genera values over all orientable surfaces into which a graph is embedded cellularly, and the embeddings of rigid vertex graphs are required to preserve the prescribed cyclic order of incident edges at every vertex. The genus ranges of 4-regular rigid vertex graphs are sets of consecutive integers, and we address two questions: which intervals of integers appear as genus ranges of such graphs, and what types of graphs realize a given genus range. For graphs with 2 n vertices ( n > 1), we prove that all intervals [ a, b ] for all a genus ranges. For graphs with 2 n - 1 vertices ( n ≥ 1), we prove that all intervals [ a, b ] for all a genus ranges. We also provide constructions of graphs that realize these ranges.

  16. INTRINSIC TOPOLOGY AND REFINEMENT OF HUTTON UNIT INTERVAL

    Institute of Scientific and Technical Information of China (English)

    王国俊; 徐罗山

    1992-01-01

    This paper introduces the theory of continuous lattices to the study of the Hutton unit interval I(L). some theorems related to I(L) are pithily proved. A kind of intrinsic topologies is applied to refining the topology of I(L),and a new fuzzy unit interval,called the H(λ) unit interval,is defined.Based on the H(λ) unit interval the H(λ)-complete regularity is introduced.Also,the theory of. H(λ)-stone-ech compactifications is established

  17. On the regularity of mild solutions to complete higher order differential equations on Banach spaces

    Directory of Open Access Journals (Sweden)

    Nezam Iraniparast

    2015-09-01

    Full Text Available For the complete higher order differential equation u(n(t=Σk=0n-1Aku(k(t+f(t, t∈ R (* on a Banach space E, we give a new definition of mild solutions of (*. We then characterize the regular admissibility of a translation invariant subspace al M of BUC(R, E with respect to (* in terms of solvability of the operator equation Σj=0n-1AjXal Dj-Xal Dn = C. As application, almost periodicity of mild solutions of (* is proved.

  18. The existence and regularity of time-periodic solutions to the three-dimensional Navier–Stokes equations in the whole space

    International Nuclear Information System (INIS)

    Kyed, Mads

    2014-01-01

    The existence, uniqueness and regularity of time-periodic solutions to the Navier–Stokes equations in the three-dimensional whole space are investigated. We consider the Navier–Stokes equations with a non-zero drift term corresponding to the physical model of a fluid flow around a body that moves with a non-zero constant velocity. The existence of a strong time-periodic solution is shown for small time-periodic data. It is further shown that this solution is unique in a large class of weak solutions that can be considered physically reasonable. Finally, we establish regularity properties for any strong solution regardless of its size. (paper)

  19. Manifold Regularized Correlation Object Tracking

    OpenAIRE

    Hu, Hongwei; Ma, Bo; Shen, Jianbing; Shao, Ling

    2017-01-01

    In this paper, we propose a manifold regularized correlation tracking method with augmented samples. To make better use of the unlabeled data and the manifold structure of the sample space, a manifold regularization-based correlation filter is introduced, which aims to assign similar labels to neighbor samples. Meanwhile, the regression model is learned by exploiting the block-circulant structure of matrices resulting from the augmented translated samples over multiple base samples cropped fr...

  20. Physical model of dimensional regularization

    Energy Technology Data Exchange (ETDEWEB)

    Schonfeld, Jonathan F.

    2016-12-15

    We explicitly construct fractals of dimension 4-ε on which dimensional regularization approximates scalar-field-only quantum-field theory amplitudes. The construction does not require fractals to be Lorentz-invariant in any sense, and we argue that there probably is no Lorentz-invariant fractal of dimension greater than 2. We derive dimensional regularization's power-law screening first for fractals obtained by removing voids from 3-dimensional Euclidean space. The derivation applies techniques from elementary dielectric theory. Surprisingly, fractal geometry by itself does not guarantee the appropriate power-law behavior; boundary conditions at fractal voids also play an important role. We then extend the derivation to 4-dimensional Minkowski space. We comment on generalization to non-scalar fields, and speculate about implications for quantum gravity. (orig.)

  1. Hierarchical regular small-world networks

    International Nuclear Information System (INIS)

    Boettcher, Stefan; Goncalves, Bruno; Guclu, Hasan

    2008-01-01

    Two new networks are introduced that resemble small-world properties. These networks are recursively constructed but retain a fixed, regular degree. They possess a unique one-dimensional lattice backbone overlaid by a hierarchical sequence of long-distance links, mixing real-space and small-world features. Both networks, one 3-regular and the other 4-regular, lead to distinct behaviors, as revealed by renormalization group studies. The 3-regular network is planar, has a diameter growing as √N with system size N, and leads to super-diffusion with an exact, anomalous exponent d w = 1.306..., but possesses only a trivial fixed point T c = 0 for the Ising ferromagnet. In turn, the 4-regular network is non-planar, has a diameter growing as ∼2 √(log 2 N 2 ) , exhibits 'ballistic' diffusion (d w = 1), and a non-trivial ferromagnetic transition, T c > 0. It suggests that the 3-regular network is still quite 'geometric', while the 4-regular network qualifies as a true small world with mean-field properties. As an engineering application we discuss synchronization of processors on these networks. (fast track communication)

  2. Graph Regularized Auto-Encoders for Image Representation.

    Science.gov (United States)

    Yiyi Liao; Yue Wang; Yong Liu

    2017-06-01

    Image representation has been intensively explored in the domain of computer vision for its significant influence on the relative tasks such as image clustering and classification. It is valuable to learn a low-dimensional representation of an image which preserves its inherent information from the original image space. At the perspective of manifold learning, this is implemented with the local invariant idea to capture the intrinsic low-dimensional manifold embedded in the high-dimensional input space. Inspired by the recent successes of deep architectures, we propose a local invariant deep nonlinear mapping algorithm, called graph regularized auto-encoder (GAE). With the graph regularization, the proposed method preserves the local connectivity from the original image space to the representation space, while the stacked auto-encoders provide explicit encoding model for fast inference and powerful expressive capacity for complex modeling. Theoretical analysis shows that the graph regularizer penalizes the weighted Frobenius norm of the Jacobian matrix of the encoder mapping, where the weight matrix captures the local property in the input space. Furthermore, the underlying effects on the hidden representation space are revealed, providing insightful explanation to the advantage of the proposed method. Finally, the experimental results on both clustering and classification tasks demonstrate the effectiveness of our GAE as well as the correctness of the proposed theoretical analysis, and it also suggests that GAE is a superior solution to the current deep representation learning techniques comparing with variant auto-encoders and existing local invariant methods.

  3. Geometric regularizations and dual conifold transitions

    International Nuclear Information System (INIS)

    Landsteiner, Karl; Lazaroiu, Calin I.

    2003-01-01

    We consider a geometric regularization for the class of conifold transitions relating D-brane systems on noncompact Calabi-Yau spaces to certain flux backgrounds. This regularization respects the SL(2,Z) invariance of the flux superpotential, and allows for computation of the relevant periods through the method of Picard-Fuchs equations. The regularized geometry is a noncompact Calabi-Yau which can be viewed as a monodromic fibration, with the nontrivial monodromy being induced by the regulator. It reduces to the original, non-monodromic background when the regulator is removed. Using this regularization, we discuss the simple case of the local conifold, and show how the relevant field-theoretic information can be extracted in this approach. (author)

  4. Generalization Performance of Regularized Ranking With Multiscale Kernels.

    Science.gov (United States)

    Zhou, Yicong; Chen, Hong; Lan, Rushi; Pan, Zhibin

    2016-05-01

    The regularized kernel method for the ranking problem has attracted increasing attentions in machine learning. The previous regularized ranking algorithms are usually based on reproducing kernel Hilbert spaces with a single kernel. In this paper, we go beyond this framework by investigating the generalization performance of the regularized ranking with multiscale kernels. A novel ranking algorithm with multiscale kernels is proposed and its representer theorem is proved. We establish the upper bound of the generalization error in terms of the complexity of hypothesis spaces. It shows that the multiscale ranking algorithm can achieve satisfactory learning rates under mild conditions. Experiments demonstrate the effectiveness of the proposed method for drug discovery and recommendation tasks.

  5. Explicit isospectral flows associated to the AKNS operator on the unit interval. II

    Science.gov (United States)

    Amour, Laurent

    2012-10-01

    Explicit flows associated to any tangent vector fields on any isospectral manifold for the AKNS operator acting in L2 × L2 on the unit interval are written down. The manifolds are of infinite dimension (and infinite codimension). The flows are called isospectral and also are Hamiltonian flows. It is proven that they may be explicitly expressed in terms of regularized determinants of infinite matrix-valued functions with entries depending only on the spectral data at the starting point of the flow. The tangent vector fields are decomposed as ∑ξkTk where ξ ∈ ℓ2 and the Tk ∈ L2 × L2 form a particular basis of the tangent vector spaces of the infinite dimensional manifold. The paper here is a continuation of Amour ["Explicit isospectral flows for the AKNS operator on the unit interval," Inverse Probl. 25, 095008 (2009)], 10.1088/0266-5611/25/9/095008 where, except for a finite number, all the components of the sequence ξ are zero in order to obtain an explicit expression for the isospectral flows. The regularized determinants induce counter-terms allowing for the consideration of finite quantities when the sequences ξ run all over ℓ2.

  6. Application of Fourier-wavelet regularized deconvolution for improving image quality of free space propagation x-ray phase contrast imaging.

    Science.gov (United States)

    Zhou, Zhongxing; Gao, Feng; Zhao, Huijuan; Zhang, Lixin

    2012-11-21

    New x-ray phase contrast imaging techniques without using synchrotron radiation confront a common problem from the negative effects of finite source size and limited spatial resolution. These negative effects swamp the fine phase contrast fringes and make them almost undetectable. In order to alleviate this problem, deconvolution procedures should be applied to the blurred x-ray phase contrast images. In this study, three different deconvolution techniques, including Wiener filtering, Tikhonov regularization and Fourier-wavelet regularized deconvolution (ForWaRD), were applied to the simulated and experimental free space propagation x-ray phase contrast images of simple geometric phantoms. These algorithms were evaluated in terms of phase contrast improvement and signal-to-noise ratio. The results demonstrate that the ForWaRD algorithm is most appropriate for phase contrast image restoration among above-mentioned methods; it can effectively restore the lost information of phase contrast fringes while reduce the amplified noise during Fourier regularization.

  7. Periodontal Disease, Regular Dental Care Use, and Incident Ischemic Stroke.

    Science.gov (United States)

    Sen, Souvik; Giamberardino, Lauren D; Moss, Kevin; Morelli, Thiago; Rosamond, Wayne D; Gottesman, Rebecca F; Beck, James; Offenbacher, Steven

    2018-02-01

    Periodontal disease is independently associated with cardiovascular disease. Identification of periodontal disease as a risk factor for incident ischemic stroke raises the possibility that regular dental care utilization may reduce the stroke risk. In the ARIC (Atherosclerosis Risk in Communities) study, pattern of dental visits were classified as regular or episodic dental care users. In the ancillary dental ARIC study, selected subjects from ARIC underwent fullmouth periodontal measurements collected at 6 sites per tooth and classified into 7 periodontal profile classes (PPCs). In the ARIC study 10 362 stroke-free participants, 584 participants had incident ischemic strokes over a 15-year period. In the dental ARIC study, 6736 dentate subjects were assessed for periodontal disease status using PPC with a total of 299 incident ischemic strokes over the 15-year period. The 7 levels of PPC showed a trend toward an increased stroke risk (χ 2 trend P periodontal disease). Periodontal disease was significantly associated with cardioembolic (hazard ratio, 2.6; 95% confidence interval, 1.2-5.6) and thrombotic (hazard ratio, 2.2; 95% confidence interval, 1.3-3.8) stroke subtypes. Regular dental care utilization was associated with lower adjusted stroke risk (hazard ratio, 0.77; 95% confidence interval, 0.63-0.94). We confirm an independent association between periodontal disease and incident stroke risk, particularly cardioembolic and thrombotic stroke subtype. Further, we report that regular dental care utilization may lower this risk for stroke. © 2018 American Heart Association, Inc.

  8. Regular non-twisting S-branes

    International Nuclear Information System (INIS)

    Obregon, Octavio; Quevedo, Hernando; Ryan, Michael P.

    2004-01-01

    We construct a family of time and angular dependent, regular S-brane solutions which corresponds to a simple analytical continuation of the Zipoy-Voorhees 4-dimensional vacuum spacetime. The solutions are asymptotically flat and turn out to be free of singularities without requiring a twist in space. They can be considered as the simplest non-singular generalization of the singular S0-brane solution. We analyze the properties of a representative of this family of solutions and show that it resembles to some extent the asymptotic properties of the regular Kerr S-brane. The R-symmetry corresponds, however, to the general lorentzian symmetry. Several generalizations of this regular solution are derived which include a charged S-brane and an additional dilatonic field. (author)

  9. Manifold Regularized Reinforcement Learning.

    Science.gov (United States)

    Li, Hongliang; Liu, Derong; Wang, Ding

    2018-04-01

    This paper introduces a novel manifold regularized reinforcement learning scheme for continuous Markov decision processes. Smooth feature representations for value function approximation can be automatically learned using the unsupervised manifold regularization method. The learned features are data-driven, and can be adapted to the geometry of the state space. Furthermore, the scheme provides a direct basis representation extension for novel samples during policy learning and control. The performance of the proposed scheme is evaluated on two benchmark control tasks, i.e., the inverted pendulum and the energy storage problem. Simulation results illustrate the concepts of the proposed scheme and show that it can obtain excellent performance.

  10. Matrix regularization of 4-manifolds

    OpenAIRE

    Trzetrzelewski, M.

    2012-01-01

    We consider products of two 2-manifolds such as S^2 x S^2, embedded in Euclidean space and show that the corresponding 4-volume preserving diffeomorphism algebra can be approximated by a tensor product SU(N)xSU(N) i.e. functions on a manifold are approximated by the Kronecker product of two SU(N) matrices. A regularization of the 4-sphere is also performed by constructing N^2 x N^2 matrix representations of the 4-algebra (and as a byproduct of the 3-algebra which makes the regularization of S...

  11. Regularity and chaos in cavity QED

    International Nuclear Information System (INIS)

    Bastarrachea-Magnani, Miguel Angel; López-del-Carpio, Baldemar; Chávez-Carlos, Jorge; Lerma-Hernández, Sergio; Hirsch, Jorge G

    2017-01-01

    The interaction of a quantized electromagnetic field in a cavity with a set of two-level atoms inside it can be described with algebraic Hamiltonians of increasing complexity, from the Rabi to the Dicke models. Their algebraic character allows, through the use of coherent states, a semiclassical description in phase space, where the non-integrable Dicke model has regions associated with regular and chaotic motion. The appearance of classical chaos can be quantified calculating the largest Lyapunov exponent over the whole available phase space for a given energy. In the quantum regime, employing efficient diagonalization techniques, we are able to perform a detailed quantitative study of the regular and chaotic regions, where the quantum participation ratio (P R ) of coherent states on the eigenenergy basis plays a role equivalent to the Lyapunov exponent. It is noted that, in the thermodynamic limit, dividing the participation ratio by the number of atoms leads to a positive value in chaotic regions, while it tends to zero in the regular ones. (paper)

  12. Regularity for a clamped grid equation $u_{xxxx}+u_{yyyy}=f $ on a domain with a corner

    Directory of Open Access Journals (Sweden)

    Tymofiy Gerasimov

    2009-04-01

    Full Text Available The operator $L=frac{partial ^{4}}{partial x^{4}} +frac{partial ^{4}}{partial y^{4}}$ appears in a model for the vertical displacement of a two-dimensional grid that consists of two perpendicular sets of elastic fibers or rods. We are interested in the behaviour of such a grid that is clamped at the boundary and more specifically near a corner of the domain. Kondratiev supplied the appropriate setting in the sense of Sobolev type spaces tailored to find the optimal regularity. Inspired by the Laplacian and the Bilaplacian models one expect, except maybe for some special angles that the optimal regularity improves when angle decreases. For the homogeneous Dirichlet problem with this special non-isotropic fourth order operator such a result does not hold true. We will show the existence of an interval $( frac{1}{2}pi ,omega _{star }$, $omega _{star }/pi approx 0.528dots$ (in degrees $omega _{star }approx 95.1dots^{circ} $, in which the optimal regularity improves with increasing opening angle.

  13. Learning About Time Within the Spinal Cord II: Evidence that Temporal Regularity is Encoded by a Spinal Oscillator

    Directory of Open Access Journals (Sweden)

    Kuan Hsien Lee

    2016-02-01

    Full Text Available How a stimulus impacts spinal cord function depends upon temporal relations. When intermittent noxious stimulation (shock is applied and the interval between shock pulses is varied (unpredictable, it induces a lasting alteration that inhibits adaptive learning. If the same stimulus is applied in a temporally regular (predictable manner, the capacity to learn is preserved and a protective/restorative effect is engaged that counters the adverse effect of variable stimulation. Sensitivity to temporal relations implies a capacity to encode time. This study explores how spinal neurons discriminate variable and fixed spaced stimulation. Communication with the brain was blocked by means of a spinal transection and adaptive capacity was tested using an instrumental learning task. In this task, subjects must learn to maintain a hind limb in a flexed position to minimize shock exposure. To evaluate the possibility that a distinct class of afferent fibers provide a sensory cue for regularity, we manipulated the temporal relation between shocks given to two dermatomes (leg and tail. Evidence for timing emerged when the stimuli were applied in a coherent manner across dermatomes, implying that a central (spinal process detects regularity. Next, we show that fixed spaced stimulation has a restorative effect when half the physical stimuli are randomly omitted, as long as the stimuli remain in phase, suggesting that stimulus regularity is encoded by an internal oscillator Research suggests that the oscillator that drives the tempo of stepping depends upon neurons within the rostral lumbar (L1-L2 region. Disrupting communication with the L1-L2 tissue by means of a L3 transection eliminated the restorative effect of fixed spaced stimulation. Implications of the results for step training and rehabilitation after injury are discussed.

  14. Fast and compact regular expression matching

    DEFF Research Database (Denmark)

    Bille, Philip; Farach-Colton, Martin

    2008-01-01

    We study 4 problems in string matching, namely, regular expression matching, approximate regular expression matching, string edit distance, and subsequence indexing, on a standard word RAM model of computation that allows logarithmic-sized words to be manipulated in constant time. We show how...... to improve the space and/or remove a dependency on the alphabet size for each problem using either an improved tabulation technique of an existing algorithm or by combining known algorithms in a new way....

  15. Accreting fluids onto regular black holes via Hamiltonian approach

    Energy Technology Data Exchange (ETDEWEB)

    Jawad, Abdul [COMSATS Institute of Information Technology, Department of Mathematics, Lahore (Pakistan); Shahzad, M.U. [COMSATS Institute of Information Technology, Department of Mathematics, Lahore (Pakistan); University of Central Punjab, CAMS, UCP Business School, Lahore (Pakistan)

    2017-08-15

    We investigate the accretion of test fluids onto regular black holes such as Kehagias-Sfetsos black holes and regular black holes with Dagum distribution function. We analyze the accretion process when different test fluids are falling onto these regular black holes. The accreting fluid is being classified through the equation of state according to the features of regular black holes. The behavior of fluid flow and the existence of sonic points is being checked for these regular black holes. It is noted that the three-velocity depends on critical points and the equation of state parameter on phase space. (orig.)

  16. Manifold Regularized Correlation Object Tracking.

    Science.gov (United States)

    Hu, Hongwei; Ma, Bo; Shen, Jianbing; Shao, Ling

    2018-05-01

    In this paper, we propose a manifold regularized correlation tracking method with augmented samples. To make better use of the unlabeled data and the manifold structure of the sample space, a manifold regularization-based correlation filter is introduced, which aims to assign similar labels to neighbor samples. Meanwhile, the regression model is learned by exploiting the block-circulant structure of matrices resulting from the augmented translated samples over multiple base samples cropped from both target and nontarget regions. Thus, the final classifier in our method is trained with positive, negative, and unlabeled base samples, which is a semisupervised learning framework. A block optimization strategy is further introduced to learn a manifold regularization-based correlation filter for efficient online tracking. Experiments on two public tracking data sets demonstrate the superior performance of our tracker compared with the state-of-the-art tracking approaches.

  17. UNFOLDED REGULAR AND SEMI-REGULAR POLYHEDRA

    Directory of Open Access Journals (Sweden)

    IONIŢĂ Elena

    2015-06-01

    Full Text Available This paper proposes a presentation unfolding regular and semi-regular polyhedra. Regular polyhedra are convex polyhedra whose faces are regular and equal polygons, with the same number of sides, and whose polyhedral angles are also regular and equal. Semi-regular polyhedra are convex polyhedra with regular polygon faces, several types and equal solid angles of the same type. A net of a polyhedron is a collection of edges in the plane which are the unfolded edges of the solid. Modeling and unfolding Platonic and Arhimediene polyhedra will be using 3dsMAX program. This paper is intended as an example of descriptive geometry applications.

  18. The geometric $\\beta$-function in curved space-time under operator regularization

    OpenAIRE

    Agarwala, Susama

    2009-01-01

    In this paper, I compare the generators of the renormalization group flow, or the geometric $\\beta$-functions for dimensional regularization and operator regularization. I then extend the analysis to show that the geometric $\\beta$-function for a scalar field theory on a closed compact Riemannian manifold is defined on the entire manifold. I then extend the analysis to find the generator of the renormalization group flow for a conformal scalar-field theories on the same manifolds. The geometr...

  19. Variational analysis of regular mappings theory and applications

    CERN Document Server

    Ioffe, Alexander D

    2017-01-01

    This monograph offers the first systematic account of (metric) regularity theory in variational analysis. It presents new developments alongside classical results and demonstrates the power of the theory through applications to various problems in analysis and optimization theory. The origins of metric regularity theory can be traced back to a series of fundamental ideas and results of nonlinear functional analysis and global analysis centered around problems of existence and stability of solutions of nonlinear equations. In variational analysis, regularity theory goes far beyond the classical setting and is also concerned with non-differentiable and multi-valued operators. The present volume explores all basic aspects of the theory, from the most general problems for mappings between metric spaces to those connected with fairly concrete and important classes of operators acting in Banach and finite dimensional spaces. Written by a leading expert in the field, the book covers new and powerful techniques, whic...

  20. The patterning of retinal horizontal cells: normalizing the regularity index enhances the detection of genomic linkage

    Directory of Open Access Journals (Sweden)

    Patrick W. Keeley

    2014-10-01

    Full Text Available Retinal neurons are often arranged as non-random distributions called mosaics, as their somata minimize proximity to neighboring cells of the same type. The horizontal cells serve as an example of such a mosaic, but little is known about the developmental mechanisms that underlie their patterning. To identify genes involved in this process, we have used three different spatial statistics to assess the patterning of the horizontal cell mosaic across a panel of genetically distinct recombinant inbred strains. To avoid the confounding effect cell density, which varies two-fold across these different strains, we computed the real/random regularity ratio, expressing the regularity of a mosaic relative to a randomly distributed simulation of similarly sized cells. To test whether this latter statistic better reflects the variation in biological processes that contribute to horizontal cell spacing, we subsequently compared the genetic linkage for each of these two traits, the regularity index and the real/random regularity ratio, each computed from the distribution of nearest neighbor (NN distances and from the Voronoi domain (VD areas. Finally, we compared each of these analyses with another index of patterning, the packing factor. Variation in the regularity indexes, as well as their real/random regularity ratios, and the packing factor, mapped quantitative trait loci (QTL to the distal ends of Chromosomes 1 and 14. For the NN and VD analyses, we found that the degree of linkage was greater when using the real/random regularity ratio rather than the respective regularity index. Using informatic resources, we narrow the list of prospective genes positioned at these two intervals to a small collection of six genes that warrant further investigation to determine their potential role in shaping the patterning of the horizontal cell mosaic.

  1. The geometric β-function in curved space-time under operator regularization

    Energy Technology Data Exchange (ETDEWEB)

    Agarwala, Susama [Mathematical Institute, Oxford University, Oxford OX2 6GG (United Kingdom)

    2015-06-15

    In this paper, I compare the generators of the renormalization group flow, or the geometric β-functions, for dimensional regularization and operator regularization. I then extend the analysis to show that the geometric β-function for a scalar field theory on a closed compact Riemannian manifold is defined on the entire manifold. I then extend the analysis to find the generator of the renormalization group flow to conformally coupled scalar-field theories on the same manifolds. The geometric β-function in this case is not defined.

  2. The geometric β-function in curved space-time under operator regularization

    International Nuclear Information System (INIS)

    Agarwala, Susama

    2015-01-01

    In this paper, I compare the generators of the renormalization group flow, or the geometric β-functions, for dimensional regularization and operator regularization. I then extend the analysis to show that the geometric β-function for a scalar field theory on a closed compact Riemannian manifold is defined on the entire manifold. I then extend the analysis to find the generator of the renormalization group flow to conformally coupled scalar-field theories on the same manifolds. The geometric β-function in this case is not defined

  3. Randomness control of vehicular motion through a sequence of traffic signals at irregular intervals

    International Nuclear Information System (INIS)

    Nagatani, Takashi

    2010-01-01

    We study the regularization of irregular motion of a vehicle moving through the sequence of traffic signals with a disordered configuration. Each traffic signal is controlled by both cycle time and phase shift. The cycle time is the same for all signals, while the phase shift varies from signal to signal by synchronizing with intervals between a signal and the next signal. The nonlinear dynamic model of the vehicular motion is presented by the stochastic nonlinear map. The vehicle exhibits the very complex behavior with varying both cycle time and strength of irregular intervals. The irregular motion induced by the disordered configuration is regularized by adjusting the phase shift within the regularization regions.

  4. Partial Regularity for Holonomic Minimisers of Quasiconvex Functionals

    Science.gov (United States)

    Hopper, Christopher P.

    2016-10-01

    We prove partial regularity for local minimisers of certain strictly quasiconvex integral functionals, over a class of Sobolev mappings into a compact Riemannian manifold, to which such mappings are said to be holonomically constrained. Our approach uses the lifting of Sobolev mappings to the universal covering space, the connectedness of the covering space, an application of Ekeland's variational principle and a certain tangential A-harmonic approximation lemma obtained directly via a Lipschitz approximation argument. This allows regularity to be established directly on the level of the gradient. Several applications to variational problems in condensed matter physics with broken symmetries are also discussed, in particular those concerning the superfluidity of liquid helium-3 and nematic liquid crystals.

  5. Explicit Inverse of an Interval Matrix with Unit Midpoint

    Czech Academy of Sciences Publication Activity Database

    Rohn, Jiří

    2011-01-01

    Roč. 22, - (2011), s. 138-150 E-ISSN 1081-3810 R&D Projects: GA ČR GA201/09/1957; GA ČR GC201/08/J020 Institutional research plan: CEZ:AV0Z10300504 Keywords : interval matrix * unit midpoint * inverse interval matrix * regularity Subject RIV: BA - General Mathematics Impact factor: 0.808, year: 2010 http://www.math.technion.ac.il/iic/ ela / ela -articles/articles/vol22_pp138-150.pdf

  6. Consistent momentum space regularization/renormalization of supersymmetric quantum field theories: the three-loop β-function for the Wess-Zumino model

    International Nuclear Information System (INIS)

    Carneiro, David; Sampaio, Marcos; Nemes, Maria Carolina; Scarpelli, Antonio Paulo Baeta

    2003-01-01

    We compute the three loop β function of the Wess-Zumino model to motivate implicit regularization (IR) as a consistent and practical momentum-space framework to study supersymmetric quantum field theories. In this framework which works essentially in the physical dimension of the theory we show that ultraviolet are clearly disentangled from infrared divergences. We obtain consistent results which motivate the method as a good choice to study supersymmetry anomalies in quantum field theories. (author)

  7. Parameter identification for continuous point emission source based on Tikhonov regularization method coupled with particle swarm optimization algorithm.

    Science.gov (United States)

    Ma, Denglong; Tan, Wei; Zhang, Zaoxiao; Hu, Jun

    2017-03-05

    In order to identify the parameters of hazardous gas emission source in atmosphere with less previous information and reliable probability estimation, a hybrid algorithm coupling Tikhonov regularization with particle swarm optimization (PSO) was proposed. When the source location is known, the source strength can be estimated successfully by common Tikhonov regularization method, but it is invalid when the information about both source strength and location is absent. Therefore, a hybrid method combining linear Tikhonov regularization and PSO algorithm was designed. With this method, the nonlinear inverse dispersion model was transformed to a linear form under some assumptions, and the source parameters including source strength and location were identified simultaneously by linear Tikhonov-PSO regularization method. The regularization parameters were selected by L-curve method. The estimation results with different regularization matrixes showed that the confidence interval with high-order regularization matrix is narrower than that with zero-order regularization matrix. But the estimation results of different source parameters are close to each other with different regularization matrixes. A nonlinear Tikhonov-PSO hybrid regularization was also designed with primary nonlinear dispersion model to estimate the source parameters. The comparison results of simulation and experiment case showed that the linear Tikhonov-PSO method with transformed linear inverse model has higher computation efficiency than nonlinear Tikhonov-PSO method. The confidence intervals from linear Tikhonov-PSO are more reasonable than that from nonlinear method. The estimation results from linear Tikhonov-PSO method are similar to that from single PSO algorithm, and a reasonable confidence interval with some probability levels can be additionally given by Tikhonov-PSO method. Therefore, the presented linear Tikhonov-PSO regularization method is a good potential method for hazardous emission

  8. Matrix regularization of embedded 4-manifolds

    International Nuclear Information System (INIS)

    Trzetrzelewski, Maciej

    2012-01-01

    We consider products of two 2-manifolds such as S 2 ×S 2 , embedded in Euclidean space and show that the corresponding 4-volume preserving diffeomorphism algebra can be approximated by a tensor product SU(N)⊗SU(N) i.e. functions on a manifold are approximated by the Kronecker product of two SU(N) matrices. A regularization of the 4-sphere is also performed by constructing N 2 ×N 2 matrix representations of the 4-algebra (and as a byproduct of the 3-algebra which makes the regularization of S 3 also possible).

  9. THE REGULARITIES OF THE SPACE-TEMPORAL DISTRIBUTION OF THE RADIATION BALANCE OF THE UNDERLYING SURFACE IN ARAKS BASIN ON MOUNTAINOUS TERRITORY OF THE REPUBLIC OF ARMENIA

    Directory of Open Access Journals (Sweden)

    V. G. Margaryan

    2017-12-01

    Full Text Available The regularities of the space-temporal distribution of the radiation balance of the underlying surface for the conditions of the mountainous territory of the Republic of Armenia were discussed and analyzed.

  10. Generalized Bregman distances and convergence rates for non-convex regularization methods

    International Nuclear Information System (INIS)

    Grasmair, Markus

    2010-01-01

    We generalize the notion of Bregman distance using concepts from abstract convexity in order to derive convergence rates for Tikhonov regularization with non-convex regularization terms. In particular, we study the non-convex regularization of linear operator equations on Hilbert spaces, showing that the conditions required for the application of the convergence rates results are strongly related to the standard range conditions from the convex case. Moreover, we consider the setting of sparse regularization, where we show that a rate of order δ 1/p holds, if the regularization term has a slightly faster growth at zero than |t| p

  11. Lavrentiev regularization method for nonlinear ill-posed problems

    International Nuclear Information System (INIS)

    Kinh, Nguyen Van

    2002-10-01

    In this paper we shall be concerned with Lavientiev regularization method to reconstruct solutions x 0 of non ill-posed problems F(x)=y o , where instead of y 0 noisy data y δ is an element of X with absolut(y δ -y 0 ) ≤ δ are given and F:X→X is an accretive nonlinear operator from a real reflexive Banach space X into itself. In this regularization method solutions x α δ are obtained by solving the singularly perturbed nonlinear operator equation F(x)+α(x-x*)=y δ with some initial guess x*. Assuming certain conditions concerning the operator F and the smoothness of the element x*-x 0 we derive stability estimates which show that the accuracy of the regularized solutions is order optimal provided that the regularization parameter α has been chosen properly. (author)

  12. Regular Gleason Measures and Generalized Effect Algebras

    Science.gov (United States)

    Dvurečenskij, Anatolij; Janda, Jiří

    2015-12-01

    We study measures, finitely additive measures, regular measures, and σ-additive measures that can attain even infinite values on the quantum logic of a Hilbert space. We show when particular classes of non-negative measures can be studied in the frame of generalized effect algebras.

  13. Regular exercisers have stronger pelvic floor muscles than nonregular exercisers at midpregnancy.

    Science.gov (United States)

    Bø, Kari; Ellstrøm Engh, Marie; Hilde, Gunvor

    2018-04-01

    Today all healthy pregnant women are encouraged to be physically active throughout pregnancy, with recommendations to participate in at least 30 minutes of aerobic activity on most days of the week in addition to performing strength training of the major muscle groups 2-3 days per week and also pelvic floor muscle training. There is, however, an ongoing debate whether general physical activity enhances or declines pelvic floor muscle function. The objectives of the study were to compare vaginal resting pressure, pelvic floor muscle strength, and endurance in regular exercisers (exercise ≥30 minutes 3 or more times per week) and nonexercisers at midpregnancy. Furthermore, another objective was to assess whether regular general exercise or pelvic floor muscle strength was associated with urinary incontinence. This was a cross-sectional study at mean gestational week 20.9 (±1.4) including 218 nulliparous pregnant women, with a mean age of 28.6 years (range, 19-40 years) and prepregnancy body mass index of 23.9 kg/m 2 (SD, 4.0). Vaginal resting pressure, pelvic floor muscle strength, and pelvic floor muscle endurance were measured by a high-precision pressure transducer connected to a vaginal balloon. The International Consultation on Incontinence Questionnaire Urinary Incontinence Short Form was used to assess urinary incontinence. Differences between groups were analyzed using an independent-sample Student t test. Linear regression analysis was conducted to adjust for prepregnancy body mass index, age, smoking during pregnancy, and regular pelvic floor muscle training during pregnancy. The significance value was set to P ≤ .05. Regular exercisers had statistically significant stronger (mean 6.4 cm H 2 O [95% confidence interval, 1.7-11.2]) and more enduring (mean 39.9 cm H 2 Osec [95% confidence interval, 42.2-75.7]) pelvic floor muscles. Only pelvic floor muscle strength remained statistically significant, when adjusting for possible confounders. Pelvic floor

  14. A multiresolution method for solving the Poisson equation using high order regularization

    DEFF Research Database (Denmark)

    Hejlesen, Mads Mølholm; Walther, Jens Honore

    2016-01-01

    We present a novel high order multiresolution Poisson solver based on regularized Green's function solutions to obtain exact free-space boundary conditions while using fast Fourier transforms for computational efficiency. Multiresolution is a achieved through local refinement patches and regulari......We present a novel high order multiresolution Poisson solver based on regularized Green's function solutions to obtain exact free-space boundary conditions while using fast Fourier transforms for computational efficiency. Multiresolution is a achieved through local refinement patches...... and regularized Green's functions corresponding to the difference in the spatial resolution between the patches. The full solution is obtained utilizing the linearity of the Poisson equation enabling super-position of solutions. We show that the multiresolution Poisson solver produces convergence rates...

  15. Diverse Regular Employees and Non-regular Employment (Japanese)

    OpenAIRE

    MORISHIMA Motohiro

    2011-01-01

    Currently there are high expectations for the introduction of policies related to diverse regular employees. These policies are a response to the problem of disparities between regular and non-regular employees (part-time, temporary, contract and other non-regular employees) and will make it more likely that workers can balance work and their private lives while companies benefit from the advantages of regular employment. In this paper, I look at two issues that underlie this discussion. The ...

  16. Incidence of Interval Colorectal Cancer Among Inflammatory Bowel Disease Patients Undergoing Regular Colonoscopic Surveillance

    NARCIS (Netherlands)

    Mooiweer, Erik; van der Meulen-de Jong, Andrea E.; Ponsioen, Cyriel Y.; van der Woude, C. Janneke; van Bodegraven, Ad A.; Jansen, Jeroen M.; Mahmmod, Nofel; Kremer, Willemijn; Siersema, Peter D.; Oldenburg, Bas

    2015-01-01

    Surveillance is recommended for patients with long-term inflammatory bowel disease because they have an increased risk of colorectal cancer (CRC). To study the effectiveness of surveillance, we determined the incidence of CRC after negative findings from surveillance colonoscopies (interval CRC).

  17. A blind deconvolution method based on L1/L2 regularization prior in the gradient space

    Science.gov (United States)

    Cai, Ying; Shi, Yu; Hua, Xia

    2018-02-01

    In the process of image restoration, the result of image restoration is very different from the real image because of the existence of noise, in order to solve the ill posed problem in image restoration, a blind deconvolution method based on L1/L2 regularization prior to gradient domain is proposed. The method presented in this paper first adds a function to the prior knowledge, which is the ratio of the L1 norm to the L2 norm, and takes the function as the penalty term in the high frequency domain of the image. Then, the function is iteratively updated, and the iterative shrinkage threshold algorithm is applied to solve the high frequency image. In this paper, it is considered that the information in the gradient domain is better for the estimation of blur kernel, so the blur kernel is estimated in the gradient domain. This problem can be quickly implemented in the frequency domain by fast Fast Fourier Transform. In addition, in order to improve the effectiveness of the algorithm, we have added a multi-scale iterative optimization method. This paper proposes the blind deconvolution method based on L1/L2 regularization priors in the gradient space can obtain the unique and stable solution in the process of image restoration, which not only keeps the edges and details of the image, but also ensures the accuracy of the results.

  18. Regularity criteria for incompressible magnetohydrodynamics equations in three dimensions

    International Nuclear Information System (INIS)

    Lin, Hongxia; Du, Lili

    2013-01-01

    In this paper, we give some new global regularity criteria for three-dimensional incompressible magnetohydrodynamics (MHD) equations. More precisely, we provide some sufficient conditions in terms of the derivatives of the velocity or pressure, for the global regularity of strong solutions to 3D incompressible MHD equations in the whole space, as well as for periodic boundary conditions. Moreover, the regularity criterion involving three of the nine components of the velocity gradient tensor is also obtained. The main results generalize the recent work by Cao and Wu (2010 Two regularity criteria for the 3D MHD equations J. Diff. Eqns 248 2263–74) and the analysis in part is based on the works by Cao C and Titi E (2008 Regularity criteria for the three-dimensional Navier–Stokes equations Indiana Univ. Math. J. 57 2643–61; 2011 Gobal regularity criterion for the 3D Navier–Stokes equations involving one entry of the velocity gradient tensor Arch. Rational Mech. Anal. 202 919–32) for 3D incompressible Navier–Stokes equations. (paper)

  19. Incidence of Interval Colorectal Cancer Among Inflammatory Bowel Disease Patients Undergoing Regular Colonoscopic Surveillance

    NARCIS (Netherlands)

    Mooiweer, Erik; van der Meulen-de Jong, Andrea E.; Ponsioen, Cyriel Y.; van der Woude, C. Janneke; van Bodegraven, Ad A.; Jansen, Jeroen M.; Mahmmod, Nofel; Kremer, Willemijn; Siersema, Peter D.; Oldenburg, Bas

    2015-01-01

    Surveillance is recommended for patients with long-term inflammatory bowel disease because they have an increased risk of colorectal cancer (CRC). To study the effectiveness of surveillance, we determined the incidence of CRC after negative findings from surveillance colonoscopies (interval CRC). We

  20. Regular black hole in three dimensions

    OpenAIRE

    Myung, Yun Soo; Yoon, Myungseok

    2008-01-01

    We find a new black hole in three dimensional anti-de Sitter space by introducing an anisotropic perfect fluid inspired by the noncommutative black hole. This is a regular black hole with two horizons. We compare thermodynamics of this black hole with that of non-rotating BTZ black hole. The first-law of thermodynamics is not compatible with the Bekenstein-Hawking entropy.

  1. Dimensional versus lattice regularization within Luescher's Yang Mills theory

    International Nuclear Information System (INIS)

    Diekmann, B.; Langer, M.; Schuette, D.

    1993-01-01

    It is pointed out that the coefficients of Luescher's effective model space Hamiltonian, which is based upon dimensional regularization techniques, can be reproduced by applying folded diagram perturbation theory to the Kogut Susskind Hamiltonian and by performing a lattice continuum limit (keeping the volume fixed). Alternative cutoff regularizations of the Hamiltonian are in general inconsistent, the critical point beeing the correct prediction for Luescher's tadpole coefficient which is formally quadratically divergent and which has to become a well defined (negative) number. (orig.)

  2. Regularization and asymptotic expansion of certain distributions defined by divergent series

    Directory of Open Access Journals (Sweden)

    Ricardo Estrada

    1995-01-01

    Full Text Available The regularization of the distribution ∑n=−∞∞δ(x−pn. which gives a regularized value to the divergent series ∑n=−∞∞φ(pn is obtained in several spaces of test functions. The asymptotic expansion as ϵ→0+of series of the type ∑n=0∞φ(ϵ pn is also obtained.

  3. Experiment Description and Results for Arrival Operations Using Interval Management with Spacing to Parallel Dependent Runways (IMSPiDR)

    Science.gov (United States)

    Baxley, Brian T.; Murdoch, Jennifer L.; Swieringa, Kurt A.; Barmore, Bryan E.; Capron, William R.; Hubbs, Clay E.; Shay, Richard F.; Abbott, Terence S.

    2013-01-01

    The predicted increase in the number of commercial aircraft operations creates a need for improved operational efficiency. Two areas believed to offer increases in aircraft efficiency are optimized profile descents and dependent parallel runway operations. Using Flight deck Interval Management (FIM) software and procedures during these operations, flight crews can achieve by the runway threshold an interval assigned by air traffic control (ATC) behind the preceding aircraft that maximizes runway throughput while minimizing additional fuel consumption and pilot workload. This document describes an experiment where 24 pilots flew arrivals into the Dallas Fort-Worth terminal environment using one of three simulators at NASA?s Langley Research Center. Results indicate that pilots delivered their aircraft to the runway threshold within +/- 3.5 seconds of their assigned time interval, and reported low workload levels. In general, pilots found the FIM concept, procedures, speeds, and interface acceptable. Analysis of the time error and FIM speed changes as a function of arrival stream position suggest the spacing algorithm generates stable behavior while in the presence of continuous (wind) or impulse (offset) error. Concerns reported included multiple speed changes within a short time period, and an airspeed increase followed shortly by an airspeed decrease.

  4. Simulation of Canopy CO2/H2O Fluxes for a Rubber (Hevea Brasiliensis) Plantation in Central Cambodia: The Effect of the Regular Spacing of Planted Trees

    Energy Technology Data Exchange (ETDEWEB)

    Kumagai, Tomo' omi; Mudd, Ryan; Miyazawa, Yoshiyuki; Liu, Wen; Giambelluca, Thomas; Kobayashi, N.; Lim, Tiva Khan; Jomura, Mayuko; Matsumoto, Kazuho; Huang, Maoyi; Chen, Qi; Ziegler, Alan; Yin, Song

    2013-09-10

    We developed a soil-vegetation-atmosphere transfer (SVAT) model applicable to simulating CO2 and H2O fluxes from the canopies of rubber plantations, which are characterized by distinct canopy clumping produced by regular spacing of plantation trees. Rubber (Hevea brasiliensis Müll. Arg.) plantations, which are rapidly expanding into both climatically optimal and sub-optimal environments throughout mainland Southeast Asia, potentially change the partitioning of water, energy, and carbon at multiple scales, compared with traditional land covers it is replacing. Describing the biosphere-atmosphere exchange in rubber plantations via SVAT modeling is therefore essential to understanding the impacts on environmental processes. The regular spacing of plantation trees creates a peculiar canopy structure that is not well represented in most SVAT models, which generally assumes a non-uniform spacing of vegetation. Herein we develop a SVAT model applicable to rubber plantation and an evaluation method for its canopy structure, and examine how the peculiar canopy structure of rubber plantations affects canopy CO2 and H2O exchanges. Model results are compared with measurements collected at a field site in central Cambodia. Our findings suggest that it is crucial to account for intensive canopy clumping in order to reproduce observed rubber plantation fluxes. These results suggest a potentially optimal spacing of rubber trees to produce high productivity and water use efficiency.

  5. Regular graph construction for semi-supervised learning

    International Nuclear Information System (INIS)

    Vega-Oliveros, Didier A; Berton, Lilian; Eberle, Andre Mantini; Lopes, Alneu de Andrade; Zhao, Liang

    2014-01-01

    Semi-supervised learning (SSL) stands out for using a small amount of labeled points for data clustering and classification. In this scenario graph-based methods allow the analysis of local and global characteristics of the available data by identifying classes or groups regardless data distribution and representing submanifold in Euclidean space. Most of methods used in literature for SSL classification do not worry about graph construction. However, regular graphs can obtain better classification accuracy compared to traditional methods such as k-nearest neighbor (kNN), since kNN benefits the generation of hubs and it is not appropriate for high-dimensionality data. Nevertheless, methods commonly used for generating regular graphs have high computational cost. We tackle this problem introducing an alternative method for generation of regular graphs with better runtime performance compared to methods usually find in the area. Our technique is based on the preferential selection of vertices according some topological measures, like closeness, generating at the end of the process a regular graph. Experiments using the global and local consistency method for label propagation show that our method provides better or equal classification rate in comparison with kNN

  6. Concept of Operations for Interval Management Arrivals and Approach

    Science.gov (United States)

    Hicok, Daniel S.; Barmore, Bryan E.

    2016-01-01

    This paper presents the concept of operations for interval management operations to be deployed in the US National Airspace System (NAS) by the Federal Aviation Administration (FAA) after 2020. The use of interval management operations is described that begin in en route airspace and continue to a termination point inside the arrival terminal area, in a terminal environment that includes other arrival management tools such as arrival metering, Ground-based Interval Management - Spacing (GIM-S), and Terminal Sequencing and Spacing (TSAS). The roles of Air Traffic Controllers and Flight Crews and the ground automation tools that are used by Air Traffic Controllers to enable the primary operation and variations are described.

  7. Regularity of p(ṡ)-superharmonic functions, the Kellogg property and semiregular boundary points

    Science.gov (United States)

    Adamowicz, Tomasz; Björn, Anders; Björn, Jana

    2014-11-01

    We study various boundary and inner regularity questions for $p(\\cdot)$-(super)harmonic functions in Euclidean domains. In particular, we prove the Kellogg property and introduce a classification of boundary points for $p(\\cdot)$-harmonic functions into three disjoint classes: regular, semiregular and strongly irregular points. Regular and especially semiregular points are characterized in many ways. The discussion is illustrated by examples. Along the way, we present a removability result for bounded $p(\\cdot)$-harmonic functions and give some new characterizations of $W^{1, p(\\cdot)}_0$ spaces. We also show that $p(\\cdot)$-superharmonic functions are lower semicontinuously regularized, and characterize them in terms of lower semicontinuously regularized supersolutions.

  8. Timing intervals using population synchrony and spike timing dependent plasticity

    Directory of Open Access Journals (Sweden)

    Wei Xu

    2016-12-01

    Full Text Available We present a computational model by which ensembles of regularly spiking neurons can encode different time intervals through synchronous firing. We show that a neuron responding to a large population of convergent inputs has the potential to learn to produce an appropriately-timed output via spike-time dependent plasticity. We explain why temporal variability of this population synchrony increases with increasing time intervals. We also show that the scalar property of timing and its violation at short intervals can be explained by the spike-wise accumulation of jitter in the inter-spike intervals of timing neurons. We explore how the challenge of encoding longer time intervals can be overcome and conclude that this may involve a switch to a different population of neurons with lower firing rate, with the added effect of producing an earlier bias in response. Experimental data on human timing performance show features in agreement with the model’s output.

  9. Selection of regularization parameter for l1-regularized damage detection

    Science.gov (United States)

    Hou, Rongrong; Xia, Yong; Bao, Yuequan; Zhou, Xiaoqing

    2018-06-01

    The l1 regularization technique has been developed for structural health monitoring and damage detection through employing the sparsity condition of structural damage. The regularization parameter, which controls the trade-off between data fidelity and solution size of the regularization problem, exerts a crucial effect on the solution. However, the l1 regularization problem has no closed-form solution, and the regularization parameter is usually selected by experience. This study proposes two strategies of selecting the regularization parameter for the l1-regularized damage detection problem. The first method utilizes the residual and solution norms of the optimization problem and ensures that they are both small. The other method is based on the discrepancy principle, which requires that the variance of the discrepancy between the calculated and measured responses is close to the variance of the measurement noise. The two methods are applied to a cantilever beam and a three-story frame. A range of the regularization parameter, rather than one single value, can be determined. When the regularization parameter in this range is selected, the damage can be accurately identified even for multiple damage scenarios. This range also indicates the sensitivity degree of the damage identification problem to the regularization parameter.

  10. Chord length distributions between hard disks and spheres in regular, semi-regular, and quasi-random structures

    International Nuclear Information System (INIS)

    Olson, Gordon L.

    2008-01-01

    In binary stochastic media in two- and three-dimensions consisting of randomly placed impenetrable disks or spheres, the chord lengths in the background material between disks and spheres closely follow exponential distributions if the disks and spheres occupy less than 10% of the medium. This work demonstrates that for regular spatial structures of disks and spheres, the tails of the chord length distributions (CLDs) follow power laws rather than exponentials. In dilute media, when the disks and spheres are widely spaced, the slope of the power law seems to be independent of the details of the structure. When approaching a close-packed arrangement, the exact placement of the spheres can make a significant difference. When regular structures are perturbed by small random displacements, the CLDs become power laws with steeper slopes. An example CLD from a quasi-random distribution of spheres in clusters shows a modified exponential distribution

  11. Chord length distributions between hard disks and spheres in regular, semi-regular, and quasi-random structures

    Energy Technology Data Exchange (ETDEWEB)

    Olson, Gordon L. [Computer and Computational Sciences Division (CCS-2), Los Alamos National Laboratory, 5 Foxglove Circle, Madison, WI 53717 (United States)], E-mail: olson99@tds.net

    2008-11-15

    In binary stochastic media in two- and three-dimensions consisting of randomly placed impenetrable disks or spheres, the chord lengths in the background material between disks and spheres closely follow exponential distributions if the disks and spheres occupy less than 10% of the medium. This work demonstrates that for regular spatial structures of disks and spheres, the tails of the chord length distributions (CLDs) follow power laws rather than exponentials. In dilute media, when the disks and spheres are widely spaced, the slope of the power law seems to be independent of the details of the structure. When approaching a close-packed arrangement, the exact placement of the spheres can make a significant difference. When regular structures are perturbed by small random displacements, the CLDs become power laws with steeper slopes. An example CLD from a quasi-random distribution of spheres in clusters shows a modified exponential distribution.

  12. Convergence rates in constrained Tikhonov regularization: equivalence of projected source conditions and variational inequalities

    International Nuclear Information System (INIS)

    Flemming, Jens; Hofmann, Bernd

    2011-01-01

    In this paper, we enlighten the role of variational inequalities for obtaining convergence rates in Tikhonov regularization of nonlinear ill-posed problems with convex penalty functionals under convexity constraints in Banach spaces. Variational inequalities are able to cover solution smoothness and the structure of nonlinearity in a uniform manner, not only for unconstrained but, as we indicate, also for constrained Tikhonov regularization. In this context, we extend the concept of projected source conditions already known in Hilbert spaces to Banach spaces, and we show in the main theorem that such projected source conditions are to some extent equivalent to certain variational inequalities. The derived variational inequalities immediately yield convergence rates measured by Bregman distances

  13. Nonlocal Regularized Algebraic Reconstruction Techniques for MRI: An Experimental Study

    Directory of Open Access Journals (Sweden)

    Xin Li

    2013-01-01

    Full Text Available We attempt to revitalize researchers' interest in algebraic reconstruction techniques (ART by expanding their capabilities and demonstrating their potential in speeding up the process of MRI acquisition. Using a continuous-to-discrete model, we experimentally study the application of ART into MRI reconstruction which unifies previous nonuniform-fast-Fourier-transform- (NUFFT- based and gridding-based approaches. Under the framework of ART, we advocate the use of nonlocal regularization techniques which are leveraged from our previous research on modeling photographic images. It is experimentally shown that nonlocal regularization ART (NR-ART can often outperform their local counterparts in terms of both subjective and objective qualities of reconstructed images. On one real-world k-space data set, we find that nonlocal regularization can achieve satisfactory reconstruction from as few as one-third of samples. We also address an issue related to image reconstruction from real-world k-space data but overlooked in the open literature: the consistency of reconstructed images across different resolutions. A resolution-consistent extension of NR-ART is developed and shown to effectively suppress the artifacts arising from frequency extrapolation. Both source codes and experimental results of this work are made fully reproducible.

  14. Development of a swim-type ROV for narrow space inspection

    International Nuclear Information System (INIS)

    Okada, Satoshi; Otani, Kenichi; Kobayashi, Ryosuke; Ohno, Kazunori

    2017-01-01

    The swim-type remotely operated vehicle (ROV) for inspection of narrow spaces in nuclear power plants has been developed. Many structures are crowded in a confined space at regular intervals in the bottom area of a reactor. So, the thickness of the ROV shape is an important design point to ensure that the ROV can move in the space. The developed ROV has a three-dimensional swimming mechanism using six thrusters, three cameras for observing the position while moving and for making inspections easily, and a localization system. The localization system combines two elements: a gyroscope to detect the progression direction; and a slit laser that detects the progression distance using the optical cutting method. The localization method is called the modified inertial navigation (MIN) method and it was evaluated in a mock-up examination. The ROV was able to move smoothly using the MIN method and its position could be detected without making a mistake in the route followed. (author)

  15. Borderline personality disorder and regularly drinking alcohol before sex.

    Science.gov (United States)

    Thompson, Ronald G; Eaton, Nicholas R; Hu, Mei-Chen; Hasin, Deborah S

    2017-07-01

    Drinking alcohol before sex increases the likelihood of engaging in unprotected intercourse, having multiple sexual partners and becoming infected with sexually transmitted infections. Borderline personality disorder (BPD), a complex psychiatric disorder characterised by pervasive instability in emotional regulation, self-image, interpersonal relationships and impulse control, is associated with substance use disorders and sexual risk behaviours. However, no study has examined the relationship between BPD and drinking alcohol before sex in the USA. This study examined the association between BPD and regularly drinking before sex in a nationally representative adult sample. Participants were 17 491 sexually active drinkers from Wave 2 of the National Epidemiologic Survey on Alcohol and Related Conditions. Logistic regression models estimated effects of BPD diagnosis, specific borderline diagnostic criteria and BPD criterion count on the likelihood of regularly (mostly or always) drinking alcohol before sex, adjusted for controls. Borderline personality disorder diagnosis doubled the odds of regularly drinking before sex [adjusted odds ratio (AOR) = 2.26; confidence interval (CI) = 1.63, 3.14]. Of nine diagnostic criteria, impulsivity in areas that are self-damaging remained a significant predictor of regularly drinking before sex (AOR = 1.82; CI = 1.42, 2.35). The odds of regularly drinking before sex increased by 20% for each endorsed criterion (AOR = 1.20; CI = 1.14, 1.27) DISCUSSION AND CONCLUSIONS: This is the first study to examine the relationship between BPD and regularly drinking alcohol before sex in the USA. Substance misuse treatment should assess regularly drinking before sex, particularly among patients with BPD, and BPD treatment should assess risk at the intersection of impulsivity, sexual behaviour and substance use. [Thompson Jr RG, Eaton NR, Hu M-C, Hasin DS Borderline personality disorder and regularly drinking alcohol

  16. Mathematical Modeling the Geometric Regularity in Proteus Mirabilis Colonies

    Science.gov (United States)

    Zhang, Bin; Jiang, Yi; Minsu Kim Collaboration

    Proteus Mirabilis colony exhibits striking spatiotemporal regularity, with concentric ring patterns with alternative high and low bacteria density in space, and periodicity for repetition process of growth and swarm in time. We present a simple mathematical model to explain the spatiotemporal regularity of P. Mirabilis colonies. We study a one-dimensional system. Using a reaction-diffusion model with thresholds in cell density and nutrient concentration, we recreated periodic growth and spread patterns, suggesting that the nutrient constraint and cell density regulation might be sufficient to explain the spatiotemporal periodicity in P. Mirabilis colonies. We further verify this result using a cell based model.

  17. Nonparametric Estimation of Interval Reliability for Discrete-Time Semi-Markov Systems

    DEFF Research Database (Denmark)

    Georgiadis, Stylianos; Limnios, Nikolaos

    2016-01-01

    In this article, we consider a repairable discrete-time semi-Markov system with finite state space. The measure of the interval reliability is given as the probability of the system being operational over a given finite-length time interval. A nonparametric estimator is proposed for the interval...

  18. Regular and stochastic particle motion in plasma dynamics

    International Nuclear Information System (INIS)

    Kaufman, A.N.

    1979-08-01

    A Hamiltonian formalism is presented for the study of charged-particle trajectories in the self-consistent field of the particles. The intention is to develop a general approach to plasma dynamics. Transformations of phase-space variables are used to separate out the regular, adiabatic motion from the irregular, stochastic trajectories. Several new techniques are included in this presentation

  19. Intrinsic Regularization in a Lorentz invariant non-orthogonal Euclidean Space

    OpenAIRE

    Tornow, Carmen

    2006-01-01

    It is shown that the Lorentz transformations can be derived for a non-orthogonal Euclidean space. In this geometry one finds the same relations of special relativity as the ones known from the orthogonal Minkowski space. In order to illustrate the advantage of a non-orthogonal Euclidean metric the two-point Green’s function at x = 0 for a self-interacting scalar field is calculated. In contrast to the Minkowski space the one loop mass correction derived from this function gives a convergent r...

  20. Experimental uncertainty estimation and statistics for data having interval uncertainty.

    Energy Technology Data Exchange (ETDEWEB)

    Kreinovich, Vladik (Applied Biomathematics, Setauket, New York); Oberkampf, William Louis (Applied Biomathematics, Setauket, New York); Ginzburg, Lev (Applied Biomathematics, Setauket, New York); Ferson, Scott (Applied Biomathematics, Setauket, New York); Hajagos, Janos (Applied Biomathematics, Setauket, New York)

    2007-05-01

    This report addresses the characterization of measurements that include epistemic uncertainties in the form of intervals. It reviews the application of basic descriptive statistics to data sets which contain intervals rather than exclusively point estimates. It describes algorithms to compute various means, the median and other percentiles, variance, interquartile range, moments, confidence limits, and other important statistics and summarizes the computability of these statistics as a function of sample size and characteristics of the intervals in the data (degree of overlap, size and regularity of widths, etc.). It also reviews the prospects for analyzing such data sets with the methods of inferential statistics such as outlier detection and regressions. The report explores the tradeoff between measurement precision and sample size in statistical results that are sensitive to both. It also argues that an approach based on interval statistics could be a reasonable alternative to current standard methods for evaluating, expressing and propagating measurement uncertainties.

  1. Effective action for scalar fields and generalized zeta-function regularization

    International Nuclear Information System (INIS)

    Cognola, Guido; Zerbini, Sergio

    2004-01-01

    Motivated by the study of quantum fields in a Friedmann-Robertson-Walker space-time, the one-loop effective action for a scalar field defined in the ultrastatic manifold RxH 3 /Γ, H 3 /Γ being the finite volume, noncompact, hyperbolic spatial section, is investigated by a generalization of zeta-function regularization. It is shown that additional divergences may appear at the one-loop level. The one-loop renormalizability of the model is discussed and, making use of a generalization of zeta-function regularization, the one-loop renormalization group equations are derived

  2. Entrained rhythmic activities of neuronal ensembles as perceptual memory of time interval.

    Science.gov (United States)

    Sumbre, Germán; Muto, Akira; Baier, Herwig; Poo, Mu-ming

    2008-11-06

    The ability to process temporal information is fundamental to sensory perception, cognitive processing and motor behaviour of all living organisms, from amoebae to humans. Neural circuit mechanisms based on neuronal and synaptic properties have been shown to process temporal information over the range of tens of microseconds to hundreds of milliseconds. How neural circuits process temporal information in the range of seconds to minutes is much less understood. Studies of working memory in monkeys and rats have shown that neurons in the prefrontal cortex, the parietal cortex and the thalamus exhibit ramping activities that linearly correlate with the lapse of time until the end of a specific time interval of several seconds that the animal is trained to memorize. Many organisms can also memorize the time interval of rhythmic sensory stimuli in the timescale of seconds and can coordinate motor behaviour accordingly, for example, by keeping the rhythm after exposure to the beat of music. Here we report a form of rhythmic activity among specific neuronal ensembles in the zebrafish optic tectum, which retains the memory of the time interval (in the order of seconds) of repetitive sensory stimuli for a duration of up to approximately 20 s. After repetitive visual conditioning stimulation (CS) of zebrafish larvae, we observed rhythmic post-CS activities among specific tectal neuronal ensembles, with a regular interval that closely matched the CS. Visuomotor behaviour of the zebrafish larvae also showed regular post-CS repetitions at the entrained time interval that correlated with rhythmic neuronal ensemble activities in the tectum. Thus, rhythmic activities among specific neuronal ensembles may act as an adjustable 'metronome' for time intervals in the order of seconds, and serve as a mechanism for the short-term perceptual memory of rhythmic sensory experience.

  3. Mixed Total Variation and L1 Regularization Method for Optical Tomography Based on Radiative Transfer Equation

    Directory of Open Access Journals (Sweden)

    Jinping Tang

    2017-01-01

    Full Text Available Optical tomography is an emerging and important molecular imaging modality. The aim of optical tomography is to reconstruct optical properties of human tissues. In this paper, we focus on reconstructing the absorption coefficient based on the radiative transfer equation (RTE. It is an ill-posed parameter identification problem. Regularization methods have been broadly applied to reconstruct the optical coefficients, such as the total variation (TV regularization and the L1 regularization. In order to better reconstruct the piecewise constant and sparse coefficient distributions, TV and L1 norms are combined as the regularization. The forward problem is discretized with the discontinuous Galerkin method on the spatial space and the finite element method on the angular space. The minimization problem is solved by a Jacobian-based Levenberg-Marquardt type method which is equipped with a split Bregman algorithms for the L1 regularization. We use the adjoint method to compute the Jacobian matrix which dramatically improves the computation efficiency. By comparing with the other imaging reconstruction methods based on TV and L1 regularizations, the simulation results show the validity and efficiency of the proposed method.

  4. An optimal design of cluster spacing intervals for staged fracturing in horizontal shale gas wells based on the optimal SRVs

    Directory of Open Access Journals (Sweden)

    Lan Ren

    2017-09-01

    Full Text Available When horizontal well staged cluster fracturing is applied in shale gas reservoirs, the cluster spacing is essential to fracturing performance. If the cluster spacing is too small, the stimulated area between major fractures will be overlapped, and the efficiency of fracturing stimulation will be decreased. If the cluster spacing is too large, the area between major fractures cannot be stimulated completely and reservoir recovery extent will be adversely impacted. At present, cluster spacing design is mainly based on the static model with the potential reservoir stimulation area as the target, and there is no cluster spacing design method in accordance with the actual fracturing process and targets dynamic stimulated reservoir volume (SRV. In this paper, a dynamic SRV calculation model for cluster fracture propagation was established by analyzing the coupling mechanisms among fracture propagation, fracturing fluid loss and stress. Then, the cluster spacing was optimized to reach the target of the optimal SRVs. This model was applied for validation on site in the Jiaoshiba shale gasfield in the Fuling area of the Sichuan Basin. The key geological engineering parameters influencing the optimal cluster spacing intervals were analyzed. The reference charts for the optimal cluster spacing design were prepared based on the geological characteristics of south and north blocks in the Jiaoshiba shale gasfield. It is concluded that the cluster spacing optimal design method proposed in this paper is of great significance in overcoming the blindness in current cluster perforation design and guiding the optimal design of volume fracturing in shale gas reservoirs. Keywords: Shale gas, Horizontal well, Staged fracturing, Cluster spacing, Reservoir, Stimulated reservoir volume (SRV, Mathematical model, Optimal method, Sichuan basin, Jiaoshiba shale gasfield

  5. Major earthquakes occur regularly on an isolated plate boundary fault.

    Science.gov (United States)

    Berryman, Kelvin R; Cochran, Ursula A; Clark, Kate J; Biasi, Glenn P; Langridge, Robert M; Villamor, Pilar

    2012-06-29

    The scarcity of long geological records of major earthquakes, on different types of faults, makes testing hypotheses of regular versus random or clustered earthquake recurrence behavior difficult. We provide a fault-proximal major earthquake record spanning 8000 years on the strike-slip Alpine Fault in New Zealand. Cyclic stratigraphy at Hokuri Creek suggests that the fault ruptured to the surface 24 times, and event ages yield a 0.33 coefficient of variation in recurrence interval. We associate this near-regular earthquake recurrence with a geometrically simple strike-slip fault, with high slip rate, accommodating a high proportion of plate boundary motion that works in isolation from other faults. We propose that it is valid to apply time-dependent earthquake recurrence models for seismic hazard estimation to similar faults worldwide.

  6. Feature selection and multi-kernel learning for adaptive graph regularized nonnegative matrix factorization

    KAUST Repository

    Wang, Jim Jing-Yan

    2014-09-20

    Nonnegative matrix factorization (NMF), a popular part-based representation technique, does not capture the intrinsic local geometric structure of the data space. Graph regularized NMF (GNMF) was recently proposed to avoid this limitation by regularizing NMF with a nearest neighbor graph constructed from the input data set. However, GNMF has two main bottlenecks. First, using the original feature space directly to construct the graph is not necessarily optimal because of the noisy and irrelevant features and nonlinear distributions of data samples. Second, one possible way to handle the nonlinear distribution of data samples is by kernel embedding. However, it is often difficult to choose the most suitable kernel. To solve these bottlenecks, we propose two novel graph-regularized NMF methods, AGNMFFS and AGNMFMK, by introducing feature selection and multiple-kernel learning to the graph regularized NMF, respectively. Instead of using a fixed graph as in GNMF, the two proposed methods learn the nearest neighbor graph that is adaptive to the selected features and learned multiple kernels, respectively. For each method, we propose a unified objective function to conduct feature selection/multi-kernel learning, NMF and adaptive graph regularization simultaneously. We further develop two iterative algorithms to solve the two optimization problems. Experimental results on two challenging pattern classification tasks demonstrate that the proposed methods significantly outperform state-of-the-art data representation methods.

  7. Regularizing Feynman path integrals using the generalized Kontsevich-Vishik trace

    Science.gov (United States)

    Hartung, Tobias

    2017-12-01

    A fully regulated definition of Feynman's path integral is presented here. The proposed re-formulation of the path integral coincides with the familiar formulation whenever the path integral is well defined. In particular, it is consistent with respect to lattice formulations and Wick rotations, i.e., it can be used in Euclidean and Minkowski space-time. The path integral regularization is introduced through the generalized Kontsevich-Vishik trace, that is, the extension of the classical trace to Fourier integral operators. Physically, we are replacing the time-evolution semi-group by a holomorphic family of operators such that the corresponding path integrals are well defined in some half space of C . The regularized path integral is, thus, defined through analytic continuation. This regularization can be performed by means of stationary phase approximation or computed analytically depending only on the Hamiltonian and the observable (i.e., known a priori). In either case, the computational effort to evaluate path integrals or expectations of observables reduces to the evaluation of integrals over spheres. Furthermore, computations can be performed directly in the continuum and applications (analytic computations and their implementations) to a number of models including the non-trivial cases of the massive Schwinger model and a φ4 theory.

  8. On entire functions restricted to intervals, partition of unities, and dual Gabor frames

    DEFF Research Database (Denmark)

    Christensen, Ole; Kim, Hong Oh; Kim, Rae Young

    2014-01-01

    Partition of unities appears in many places in analysis. Typically it is generated by compactly supported functions with a certain regularity. In this paper we consider partition of unities obtained as integer-translates of entire functions restricted to finite intervals. We characterize the enti...

  9. L1-norm locally linear representation regularization multi-source adaptation learning.

    Science.gov (United States)

    Tao, Jianwen; Wen, Shiting; Hu, Wenjun

    2015-09-01

    In most supervised domain adaptation learning (DAL) tasks, one has access only to a small number of labeled examples from target domain. Therefore the success of supervised DAL in this "small sample" regime needs the effective utilization of the large amounts of unlabeled data to extract information that is useful for generalization. Toward this end, we here use the geometric intuition of manifold assumption to extend the established frameworks in existing model-based DAL methods for function learning by incorporating additional information about the target geometric structure of the marginal distribution. We would like to ensure that the solution is smooth with respect to both the ambient space and the target marginal distribution. In doing this, we propose a novel L1-norm locally linear representation regularization multi-source adaptation learning framework which exploits the geometry of the probability distribution, which has two techniques. Firstly, an L1-norm locally linear representation method is presented for robust graph construction by replacing the L2-norm reconstruction measure in LLE with L1-norm one, which is termed as L1-LLR for short. Secondly, considering the robust graph regularization, we replace traditional graph Laplacian regularization with our new L1-LLR graph Laplacian regularization and therefore construct new graph-based semi-supervised learning framework with multi-source adaptation constraint, which is coined as L1-MSAL method. Moreover, to deal with the nonlinear learning problem, we also generalize the L1-MSAL method by mapping the input data points from the input space to a high-dimensional reproducing kernel Hilbert space (RKHS) via a nonlinear mapping. Promising experimental results have been obtained on several real-world datasets such as face, visual video and object. Copyright © 2015 Elsevier Ltd. All rights reserved.

  10. Analysis of regularized inversion of data corrupted by white Gaussian noise

    International Nuclear Information System (INIS)

    Kekkonen, Hanne; Lassas, Matti; Siltanen, Samuli

    2014-01-01

    Tikhonov regularization is studied in the case of linear pseudodifferential operator as the forward map and additive white Gaussian noise as the measurement error. The measurement model for an unknown function u(x) is m(x) = Au(x) + δ ε (x), where δ > 0 is the noise magnitude. If ε was an L 2 -function, Tikhonov regularization gives an estimate T α (m) = u∈H r arg min { ||Au-m|| L 2 2 + α||u|| H r 2 } for u where α = α(δ) is the regularization parameter. Here penalization of the Sobolev norm ||u|| H r covers the cases of standard Tikhonov regularization (r = 0) and first derivative penalty (r = 1). Realizations of white Gaussian noise are almost never in L 2 , but do belong to H s with probability one if s < 0 is small enough. A modification of Tikhonov regularization theory is presented, covering the case of white Gaussian measurement noise. Furthermore, the convergence of regularized reconstructions to the correct solution as δ → 0 is proven in appropriate function spaces using microlocal analysis. The convergence of the related finite-dimensional problems to the infinite-dimensional problem is also analysed. (paper)

  11. 2-regularity and 2-normality conditions for systems with impulsive controls

    Directory of Open Access Journals (Sweden)

    Pavlova Natal'ya

    2007-01-01

    Full Text Available In this paper a controlled system with impulsive controls in the neighborhood of an abnormal point is investigated. The set of pairs (u,μ is considered as a class of admissible controls, where u is a measurable essentially bounded function and μ is a finite-dimensional Borel measure, such that for any Borel set B, μ(B is a subset of the given convex closed pointed cone. In this article the concepts of 2-regularity and 2-normality for the abstract mapping Ф, operating from the given Banach space into a finite-dimensional space, are introduced. The concepts of 2-regularity and 2-normality play a great role in the course of derivation of the first and the second order necessary conditions for the optimal control problem, consisting of the minimization of a certain functional on the set of the admissible processes. These concepts are also important for obtaining the sufficient conditions for the local controllability of the nonlinear systems. The convenient criterion for 2-regularity along the prescribed direction and necessary conditions for 2-normality of systems, linear in control, are introduced in this article as well.

  12. Image degradation characteristics and restoration based on regularization for diffractive imaging

    Science.gov (United States)

    Zhi, Xiyang; Jiang, Shikai; Zhang, Wei; Wang, Dawei; Li, Yun

    2017-11-01

    The diffractive membrane optical imaging system is an important development trend of ultra large aperture and lightweight space camera. However, related investigations on physics-based diffractive imaging degradation characteristics and corresponding image restoration methods are less studied. In this paper, the model of image quality degradation for the diffraction imaging system is first deduced mathematically based on diffraction theory and then the degradation characteristics are analyzed. On this basis, a novel regularization model of image restoration that contains multiple prior constraints is established. After that, the solving approach of the equation with the multi-norm coexistence and multi-regularization parameters (prior's parameters) is presented. Subsequently, the space-variant PSF image restoration method for large aperture diffractive imaging system is proposed combined with block idea of isoplanatic region. Experimentally, the proposed algorithm demonstrates its capacity to achieve multi-objective improvement including MTF enhancing, dispersion correcting, noise and artifact suppressing as well as image's detail preserving, and produce satisfactory visual quality. This can provide scientific basis for applications and possesses potential application prospects on future space applications of diffractive membrane imaging technology.

  13. Progressive image denoising through hybrid graph Laplacian regularization: a unified framework.

    Science.gov (United States)

    Liu, Xianming; Zhai, Deming; Zhao, Debin; Zhai, Guangtao; Gao, Wen

    2014-04-01

    Recovering images from corrupted observations is necessary for many real-world applications. In this paper, we propose a unified framework to perform progressive image recovery based on hybrid graph Laplacian regularized regression. We first construct a multiscale representation of the target image by Laplacian pyramid, then progressively recover the degraded image in the scale space from coarse to fine so that the sharp edges and texture can be eventually recovered. On one hand, within each scale, a graph Laplacian regularization model represented by implicit kernel is learned, which simultaneously minimizes the least square error on the measured samples and preserves the geometrical structure of the image data space. In this procedure, the intrinsic manifold structure is explicitly considered using both measured and unmeasured samples, and the nonlocal self-similarity property is utilized as a fruitful resource for abstracting a priori knowledge of the images. On the other hand, between two successive scales, the proposed model is extended to a projected high-dimensional feature space through explicit kernel mapping to describe the interscale correlation, in which the local structure regularity is learned and propagated from coarser to finer scales. In this way, the proposed algorithm gradually recovers more and more image details and edges, which could not been recovered in previous scale. We test our algorithm on one typical image recovery task: impulse noise removal. Experimental results on benchmark test images demonstrate that the proposed method achieves better performance than state-of-the-art algorithms.

  14. FOREWORD: Tackling inverse problems in a Banach space environment: from theory to applications Tackling inverse problems in a Banach space environment: from theory to applications

    Science.gov (United States)

    Schuster, Thomas; Hofmann, Bernd; Kaltenbacher, Barbara

    2012-10-01

    Inverse problems can usually be modelled as operator equations in infinite-dimensional spaces with a forward operator acting between Hilbert or Banach spaces—a formulation which quite often also serves as the basis for defining and analyzing solution methods. The additional amount of structure and geometric interpretability provided by the concept of an inner product has rendered these methods amenable to a convergence analysis, a fact which has led to a rigorous and comprehensive study of regularization methods in Hilbert spaces over the last three decades. However, for numerous problems such as x-ray diffractometry, certain inverse scattering problems and a number of parameter identification problems in PDEs, the reasons for using a Hilbert space setting seem to be based on conventions rather than an appropriate and realistic model choice, so often a Banach space setting would be closer to reality. Furthermore, non-Hilbertian regularization and data fidelity terms incorporating a priori information on solution and noise, such as general Lp-norms, TV-type norms, or the Kullback-Leibler divergence, have recently become very popular. These facts have motivated intensive investigations on regularization methods in Banach spaces, a topic which has emerged as a highly active research field within the area of inverse problems. Meanwhile some of the most well-known regularization approaches, such as Tikhonov-type methods requiring the solution of extremal problems, and iterative ones like the Landweber method, the Gauss-Newton method, as well as the approximate inverse method, have been investigated for linear and nonlinear operator equations in Banach spaces. Convergence with rates has been proven and conditions on the solution smoothness and on the structure of nonlinearity have been formulated. Still, beyond the existing results a large number of challenging open questions have arisen, due to the more involved handling of general Banach spaces and the larger variety

  15. A NOVEL APPROACH TO ARRHYTHMIA CLASSIFICATION USING RR INTERVAL AND TEAGER ENERGY

    Directory of Open Access Journals (Sweden)

    CHANDRAKAR KAMATH

    2012-12-01

    Full Text Available It is hypothesized that a key characteristic of electrocardiogram (ECG signal is its nonlinear dynamic behaviour and that the nonlinear component changes more significantly between normal and arrhythmia conditions than the linear component. The usual statistical descriptors used in RR (R to R interval analysis do not capture the nonlinear disposition of RR interval variability. In this paper we explore a novel approach to extract the features from nonlinear component of the RR interval signal using Teager energy operator (TEO. The key feature of Teager energy is that it models the energy of the source that generated the signal rather than the energy of the signal itself. Hence any deviations in regular rhythmic activity of the heart get reflected in the Teager energy function. The classification evaluated on MIT-BIH database, with RR interval and mean of Teager energy computed over RR interval as features, exhibits an average accuracy that exceeds 99.79%.

  16. Multi-view clustering via multi-manifold regularized non-negative matrix factorization.

    Science.gov (United States)

    Zong, Linlin; Zhang, Xianchao; Zhao, Long; Yu, Hong; Zhao, Qianli

    2017-04-01

    Non-negative matrix factorization based multi-view clustering algorithms have shown their competitiveness among different multi-view clustering algorithms. However, non-negative matrix factorization fails to preserve the locally geometrical structure of the data space. In this paper, we propose a multi-manifold regularized non-negative matrix factorization framework (MMNMF) which can preserve the locally geometrical structure of the manifolds for multi-view clustering. MMNMF incorporates consensus manifold and consensus coefficient matrix with multi-manifold regularization to preserve the locally geometrical structure of the multi-view data space. We use two methods to construct the consensus manifold and two methods to find the consensus coefficient matrix, which leads to four instances of the framework. Experimental results show that the proposed algorithms outperform existing non-negative matrix factorization based algorithms for multi-view clustering. Copyright © 2017 Elsevier Ltd. All rights reserved.

  17. Instabilities of the zeta-function regularization in the presence of symmetries

    International Nuclear Information System (INIS)

    Rasetti, M.

    1980-01-01

    The zeta-function regularization method requires the calculation of the spectrum-generating function zeta sub(M) of a generic real, elliptic, self-adjoint differential operator on a manifold M. An asymptotic expansion for zeta sub(M) is given for the class of all symmetric spaces of rank 1, sufficient to compute its Mellin transform and deduce the regularization of the corresponding quadratic path integrals. The summability properties of the generalized zeta-function introduce physical instabilities in the system as negative specific heat. The technique (and the instability as well) is shown to hold - under the assumed symmetry properties - in any dimension (preserving both the global and local properties of the manifold, as opposed to the dimensional regularization, where one adds extra flat dimensions only). (author)

  18. Regular character of chromatin degradation in lymphoid tissues after treatment with biological alkylating agents in vivo

    International Nuclear Information System (INIS)

    Matyasova, J.; Skalka, M.; Cejkova, M.

    1979-01-01

    The chromatin changes are reevaluated occurring in lymphoid tissues of mice treated with alkylating agents of the nitrogen-mustard type in relation to recent evidence on the nucleosomal organization of chromatin and to our new data on the regular character of chromatin degradation in lymphoid tissues of irradiated mice. DNA was isolated from nuclei at various intervals (1 to 18 h) after treatment of mice and subjected to gel electrophoresis in polyacrylamide gels. Thymus chromatin from treated mice has been shown to degrade in a regular fashion and to yield discrete DNA fragments, resembling those that originate in lymphoid tissues of irradiated mice or in thymus nuclei digested with micrococcal nuclease in vitro. With increasing interval after treatment higher amounts of smaller DNA fragments appear. Chromatin in spleen cells responds to treatment in a similar way, whilst no degradation in vivo takes place in liver chromatin. Chromatin of LS/BL lymphosarcoma cells in mice treated with alkylating agents or with irradiation suffers from a similar regular degradation. The results stress the significance of the action of liberated or activated endogenous nuclease(s) in the development of chromatin damage in lymphoid cells after treatment with alkylating agents. (author)

  19. Weighted regularized statistical shape space projection for breast 3D model reconstruction.

    Science.gov (United States)

    Ruiz, Guillermo; Ramon, Eduard; García, Jaime; Sukno, Federico M; Ballester, Miguel A González

    2018-05-02

    The use of 3D imaging has increased as a practical and useful tool for plastic and aesthetic surgery planning. Specifically, the possibility of representing the patient breast anatomy in a 3D shape and simulate aesthetic or plastic procedures is a great tool for communication between surgeon and patient during surgery planning. For the purpose of obtaining the specific 3D model of the breast of a patient, model-based reconstruction methods can be used. In particular, 3D morphable models (3DMM) are a robust and widely used method to perform 3D reconstruction. However, if additional prior information (i.e., known landmarks) is combined with the 3DMM statistical model, shape constraints can be imposed to improve the 3DMM fitting accuracy. In this paper, we present a framework to fit a 3DMM of the breast to two possible inputs: 2D photos and 3D point clouds (scans). Our method consists in a Weighted Regularized (WR) projection into the shape space. The contribution of each point in the 3DMM shape is weighted allowing to assign more relevance to those points that we want to impose as constraints. Our method is applied at multiple stages of the 3D reconstruction process. Firstly, it can be used to obtain a 3DMM initialization from a sparse set of 3D points. Additionally, we embed our method in the 3DMM fitting process in which more reliable or already known 3D points or regions of points, can be weighted in order to preserve their shape information. The proposed method has been tested in two different input settings: scans and 2D pictures assessing both reconstruction frameworks with very positive results. Copyright © 2018 Elsevier B.V. All rights reserved.

  20. Manifold regularized multitask learning for semi-supervised multilabel image classification.

    Science.gov (United States)

    Luo, Yong; Tao, Dacheng; Geng, Bo; Xu, Chao; Maybank, Stephen J

    2013-02-01

    It is a significant challenge to classify images with multiple labels by using only a small number of labeled samples. One option is to learn a binary classifier for each label and use manifold regularization to improve the classification performance by exploring the underlying geometric structure of the data distribution. However, such an approach does not perform well in practice when images from multiple concepts are represented by high-dimensional visual features. Thus, manifold regularization is insufficient to control the model complexity. In this paper, we propose a manifold regularized multitask learning (MRMTL) algorithm. MRMTL learns a discriminative subspace shared by multiple classification tasks by exploiting the common structure of these tasks. It effectively controls the model complexity because different tasks limit one another's search volume, and the manifold regularization ensures that the functions in the shared hypothesis space are smooth along the data manifold. We conduct extensive experiments, on the PASCAL VOC'07 dataset with 20 classes and the MIR dataset with 38 classes, by comparing MRMTL with popular image classification algorithms. The results suggest that MRMTL is effective for image classification.

  1. Fermion-number violation in regularizations that preserve fermion-number symmetry

    Science.gov (United States)

    Golterman, Maarten; Shamir, Yigal

    2003-01-01

    There exist both continuum and lattice regularizations of gauge theories with fermions which preserve chiral U(1) invariance (“fermion number”). Such regularizations necessarily break gauge invariance but, in a covariant gauge, one recovers gauge invariance to all orders in perturbation theory by including suitable counterterms. At the nonperturbative level, an apparent conflict then arises between the chiral U(1) symmetry of the regularized theory and the existence of ’t Hooft vertices in the renormalized theory. The only possible resolution of the paradox is that the chiral U(1) symmetry is broken spontaneously in the enlarged Hilbert space of the covariantly gauge-fixed theory. The corresponding Goldstone pole is unphysical. The theory must therefore be defined by introducing a small fermion-mass term that breaks explicitly the chiral U(1) invariance and is sent to zero after the infinite-volume limit has been taken. Using this careful definition (and a lattice regularization) for the calculation of correlation functions in the one-instanton sector, we show that the ’t Hooft vertices are recovered as expected.

  2. Asymptotic analysis of a pile-up of regular edge dislocation walls

    KAUST Repository

    Hall, Cameron L.

    2011-12-01

    The idealised problem of a pile-up of regular dislocation walls (that is, of planes each containing an infinite number of parallel, identical and equally spaced dislocations) was presented by Roy et al. [A. Roy, R.H.J. Peerlings, M.G.D. Geers, Y. Kasyanyuk, Materials Science and Engineering A 486 (2008) 653-661] as a prototype for understanding the importance of discrete dislocation interactions in dislocation-based plasticity models. They noted that analytic solutions for the dislocation wall density are available for a pile-up of regular screw dislocation walls, but that numerical methods seem to be necessary for investigating regular edge dislocation walls. In this paper, we use the techniques of discrete-to-continuum asymptotic analysis to obtain a detailed description of a pile-up of regular edge dislocation walls. To leading order, we find that the dislocation wall density is governed by a simple differential equation and that boundary layers are present at both ends of the pile-up. © 2011 Elsevier B.V.

  3. Asymptotic analysis of a pile-up of regular edge dislocation walls

    KAUST Repository

    Hall, Cameron L.

    2011-01-01

    The idealised problem of a pile-up of regular dislocation walls (that is, of planes each containing an infinite number of parallel, identical and equally spaced dislocations) was presented by Roy et al. [A. Roy, R.H.J. Peerlings, M.G.D. Geers, Y. Kasyanyuk, Materials Science and Engineering A 486 (2008) 653-661] as a prototype for understanding the importance of discrete dislocation interactions in dislocation-based plasticity models. They noted that analytic solutions for the dislocation wall density are available for a pile-up of regular screw dislocation walls, but that numerical methods seem to be necessary for investigating regular edge dislocation walls. In this paper, we use the techniques of discrete-to-continuum asymptotic analysis to obtain a detailed description of a pile-up of regular edge dislocation walls. To leading order, we find that the dislocation wall density is governed by a simple differential equation and that boundary layers are present at both ends of the pile-up. © 2011 Elsevier B.V.

  4. Does regular practice of physical activity reduce the risk of dysphonia?

    Science.gov (United States)

    Assunção, Ada Avila; de Medeiros, Adriane Mesquita; Barreto, Sandhi Maria; Gama, Ana Cristina Cortes

    2009-12-01

    The purpose of this study was to investigate the association between regular physical activity and the prevalence of dysphonia. A cross-sectional study was conducted with 3142 teachers from 129 municipal public schools in the city of Belo Horizonte, Brazil. The dependent variable, dysphonia, was classified (absent or present) according to reported symptoms (fatigue when speaking and loss of voice quality), their frequency (occasionally and daily), and duration (past 15 days). The independent variable was regular physical activity. The degree of association was estimated based on the prevalence ratio and a 95% confidence interval obtained by the Poisson regression adapted for cross-sectional studies. In the study sample, the prevalence of dysphonia in teachers was 15.63%. Nearly half (47.52%) of the teachers reported no regular practice of physical exercises. The remaining teachers (52.48%) walked and did physical exercises, sports, and other activities; 31.25% undertook these activities once or twice a week, and 21.23% exercised three or more times a week. Teachers who did not practice physical activity were more likely to present dysphonia compared to those that exercised three or more times a week. Regular physical activity was associated positively with the prevalence of dysphonia.

  5. On the structure of space-time caustics

    International Nuclear Information System (INIS)

    Rosquist, K.

    1983-01-01

    Caustics formed by timelike and null geodesics in a space-time M are investigated. Care is taken to distinguish the conjugate points in the tangent space (T-conjugate points) from conjugate points in the manifold (M-conjugate points). It is shown that most nonspacelike conjugate points are regular, i.e. with all neighbouring conjugate points having the same degree of degeneracy. The regular timelike T-conjugate locus is shown to be a smooth 3-dimensional submanifold of the tangent space. Analogously, the regular null T-conjugate locus is shown to be a smooth 2-dimensional submanifold of the light cone in the tangent space. The smoothness properties of the null caustic are used to show that if an observer sees focusing in all directions, then there will necessarily be a cusp in the caustic. If, in addition, all the null conjugate points have maximal degree of degeneracy (as in the closed Friedmann-Robertson-Walker universes), then the space-time is closed. (orig.)

  6. Double Sequences and Iterated Limits in Regular Space

    Directory of Open Access Journals (Sweden)

    Coghetto Roland

    2016-09-01

    Full Text Available First, we define in Mizar [5], the Cartesian product of two filters bases and the Cartesian product of two filters. After comparing the product of two Fréchet filters on ℕ (F1 with the Fréchet filter on ℕ × ℕ (F2, we compare limF₁ and limF₂ for all double sequences in a non empty topological space.

  7. Extreme values, regular variation and point processes

    CERN Document Server

    Resnick, Sidney I

    1987-01-01

    Extremes Values, Regular Variation and Point Processes is a readable and efficient account of the fundamental mathematical and stochastic process techniques needed to study the behavior of extreme values of phenomena based on independent and identically distributed random variables and vectors It presents a coherent treatment of the distributional and sample path fundamental properties of extremes and records It emphasizes the core primacy of three topics necessary for understanding extremes the analytical theory of regularly varying functions; the probabilistic theory of point processes and random measures; and the link to asymptotic distribution approximations provided by the theory of weak convergence of probability measures in metric spaces The book is self-contained and requires an introductory measure-theoretic course in probability as a prerequisite Almost all sections have an extensive list of exercises which extend developments in the text, offer alternate approaches, test mastery and provide for enj...

  8. Distance-regular graphs

    NARCIS (Netherlands)

    van Dam, Edwin R.; Koolen, Jack H.; Tanaka, Hajime

    2016-01-01

    This is a survey of distance-regular graphs. We present an introduction to distance-regular graphs for the reader who is unfamiliar with the subject, and then give an overview of some developments in the area of distance-regular graphs since the monograph 'BCN'[Brouwer, A.E., Cohen, A.M., Neumaier,

  9. Perturbative formulation of pure space-like axial gauge QED with infrared divergences regularized by residual gauge fields

    International Nuclear Information System (INIS)

    Nakawaki, Yuji; McCartor, Gary

    2006-01-01

    We construct a new perturbative formulation of pure space-like axial gauge QED in which the inherent infrared divergences are regularized by residual gauge fields. For this purpose, we carry out our calculations in the coordinates x μ =(x + , x - , x 1 , x 2 ), where x + =x 0 sinθ + x 3 cosθ and x - = x 0 cosθ - x 3 sinθ. Here, A=A 0 cosθ + A 3 sinθ = n·A=0 is taken as the gauge fixing condition. We show in detail that, in perturbation theory, infrared divergences resulting from the residual gauge fields cancel infrared divergences resulting from the physical parts of the gauge field. As a result, we obtain the gauge field propagator proposed by Mandelstam and Leibbrandt. By taking the limit θ→π/4, we are able to construct a light-cone formulation that is free from infrared divergences. With that analysis complete, we next calculate the one-loop electron self-energy, something not previously done in the light-cone quantization and light-cone gauge. (author)

  10. Regular expressions cookbook

    CERN Document Server

    Goyvaerts, Jan

    2009-01-01

    This cookbook provides more than 100 recipes to help you crunch data and manipulate text with regular expressions. Every programmer can find uses for regular expressions, but their power doesn't come worry-free. Even seasoned users often suffer from poor performance, false positives, false negatives, or perplexing bugs. Regular Expressions Cookbook offers step-by-step instructions for some of the most common tasks involving this tool, with recipes for C#, Java, JavaScript, Perl, PHP, Python, Ruby, and VB.NET. With this book, you will: Understand the basics of regular expressions through a

  11. HIV intertest interval among MSM in King County, Washington.

    Science.gov (United States)

    Katz, David A; Dombrowski, Julia C; Swanson, Fred; Buskin, Susan E; Golden, Matthew R; Stekler, Joanne D

    2013-02-01

    The authors examined temporal trends and correlates of HIV testing frequency among men who have sex with men (MSM) in King County, Washington. The authors evaluated data from MSM testing for HIV at the Public Health-Seattle & King County (PHSKC) STD Clinic and Gay City Health Project (GCHP) and testing history data from MSM in PHSKC HIV surveillance. The intertest interval (ITI) was defined as the number of days between the last negative HIV test and the current testing visit or first positive test. Correlates of the log(10)-transformed ITI were determined using generalised estimating equations linear regression. Between 2003 and 2010, the median ITI among MSM seeking HIV testing at the STD Clinic and GCHP were 215 (IQR: 124-409) and 257 (IQR: 148-503) days, respectively. In multivariate analyses, younger age, having only male partners and reporting ≥10 male sex partners in the last year were associated with shorter ITIs at both testing sites (pGCHP attendees, having a regular healthcare provider, seeking a test as part of a regular schedule and inhaled nitrite use in the last year were also associated with shorter ITIs (pGCHP (median 359 vs 255 days, p=0.02). Although MSM in King County appear to be testing at frequent intervals, further efforts are needed to reduce the time that HIV-infected persons are unaware of their status.

  12. Asymptotic properties of spherically symmetric, regular and static solutions to Yang-Mills equations

    International Nuclear Information System (INIS)

    Cronstrom, C.

    1987-01-01

    In this paper the author discusses the asymptotic properties of solutions to Yang-Mills equations with the gauge group SU(2), for spherically symmetric, regular and static potentials. It is known, that the pure Yang-Mills equations cannot have nontrivial regular solutions which vanish rapidly at space infinity (socalled finite energy solutions). So, if regular solutions exist, they must have non-trivial asymptotic properties. However, if the asymptotic behaviour of the solutions is non-trivial, then the fact must be explicitly taken into account in constructing the proper action (and energy) for the theory. The elucidation of the appropriate surface correction to the Yang-Mills action (and hence the energy-momentum tensor density) is one of the main motivations behind the present study. In this paper the author restricts to the asymptotic behaviour of the static solutions. It is shown that this asymptotic behaviour is such that surface corrections (at space-infinity) are needed in order to obtain a well-defined (classical) theory. This is of relevance in formulating a quantum Yang-Mills theory

  13. Comparison of Controller and Flight Deck Algorithm Performance During Interval Management with Dynamic Arrival Trees (STARS)

    Science.gov (United States)

    Battiste, Vernol; Lawton, George; Lachter, Joel; Brandt, Summer; Koteskey, Robert; Dao, Arik-Quang; Kraut, Josh; Ligda, Sarah; Johnson, Walter W.

    2012-01-01

    Managing the interval between arrival aircraft is a major part of the en route and TRACON controller s job. In an effort to reduce controller workload and low altitude vectoring, algorithms have been developed to allow pilots to take responsibility for, achieve and maintain proper spacing. Additionally, algorithms have been developed to create dynamic weather-free arrival routes in the presence of convective weather. In a recent study we examined an algorithm to handle dynamic re-routing in the presence of convective weather and two distinct spacing algorithms. The spacing algorithms originated from different core algorithms; both were enhanced with trajectory intent data for the study. These two algorithms were used simultaneously in a human-in-the-loop (HITL) simulation where pilots performed weather-impacted arrival operations into Louisville International Airport while also performing interval management (IM) on some trials. The controllers retained responsibility for separation and for managing the en route airspace and some trials managing IM. The goal was a stress test of dynamic arrival algorithms with ground and airborne spacing concepts. The flight deck spacing algorithms or controller managed spacing not only had to be robust to the dynamic nature of aircraft re-routing around weather but also had to be compatible with two alternative algorithms for achieving the spacing goal. Flight deck interval management spacing in this simulation provided a clear reduction in controller workload relative to when controllers were responsible for spacing the aircraft. At the same time, spacing was much less variable with the flight deck automated spacing. Even though the approaches taken by the two spacing algorithms to achieve the interval management goals were slightly different they seem to be simpatico in achieving the interval management goal of 130 sec by the TRACON boundary.

  14. Tunneling into quantum wires: regularization of the tunneling Hamiltonian and consistency between free and bosonized fermions

    OpenAIRE

    Filippone, Michele; Brouwer, Piet

    2016-01-01

    Tunneling between a point contact and a one-dimensional wire is usually described with the help of a tunneling Hamiltonian that contains a delta function in position space. Whereas the leading order contribution to the tunneling current is independent of the way this delta function is regularized, higher-order corrections with respect to the tunneling amplitude are known to depend on the regularization. Instead of regularizing the delta function in the tunneling Hamiltonian, one may also obta...

  15. LL-regular grammars

    NARCIS (Netherlands)

    Nijholt, Antinus

    1980-01-01

    Culik II and Cogen introduced the class of LR-regular grammars, an extension of the LR(k) grammars. In this paper we consider an analogous extension of the LL(k) grammars called the LL-regular grammars. The relation of this class of grammars to other classes of grammars will be shown. Any LL-regular

  16. Quantification of fetal heart rate regularity using symbolic dynamics

    Science.gov (United States)

    van Leeuwen, P.; Cysarz, D.; Lange, S.; Geue, D.; Groenemeyer, D.

    2007-03-01

    Fetal heart rate complexity was examined on the basis of RR interval time series obtained in the second and third trimester of pregnancy. In each fetal RR interval time series, short term beat-to-beat heart rate changes were coded in 8bit binary sequences. Redundancies of the 28 different binary patterns were reduced by two different procedures. The complexity of these sequences was quantified using the approximate entropy (ApEn), resulting in discrete ApEn values which were used for classifying the sequences into 17 pattern sets. Also, the sequences were grouped into 20 pattern classes with respect to identity after rotation or inversion of the binary value. There was a specific, nonuniform distribution of the sequences in the pattern sets and this differed from the distribution found in surrogate data. In the course of gestation, the number of sequences increased in seven pattern sets, decreased in four and remained unchanged in six. Sequences that occurred less often over time, both regular and irregular, were characterized by patterns reflecting frequent beat-to-beat reversals in heart rate. They were also predominant in the surrogate data, suggesting that these patterns are associated with stochastic heart beat trains. Sequences that occurred more frequently over time were relatively rare in the surrogate data. Some of these sequences had a high degree of regularity and corresponded to prolonged heart rate accelerations or decelerations which may be associated with directed fetal activity or movement or baroreflex activity. Application of the pattern classes revealed that those sequences with a high degree of irregularity correspond to heart rate patterns resulting from complex physiological activity such as fetal breathing movements. The results suggest that the development of the autonomic nervous system and the emergence of fetal behavioral states lead to increases in not only irregular but also regular heart rate patterns. Using symbolic dynamics to

  17. Sparse regularization for force identification using dictionaries

    Science.gov (United States)

    Qiao, Baijie; Zhang, Xingwu; Wang, Chenxi; Zhang, Hang; Chen, Xuefeng

    2016-04-01

    The classical function expansion method based on minimizing l2-norm of the response residual employs various basis functions to represent the unknown force. Its difficulty lies in determining the optimum number of basis functions. Considering the sparsity of force in the time domain or in other basis space, we develop a general sparse regularization method based on minimizing l1-norm of the coefficient vector of basis functions. The number of basis functions is adaptively determined by minimizing the number of nonzero components in the coefficient vector during the sparse regularization process. First, according to the profile of the unknown force, the dictionary composed of basis functions is determined. Second, a sparsity convex optimization model for force identification is constructed. Third, given the transfer function and the operational response, Sparse reconstruction by separable approximation (SpaRSA) is developed to solve the sparse regularization problem of force identification. Finally, experiments including identification of impact and harmonic forces are conducted on a cantilever thin plate structure to illustrate the effectiveness and applicability of SpaRSA. Besides the Dirac dictionary, other three sparse dictionaries including Db6 wavelets, Sym4 wavelets and cubic B-spline functions can also accurately identify both the single and double impact forces from highly noisy responses in a sparse representation frame. The discrete cosine functions can also successfully reconstruct the harmonic forces including the sinusoidal, square and triangular forces. Conversely, the traditional Tikhonov regularization method with the L-curve criterion fails to identify both the impact and harmonic forces in these cases.

  18. Regularity criteria for the Navier–Stokes equations based on one component of velocity

    Czech Academy of Sciences Publication Activity Database

    Guo, Z.; Caggio, M.; Skalák, Zdeněk

    2017-01-01

    Roč. 35, June (2017), s. 379-396 ISSN 1468-1218 R&D Projects: GA ČR GA14-02067S Grant - others:Západočeská univerzita(CZ) SGS-2016-003; National Natural Science Foundation of China (CN) 11301394 Institutional support: RVO:67985874 Keywords : Navier–Stokes equations * regularity of solutions * regularity criteria * Anisotropic Lebesgue spaces Subject RIV: BK - Fluid Dynamics OBOR OECD: Fluids and plasma physics (including surface physics) Impact factor: 1.659, year: 2016

  19. Regularity criteria for the Navier–Stokes equations based on one component of velocity

    Czech Academy of Sciences Publication Activity Database

    Guo, Z.; Caggio, M.; Skalák, Zdeněk

    2017-01-01

    Roč. 35, June (2017), s. 379-396 ISSN 1468-1218 R&D Projects: GA ČR GA14-02067S Grant - others:Západočeská univerzita(CZ) SGS-2016-003; National Natural Science Foundation of China(CN) 11301394 Institutional support: RVO:67985874 Keywords : Navier–Stokes equations * regularity of solutions * regularity criteria * Anisotropic Lebesgue spaces Subject RIV: BK - Fluid Dynamics OBOR OECD: Fluids and plasma physics (including surface physics) Impact factor: 1.659, year: 2016

  20. Describing chaotic attractors: Regular and perpetual points

    Science.gov (United States)

    Dudkowski, Dawid; Prasad, Awadhesh; Kapitaniak, Tomasz

    2018-03-01

    We study the concepts of regular and perpetual points for describing the behavior of chaotic attractors in dynamical systems. The idea of these points, which have been recently introduced to theoretical investigations, is thoroughly discussed and extended into new types of models. We analyze the correlation between regular and perpetual points, as well as their relation with phase space, showing the potential usefulness of both types of points in the qualitative description of co-existing states. The ability of perpetual points in finding attractors is indicated, along with its potential cause. The location of chaotic trajectories and sets of considered points is investigated and the study on the stability of systems is shown. The statistical analysis of the observing desired states is performed. We focus on various types of dynamical systems, i.e., chaotic flows with self-excited and hidden attractors, forced mechanical models, and semiconductor superlattices, exhibiting the universality of appearance of the observed patterns and relations.

  1. Interval selection with machine-dependent intervals

    OpenAIRE

    Bohmova K.; Disser Y.; Mihalak M.; Widmayer P.

    2013-01-01

    We study an offline interval scheduling problem where every job has exactly one associated interval on every machine. To schedule a set of jobs, exactly one of the intervals associated with each job must be selected, and the intervals selected on the same machine must not intersect.We show that deciding whether all jobs can be scheduled is NP-complete already in various simple cases. In particular, by showing the NP-completeness for the case when all the intervals associated with the same job...

  2. Interval Running Training Improves Cognitive Flexibility and Aerobic Power of Young Healthy Adults.

    Science.gov (United States)

    Venckunas, Tomas; Snieckus, Audrius; Trinkunas, Eugenijus; Baranauskiene, Neringa; Solianik, Rima; Juodsnukis, Antanas; Streckis, Vytautas; Kamandulis, Sigitas

    2016-08-01

    Venckunas, T, Snieckus, A, Trinkunas, E, Baranauskiene, N, Solianik, R, Juodsnukis, A, Streckis, V, and Kamandulis, S. Interval running training improves cognitive flexibility and aerobic power of young healthy adults. J Strength Cond Res 30(8): 2114-2121, 2016-The benefits of regular physical exercise may well extend beyond the reduction of chronic diseases risk and augmentation of working capacity, to many other aspects of human well-being, including improved cognitive functioning. Although the effects of moderate intensity continuous training on cognitive performance are relatively well studied, the benefits of interval training have not been investigated in this respect so far. The aim of the current study was to assess whether 7 weeks of interval running training is effective at improving both aerobic fitness and cognitive performance. For this purpose, 8 young dinghy sailors (6 boys and 2 girls) completed the interval running program with 200 m and 2,000 m running performance, cycling maximal oxygen uptake, and cognitive function was measured before and after the intervention. The control group consisted of healthy age-matched subjects (8 boys and 2 girls) who continued their active lifestyle and were tested in the same way as the experimental group, but did not complete any regular training. In the experimental group, 200 m and 2,000 m running performance and cycling maximal oxygen uptake increased together with improved results on cognitive flexibility tasks. No changes in the results of short-term and working memory tasks were observed in the experimental group, and no changes in any of the measured indices were evident in the controls. In conclusion, 7 weeks of interval running training improved running performance and cycling aerobic power, and were sufficient to improve the ability to adjust behavior to changing demands in young active individuals.

  3. Regularized quasinormal modes for plasmonic resonators and open cavities

    Science.gov (United States)

    Kamandar Dezfouli, Mohsen; Hughes, Stephen

    2018-03-01

    Optical mode theory and analysis of open cavities and plasmonic particles is an essential component of optical resonator physics, offering considerable insight and efficiency for connecting to classical and quantum optical properties such as the Purcell effect. However, obtaining the dissipative modes in normalized form for arbitrarily shaped open-cavity systems is notoriously difficult, often involving complex spatial integrations, even after performing the necessary full space solutions to Maxwell's equations. The formal solutions are termed quasinormal modes, which are known to diverge in space, and additional techniques are frequently required to obtain more accurate field representations in the far field. In this work, we introduce a finite-difference time-domain technique that can be used to obtain normalized quasinormal modes using a simple dipole-excitation source, and an inverse Green function technique, in real frequency space, without having to perform any spatial integrations. Moreover, we show how these modes are naturally regularized to ensure the correct field decay behavior in the far field, and thus can be used at any position within and outside the resonator. We term these modes "regularized quasinormal modes" and show the reliability and generality of the theory by studying the generalized Purcell factor of dipole emitters near metallic nanoresonators, hybrid devices with metal nanoparticles coupled to dielectric waveguides, as well as coupled cavity-waveguides in photonic crystals slabs. We also directly compare our results with full-dipole simulations of Maxwell's equations without any approximations, and show excellent agreement.

  4. Reducing errors in the GRACE gravity solutions using regularization

    Science.gov (United States)

    Save, Himanshu; Bettadpur, Srinivas; Tapley, Byron D.

    2012-09-01

    solutions (RL04) from the Center for Space Research (CSR). Post-fit residual analysis shows that the regularized solutions fit the data to within the noise level of GRACE. A time series of filtered hydrological model is used to confirm that signal attenuation for basins in the Total Runoff Integrating Pathways (TRIP) database over 320 km radii is less than 1 cm equivalent water height RMS, which is within the noise level of GRACE.

  5. An iterative method for Tikhonov regularization with a general linear regularization operator

    NARCIS (Netherlands)

    Hochstenbach, M.E.; Reichel, L.

    2010-01-01

    Tikhonov regularization is one of the most popular approaches to solve discrete ill-posed problems with error-contaminated data. A regularization operator and a suitable value of a regularization parameter have to be chosen. This paper describes an iterative method, based on Golub-Kahan

  6. Regularization methods for ill-posed problems in multiple Hilbert scales

    International Nuclear Information System (INIS)

    Mazzieri, Gisela L; Spies, Ruben D

    2012-01-01

    Several convergence results in Hilbert scales under different source conditions are proved and orders of convergence and optimal orders of convergence are derived. Also, relations between those source conditions are proved. The concept of a multiple Hilbert scale on a product space is introduced, and regularization methods on these scales are defined, both for the case of a single observation and for the case of multiple observations. In the latter case, it is shown how vector-valued regularization functions in these multiple Hilbert scales can be used. In all cases, convergence is proved and orders and optimal orders of convergence are shown. Finally, some potential applications and open problems are discussed. (paper)

  7. RBSURFpred: Modeling protein accessible surface area in real and binary space using regularized and optimized regression.

    Science.gov (United States)

    Tarafder, Sumit; Toukir Ahmed, Md; Iqbal, Sumaiya; Tamjidul Hoque, Md; Sohel Rahman, M

    2018-03-14

    Accessible surface area (ASA) of a protein residue is an effective feature for protein structure prediction, binding region identification, fold recognition problems etc. Improving the prediction of ASA by the application of effective feature variables is a challenging but explorable task to consider, specially in the field of machine learning. Among the existing predictors of ASA, REGAd 3 p is a highly accurate ASA predictor which is based on regularized exact regression with polynomial kernel of degree 3. In this work, we present a new predictor RBSURFpred, which extends REGAd 3 p on several dimensions by incorporating 58 physicochemical, evolutionary and structural properties into 9-tuple peptides via Chou's general PseAAC, which allowed us to obtain higher accuracies in predicting both real-valued and binary ASA. We have compared RBSURFpred for both real and binary space predictions with state-of-the-art predictors, such as REGAd 3 p and SPIDER2. We also have carried out a rigorous analysis of the performance of RBSURFpred in terms of different amino acids and their properties, and also with biologically relevant case-studies. The performance of RBSURFpred establishes itself as a useful tool for the community. Copyright © 2018 Elsevier Ltd. All rights reserved.

  8. Regular Expression Pocket Reference

    CERN Document Server

    Stubblebine, Tony

    2007-01-01

    This handy little book offers programmers a complete overview of the syntax and semantics of regular expressions that are at the heart of every text-processing application. Ideal as a quick reference, Regular Expression Pocket Reference covers the regular expression APIs for Perl 5.8, Ruby (including some upcoming 1.9 features), Java, PHP, .NET and C#, Python, vi, JavaScript, and the PCRE regular expression libraries. This concise and easy-to-use reference puts a very powerful tool for manipulating text and data right at your fingertips. Composed of a mixture of symbols and text, regular exp

  9. Levels and correlates of non-adherence to WHO recommended inter-birth intervals in Rufiji, Tanzania

    Directory of Open Access Journals (Sweden)

    Exavery Amon

    2012-12-01

    Full Text Available Abstract Background Poorly spaced pregnancies have been documented worldwide to result in adverse maternal and child health outcomes. The World Health Organization (WHO recommends a minimum inter-birth interval of 33 months between two consecutive live births in order to reduce the risk of adverse maternal and child health outcomes. However, birth spacing practices in many developing countries, including Tanzania, remain scantly addressed. Methods Longitudinal data collected in the Rufiji Health and Demographic Surveillance System (HDSS from January 1999 to December 2010 were analyzed to investigate birth spacing practices among women of childbearing age. The outcome variable, non-adherence to the minimum inter-birth interval, constituted all inter-birth intervals Results A total of 15,373 inter-birth intervals were recorded from 8,980 women aged 15–49 years in Rufiji district over the follow-up period of 11 years. The median inter-birth interval was 33.4 months. Of the 15,373 inter-birth intervals, 48.4% were below the WHO recommended minimum length of 33 months between two live births. Non-adherence was associated with younger maternal age, low maternal education, multiple births from the preceding pregnancy, non-health facility delivery of the preceding birth, being an in-migrant resident, multi-parity and being married. Conclusion Generally, one in every two inter-birth intervals among 15–49 year-old women in Rufiji district is poorly spaced, with significant variations by socio-demographic and behavioral characteristics of mothers and newborns. Maternal, newborn and child health services should be improved with a special emphasis on community- and health facility-based optimum birth spacing education in order to enhance health outcomes of mothers and their babies, especially in rural settings.

  10. SQED two-loop beta function in the context of Implicit regularization

    International Nuclear Information System (INIS)

    Cherchiglia, Adriano Lana; Sampaio, Marcos; Nemes, Maria Carolina

    2013-01-01

    Full text: In this work we present the state-of-art for Implicit Regularization (IReg) in the context of supersymmetric theories. IReg is a four-dimensional regularization technique in momentum space which disentangles, in a consistent way at arbitrary order, the divergencies, regularization dependent and finite parts of any Feynman amplitude. Since it does not resort to modifications on the physical space-time dimensions of the underlying quantum field theoretical model, it can be consistently applied to supersymmetric theories. First we describe the technique and present previous results for supersymmetric models: the two-loop beta function for the Wess-Zumino model (both in the component and superfield formalism); the two-loop beta function for Super Yang-Mills (in the superfield formalism using the background field technique). After, we present our calculation of the two-loop beta function for massless and massive SQED using the superfield formalism with and without resorting to the background field technique. We find that only in the second case the two-loop divergence cancels out. We argue it is due to an anomalous Jacobian under the rescaling of the fields in the path-integral which is necessary for the application of the supersymmetric background field technique. We find, however, that in both cases the two-loop coefficients of beta function are non-null. Finally we briefly discuss the anomaly puzzle in the context of our technique. (author)

  11. Effort variation regularization in sound field reproduction

    DEFF Research Database (Denmark)

    Stefanakis, Nick; Jacobsen, Finn; Sarris, Ioannis

    2010-01-01

    In this paper, active control is used in order to reproduce a given sound field in an extended spatial region. A method is proposed which minimizes the reproduction error at a number of control positions with the reproduction sources holding a certain relation within their complex strengths......), and adaptive wave field synthesis (AWFS), both under free-field conditions and in reverberant rooms. It is shown that effort variation regularization overcomes the problems associated with small spaces and with a low ratio of direct to reverberant energy, improving thus the reproduction accuracy...

  12. Differential regularization and renormalization: a new method of calculation in quantum field theory

    International Nuclear Information System (INIS)

    Freedman, D.Z.; Johnson, K.; Latorre, J.I.

    1992-01-01

    Most primitively divergent Feynman diagrams are well defined in x-space but too singular at short distances for transformation to p-space. A new method of regularization is developed in which singular functions are written as derivatives of less singular functions which contain a logarithmic mass scale. The Fourier transform is then defined by formal integration by parts. The procedure is extended to graphs with divergent subgraphs. No explicit cutoff or counterterms are required, and the method automatically delivers renormalized amplitudes which satisfy Callan-Symanzik equations. These features are thoroughly explored in massless φ 4 theory through 3-loop order, and the method yields explicit functional forms for all amplitudes with less difficulty than conventional methods which use dimensional regularization in p-space. The procedure also appears to be compatible with gauge invariance and the chiral structure of the standard model. This aspect is tested in extensive 1-loop calculations which include the Ward identity in quantum electrodynamics, the chiral anomaly, and the background field algorithm in non-abelian gauge theories. (orig.)

  13. Comparing Interval Management Control Laws for Steady-State Errors and String Stability

    Science.gov (United States)

    Weitz, Lesley A.; Swieringa, Kurt A.

    2018-01-01

    Interval Management (IM) is a future airborne spacing concept that leverages avionics to provide speed guidance to an aircraft to achieve and maintain a specified spacing interval from another aircraft. The design of a speed control law to achieve the spacing goal is a key aspect in the research and development of the IM concept. In this paper, two control laws that are used in much of the contemporary IM research are analyzed and compared to characterize steady-state errors and string stability. Numerical results are used to illustrate how the choice of control laws gains impacts the size of steady-state errors and string performance and the potential trade-offs between those performance characteristics.

  14. Regularity of C*-algebras and central sequence algebras

    DEFF Research Database (Denmark)

    Christensen, Martin S.

    The main topic of this thesis is regularity properties of C*-algebras and how these regularity properties are re ected in their associated central sequence algebras. The thesis consists of an introduction followed by four papers [A], [B], [C], [D]. In [A], we show that for the class of simple...... Villadsen algebra of either the rst type with seed space a nite dimensional CW complex, or the second type, tensorial absorption of the Jiang-Su algebra is characterized by the absence of characters on the central sequence algebra. Additionally, in a joint appendix with Joan Bosa, we show that the Villadsen...... algebra of the second type with innite stable rank fails the corona factorization property. In [B], we consider the class of separable C*-algebras which do not admit characters on their central sequence algebra, and show that it has nice permanence properties. We also introduce a new divisibility property...

  15. The Validity of Dimensional Regularization Method on Fractal Spacetime

    Directory of Open Access Journals (Sweden)

    Yong Tao

    2013-01-01

    Full Text Available Svozil developed a regularization method for quantum field theory on fractal spacetime (1987. Such a method can be applied to the low-order perturbative renormalization of quantum electrodynamics but will depend on a conjectural integral formula on non-integer-dimensional topological spaces. The main purpose of this paper is to construct a fractal measure so as to guarantee the validity of the conjectural integral formula.

  16. Prevalence and Correlates of Having a Regular Physician among Women Presenting for Induced Abortion.

    Science.gov (United States)

    Chor, Julie; Hebert, Luciana E; Hasselbacher, Lee A; Whitaker, Amy K

    2016-01-01

    To determine the prevalence and correlates of having a regular physician among women presenting for induced abortion. We conducted a retrospective review of women presenting to an urban, university-based family planning clinic for abortion between January 2008 and September 2011. We conducted bivariate analyses, comparing women with and without a regular physician, and multivariable regression modeling, to identify factors associated with not having a regular physician. Of 834 women, 521 (62.5%) had a regular physician and 313 (37.5%) did not. Women with a prior pregnancy, live birth, or spontaneous abortion were more likely than women without these experiences to have a regular physician. Women with a prior induced abortion were not more likely than women who had never had a prior induced abortion to have a regular physician. Compared with women younger than 18 years, women aged 18 to 26 years were less likely to have a physician (adjusted odds ratio [aOR], 0.25; 95% confidence interval [CI], 0.10-0.62). Women with a prior live birth had increased odds of having a regular physician compared with women without a prior pregnancy (aOR, 1.89; 95% CI, 1.13-3.16). Women without medical/fetal indications and who had not been victims of sexual assault (self-indicated) were less likely to report having a regular physician compared with women with medical/fetal indications (aOR, 0.55; 95% CI, 0.17-0.82). The abortion visit is a point of contact with a large number of women without a regular physician and therefore provides an opportunity to integrate women into health care. Copyright © 2016 Jacobs Institute of Women's Health. Published by Elsevier Inc. All rights reserved.

  17. Time interval approach to the pulsed neutron logging method

    International Nuclear Information System (INIS)

    Zhao Jingwu; Su Weining

    1994-01-01

    The time interval of neighbouring neutrons emitted from a steady state neutron source can be treated as that from a time-dependent neutron source. In the rock space, the neutron flux is given by the neutron diffusion equation and is composed of an infinite terms. Each term s composed of two die-away curves. The delay action is discussed and used to measure the time interval with only one detector in the experiment. Nuclear reactions with the time distribution due to different types of radiations observed in the neutron well-logging methods are presented with a view to getting the rock nuclear parameters from the time interval technique

  18. Feature selection and multi-kernel learning for adaptive graph regularized nonnegative matrix factorization

    KAUST Repository

    Wang, Jim Jing-Yan; Huang, Jianhua Z.; Sun, Yijun; Gao, Xin

    2014-01-01

    by regularizing NMF with a nearest neighbor graph constructed from the input data set. However, GNMF has two main bottlenecks. First, using the original feature space directly to construct the graph is not necessarily optimal because of the noisy and irrelevant

  19. Regular use of aspirin and pancreatic cancer risk

    Directory of Open Access Journals (Sweden)

    Mahoney Martin C

    2002-09-01

    Full Text Available Abstract Background Regular use of aspirin and other non-steroidal anti-inflammatory drugs (NSAIDs has been consistently associated with reduced risk of colorectal cancer and adenoma, and there is some evidence for a protective effect for other types of cancer. As experimental studies reveal a possible role for NSAIDs is reducing the risk of pancreatic cancer, epidemiological studies examining similar associations in human populations become more important. Methods In this hospital-based case-control study, 194 patients with pancreatic cancer were compared to 582 age and sex-matched patients with non-neoplastic conditions to examine the association between aspirin use and risk of pancreatic cancer. All participants received medical services at the Roswell Park Cancer Institute in Buffalo, NY and completed a comprehensive epidemiologic questionnaire that included information on demographics, lifestyle factors and medical history as well as frequency and duration of aspirin use. Patients using at least one tablet per week for at least six months were classified as regular aspirin users. Unconditional logistic regression was used to compute crude and adjusted odds ratios (ORs with 95% confidence intervals (CIs. Results Pancreatic cancer risk in aspirin users was not changed relative to non-users (adjusted OR = 1.00; 95% CI 0.72–1.39. No significant change in risk was found in relation to greater frequency or prolonged duration of use, in the total sample or in either gender. Conclusions These data suggest that regular aspirin use may not be associated with lower risk of pancreatic cancer.

  20. Interpregnancy interval and risk of autistic disorder.

    Science.gov (United States)

    Gunnes, Nina; Surén, Pål; Bresnahan, Michaeline; Hornig, Mady; Lie, Kari Kveim; Lipkin, W Ian; Magnus, Per; Nilsen, Roy Miodini; Reichborn-Kjennerud, Ted; Schjølberg, Synnve; Susser, Ezra Saul; Øyen, Anne-Siri; Stoltenberg, Camilla

    2013-11-01

    A recent California study reported increased risk of autistic disorder in children conceived within a year after the birth of a sibling. We assessed the association between interpregnancy interval and risk of autistic disorder using nationwide registry data on pairs of singleton full siblings born in Norway. We defined interpregnancy interval as the time from birth of the first-born child to conception of the second-born child in a sibship. The outcome of interest was autistic disorder in the second-born child. Analyses were restricted to sibships in which the second-born child was born in 1990-2004. Odds ratios (ORs) were estimated by fitting ordinary logistic models and logistic generalized additive models. The study sample included 223,476 singleton full-sibling pairs. In sibships with interpregnancy intervals autistic disorder, compared with 0.13% in the reference category (≥ 36 months). For interpregnancy intervals shorter than 9 months, the adjusted OR of autistic disorder in the second-born child was 2.18 (95% confidence interval 1.42-3.26). The risk of autistic disorder in the second-born child was also increased for interpregnancy intervals of 9-11 months in the adjusted analysis (OR = 1.71 [95% CI = 1.07-2.64]). Consistent with a previous report from California, interpregnancy intervals shorter than 1 year were associated with increased risk of autistic disorder in the second-born child. A possible explanation is depletion of micronutrients in mothers with closely spaced pregnancies.

  1. The geometry of continuum regularization

    International Nuclear Information System (INIS)

    Halpern, M.B.

    1987-03-01

    This lecture is primarily an introduction to coordinate-invariant regularization, a recent advance in the continuum regularization program. In this context, the program is seen as fundamentally geometric, with all regularization contained in regularized DeWitt superstructures on field deformations

  2. The Dedekind completion of C ( X ): an interval-valued functions ...

    African Journals Online (AJOL)

    In his paper [1] R. Anguelov described the construction of the Dedekind order completion of C(X) the set of all real-valued continuous functions defined on a completely regular topological space X; using Hausdorff continuous real intervalvalued functions. The aim of this paper is to show that Anguelov's construction can be ...

  3. Shape-constrained regularization by statistical multiresolution for inverse problems: asymptotic analysis

    International Nuclear Information System (INIS)

    Frick, Klaus; Marnitz, Philipp; Munk, Axel

    2012-01-01

    This paper is concerned with a novel regularization technique for solving linear ill-posed operator equations in Hilbert spaces from data that are corrupted by white noise. We combine convex penalty functionals with extreme-value statistics of projections of the residuals on a given set of sub-spaces in the image space of the operator. We prove general consistency and convergence rate results in the framework of Bregman divergences which allows for a vast range of penalty functionals. Various examples that indicate the applicability of our approach will be discussed. We will illustrate in the context of signal and image processing that the presented method constitutes a locally adaptive reconstruction method. (paper)

  4. Guangxi crustal structural evolution and the formation and distribution regularities of U-rich strata

    International Nuclear Information System (INIS)

    Kang Zili.

    1989-01-01

    Based on summing up Guangxi geotectonic features and evolutionary regularities, this paper discusses the occurrence features, formation conditions and time-space distribution regularities of various U-rich strata during the development of geosyncline, platform and diwa stages, Especially, during diwa stage all those U-rich strata might be reworked to a certain degree and resulted in the mobilization of uranium, then enriching to form polygenetic composite uranium ore deposits with stratabound features. This study will be helpful for prospecting in the region

  5. Regular expression containment

    DEFF Research Database (Denmark)

    Henglein, Fritz; Nielsen, Lasse

    2011-01-01

    We present a new sound and complete axiomatization of regular expression containment. It consists of the conventional axiomatiza- tion of concatenation, alternation, empty set and (the singleton set containing) the empty string as an idempotent semiring, the fixed- point rule E* = 1 + E × E......* for Kleene-star, and a general coin- duction rule as the only additional rule. Our axiomatization gives rise to a natural computational inter- pretation of regular expressions as simple types that represent parse trees, and of containment proofs as coercions. This gives the axiom- atization a Curry......-Howard-style constructive interpretation: Con- tainment proofs do not only certify a language-theoretic contain- ment, but, under our computational interpretation, constructively transform a membership proof of a string in one regular expres- sion into a membership proof of the same string in another regular expression. We...

  6. Rule-based learning of regular past tense in children with specific language impairment.

    Science.gov (United States)

    Smith-Lock, Karen M

    2015-01-01

    The treatment of children with specific language impairment was used as a means to investigate whether a single- or dual-mechanism theory best conceptualizes the acquisition of English past tense. The dual-mechanism theory proposes that regular English past-tense forms are produced via a rule-based process whereas past-tense forms of irregular verbs are stored in the lexicon. Single-mechanism theories propose that both regular and irregular past-tense verbs are stored in the lexicon. Five 5-year-olds with specific language impairment received treatment for regular past tense. The children were tested on regular past-tense production and third-person singular "s" twice before treatment and once after treatment, at eight-week intervals. Treatment consisted of one-hour play-based sessions, once weekly, for eight weeks. Crucially, treatment focused on different lexical items from those in the test. Each child demonstrated significant improvement on the untreated past-tense test items after treatment, but no improvement on the untreated third-person singular "s". Generalization to untreated past-tense verbs could not be attributed to a frequency effect or to phonological similarity of trained and tested items. It is argued that the results are consistent with a dual-mechanism theory of past-tense inflection.

  7. Decoupling of modeling and measuring interval in groundwater time series analysis based on response characteristics

    NARCIS (Netherlands)

    Berendrecht, W.L.; Heemink, A.W.; Geer, F.C. van; Gehrels, J.C.

    2003-01-01

    A state-space representation of the transfer function-noise (TFN) model allows the choice of a modeling (input) interval that is smaller than the measuring interval of the output variable. Since in geohydrological applications the interval of the available input series (precipitation excess) is

  8. Regularization by External Variables

    DEFF Research Database (Denmark)

    Bossolini, Elena; Edwards, R.; Glendinning, P. A.

    2016-01-01

    Regularization was a big topic at the 2016 CRM Intensive Research Program on Advances in Nonsmooth Dynamics. There are many open questions concerning well known kinds of regularization (e.g., by smoothing or hysteresis). Here, we propose a framework for an alternative and important kind of regula......Regularization was a big topic at the 2016 CRM Intensive Research Program on Advances in Nonsmooth Dynamics. There are many open questions concerning well known kinds of regularization (e.g., by smoothing or hysteresis). Here, we propose a framework for an alternative and important kind...

  9. Regularization scheme dependence of virtual corrections to DY and DIS

    International Nuclear Information System (INIS)

    Khalafi, F.; Landshoff, P.V.

    1981-01-01

    One loop virtual corrections to the quark photon vertex are calculated under various assumptions and their sensitivity to the manner in which infra-red and mass singularities are regularized is studied. A method based on the use of Mellin-transforms in the Feynman parametric space is developed and shown to be convenient in calculating virtual diagrams beyond the leading logarithm in perturbative QCD. (orig.)

  10. Regular Single Valued Neutrosophic Hypergraphs

    Directory of Open Access Journals (Sweden)

    Muhammad Aslam Malik

    2016-12-01

    Full Text Available In this paper, we define the regular and totally regular single valued neutrosophic hypergraphs, and discuss the order and size along with properties of regular and totally regular single valued neutrosophic hypergraphs. We also extend work on completeness of single valued neutrosophic hypergraphs.

  11. Regular and chaotic dynamics in time-dependent relativistic mean-field theory

    International Nuclear Information System (INIS)

    Vretenar, D.; Ring, P.; Lalazissis, G.A.; Poeschl, W.

    1997-01-01

    Isoscalar and isovector monopole oscillations that correspond to giant resonances in spherical nuclei are described in the framework of time-dependent relativistic mean-field theory. Time-dependent and self-consistent calculations that reproduce experimental data on monopole resonances in 208 Pb show that the motion of the collective coordinate is regular for isoscalar oscillations, and that it becomes chaotic when initial conditions correspond to the isovector mode. Regular collective dynamics coexists with chaotic oscillations on the microscopic level. Time histories, Fourier spectra, state-space plots, Poincare sections, autocorrelation functions, and Lyapunov exponents are used to characterize the nonlinear system and to identify chaotic oscillations. Analogous considerations apply to higher multipolarities. copyright 1997 The American Physical Society

  12. Enhanced manifold regularization for semi-supervised classification.

    Science.gov (United States)

    Gan, Haitao; Luo, Zhizeng; Fan, Yingle; Sang, Nong

    2016-06-01

    Manifold regularization (MR) has become one of the most widely used approaches in the semi-supervised learning field. It has shown superiority by exploiting the local manifold structure of both labeled and unlabeled data. The manifold structure is modeled by constructing a Laplacian graph and then incorporated in learning through a smoothness regularization term. Hence the labels of labeled and unlabeled data vary smoothly along the geodesics on the manifold. However, MR has ignored the discriminative ability of the labeled and unlabeled data. To address the problem, we propose an enhanced MR framework for semi-supervised classification in which the local discriminative information of the labeled and unlabeled data is explicitly exploited. To make full use of labeled data, we firstly employ a semi-supervised clustering method to discover the underlying data space structure of the whole dataset. Then we construct a local discrimination graph to model the discriminative information of labeled and unlabeled data according to the discovered intrinsic structure. Therefore, the data points that may be from different clusters, though similar on the manifold, are enforced far away from each other. Finally, the discrimination graph is incorporated into the MR framework. In particular, we utilize semi-supervised fuzzy c-means and Laplacian regularized Kernel minimum squared error for semi-supervised clustering and classification, respectively. Experimental results on several benchmark datasets and face recognition demonstrate the effectiveness of our proposed method.

  13. Ensemble manifold regularization.

    Science.gov (United States)

    Geng, Bo; Tao, Dacheng; Xu, Chao; Yang, Linjun; Hua, Xian-Sheng

    2012-06-01

    We propose an automatic approximation of the intrinsic manifold for general semi-supervised learning (SSL) problems. Unfortunately, it is not trivial to define an optimization function to obtain optimal hyperparameters. Usually, cross validation is applied, but it does not necessarily scale up. Other problems derive from the suboptimality incurred by discrete grid search and the overfitting. Therefore, we develop an ensemble manifold regularization (EMR) framework to approximate the intrinsic manifold by combining several initial guesses. Algorithmically, we designed EMR carefully so it 1) learns both the composite manifold and the semi-supervised learner jointly, 2) is fully automatic for learning the intrinsic manifold hyperparameters implicitly, 3) is conditionally optimal for intrinsic manifold approximation under a mild and reasonable assumption, and 4) is scalable for a large number of candidate manifold hyperparameters, from both time and space perspectives. Furthermore, we prove the convergence property of EMR to the deterministic matrix at rate root-n. Extensive experiments over both synthetic and real data sets demonstrate the effectiveness of the proposed framework.

  14. Sparsity regularization for parameter identification problems

    International Nuclear Information System (INIS)

    Jin, Bangti; Maass, Peter

    2012-01-01

    The investigation of regularization schemes with sparsity promoting penalty terms has been one of the dominant topics in the field of inverse problems over the last years, and Tikhonov functionals with ℓ p -penalty terms for 1 ⩽ p ⩽ 2 have been studied extensively. The first investigations focused on regularization properties of the minimizers of such functionals with linear operators and on iteration schemes for approximating the minimizers. These results were quickly transferred to nonlinear operator equations, including nonsmooth operators and more general function space settings. The latest results on regularization properties additionally assume a sparse representation of the true solution as well as generalized source conditions, which yield some surprising and optimal convergence rates. The regularization theory with ℓ p sparsity constraints is relatively complete in this setting; see the first part of this review. In contrast, the development of efficient numerical schemes for approximating minimizers of Tikhonov functionals with sparsity constraints for nonlinear operators is still ongoing. The basic iterated soft shrinkage approach has been extended in several directions and semi-smooth Newton methods are becoming applicable in this field. In particular, the extension to more general non-convex, non-differentiable functionals by variational principles leads to a variety of generalized iteration schemes. We focus on such iteration schemes in the second part of this review. A major part of this survey is devoted to applying sparsity constrained regularization techniques to parameter identification problems for partial differential equations, which we regard as the prototypical setting for nonlinear inverse problems. Parameter identification problems exhibit different levels of complexity and we aim at characterizing a hierarchy of such problems. The operator defining these inverse problems is the parameter-to-state mapping. We first summarize some

  15. Metric modular spaces

    CERN Document Server

    Chistyakov, Vyacheslav

    2015-01-01

    Aimed toward researchers and graduate students familiar with elements of functional analysis, linear algebra, and general topology; this book contains a general study of modulars, modular spaces, and metric modular spaces. Modulars may be thought of as generalized velocity fields and serve two important purposes: generate metric spaces in a unified manner and provide a weaker convergence, the modular convergence, whose topology is non-metrizable in general. Metric modular spaces are extensions of metric spaces, metric linear spaces, and classical modular linear spaces. The topics covered include the classification of modulars, metrizability of modular spaces, modular transforms and duality between modular spaces, metric  and modular topologies. Applications illustrated in this book include: the description of superposition operators acting in modular spaces, the existence of regular selections of set-valued mappings, new interpretations of spaces of Lipschitzian and absolutely continuous mappings, the existe...

  16. On a correspondence between regular and non-regular operator monotone functions

    DEFF Research Database (Denmark)

    Gibilisco, P.; Hansen, Frank; Isola, T.

    2009-01-01

    We prove the existence of a bijection between the regular and the non-regular operator monotone functions satisfying a certain functional equation. As an application we give a new proof of the operator monotonicity of certain functions related to the Wigner-Yanase-Dyson skew information....

  17. A two-way regularization method for MEG source reconstruction

    KAUST Repository

    Tian, Tian Siva; Huang, Jianhua Z.; Shen, Haipeng; Li, Zhimin

    2012-01-01

    The MEG inverse problem refers to the reconstruction of the neural activity of the brain from magnetoencephalography (MEG) measurements. We propose a two-way regularization (TWR) method to solve the MEG inverse problem under the assumptions that only a small number of locations in space are responsible for the measured signals (focality), and each source time course is smooth in time (smoothness). The focality and smoothness of the reconstructed signals are ensured respectively by imposing a sparsity-inducing penalty and a roughness penalty in the data fitting criterion. A two-stage algorithm is developed for fast computation, where a raw estimate of the source time course is obtained in the first stage and then refined in the second stage by the two-way regularization. The proposed method is shown to be effective on both synthetic and real-world examples. © Institute of Mathematical Statistics, 2012.

  18. A two-way regularization method for MEG source reconstruction

    KAUST Repository

    Tian, Tian Siva

    2012-09-01

    The MEG inverse problem refers to the reconstruction of the neural activity of the brain from magnetoencephalography (MEG) measurements. We propose a two-way regularization (TWR) method to solve the MEG inverse problem under the assumptions that only a small number of locations in space are responsible for the measured signals (focality), and each source time course is smooth in time (smoothness). The focality and smoothness of the reconstructed signals are ensured respectively by imposing a sparsity-inducing penalty and a roughness penalty in the data fitting criterion. A two-stage algorithm is developed for fast computation, where a raw estimate of the source time course is obtained in the first stage and then refined in the second stage by the two-way regularization. The proposed method is shown to be effective on both synthetic and real-world examples. © Institute of Mathematical Statistics, 2012.

  19. Stochastic analytic regularization

    International Nuclear Information System (INIS)

    Alfaro, J.

    1984-07-01

    Stochastic regularization is reexamined, pointing out a restriction on its use due to a new type of divergence which is not present in the unregulated theory. Furthermore, we introduce a new form of stochastic regularization which permits the use of a minimal subtraction scheme to define the renormalized Green functions. (author)

  20. Analytic semigroups and optimal regularity in parabolic problems

    CERN Document Server

    Lunardi, Alessandra

    2012-01-01

    The book shows how the abstract methods of analytic semigroups and evolution equations in Banach spaces can be fruitfully applied to the study of parabolic problems. Particular attention is paid to optimal regularity results in linear equations. Furthermore, these results are used to study several other problems, especially fully nonlinear ones. Owing to the new unified approach chosen, known theorems are presented from a novel perspective and new results are derived. The book is self-contained. It is addressed to PhD students and researchers interested in abstract evolution equations and in p

  1. Functional differential equations with unbounded delay in extrapolation spaces

    Directory of Open Access Journals (Sweden)

    Mostafa Adimy

    2014-08-01

    Full Text Available We study the existence, regularity and stability of solutions for nonlinear partial neutral functional differential equations with unbounded delay and a Hille-Yosida operator on a Banach space X. We consider two nonlinear perturbations: the first one is a function taking its values in X and the second one is a function belonging to a space larger than X, an extrapolated space. We use the extrapolation techniques to prove the existence and regularity of solutions and we establish a linearization principle for the stability of the equilibria of our equation.

  2. High-intensity interval training improves insulin sensitivity in older individuals

    DEFF Research Database (Denmark)

    Søgaard, D; Lund, M T; Scheuer, C M

    2017-01-01

    AIM: Metabolic health may deteriorate with age as a result of altered body composition and decreased physical activity. Endurance exercise is known to counter these changes delaying or even preventing onset of metabolic diseases. High-intensity interval training (HIIT) is a time efficient...... alternative to regular endurance exercise, and the aim of this study was to investigate the metabolic benefit of HIIT in older subjects. METHODS: Twenty-two sedentary male (n = 11) and female (n = 11) subjects aged 63 ± 1 years performed HIIT training three times/week for 6 weeks on a bicycle ergometer. Each...... HIIT session consisted of five 1-minute intervals interspersed with 1½-minute rest. Prior to the first and after the last HIIT session whole-body insulin sensitivity, measured by a hyperinsulinaemic-euglycaemic clamp, plasma lipid levels, HbA1c, glycaemic parameters, body composition and maximal oxygen...

  3. Image super-resolution reconstruction based on regularization technique and guided filter

    Science.gov (United States)

    Huang, De-tian; Huang, Wei-qin; Gu, Pei-ting; Liu, Pei-zhong; Luo, Yan-min

    2017-06-01

    In order to improve the accuracy of sparse representation coefficients and the quality of reconstructed images, an improved image super-resolution algorithm based on sparse representation is presented. In the sparse coding stage, the autoregressive (AR) regularization and the non-local (NL) similarity regularization are introduced to improve the sparse coding objective function. A group of AR models which describe the image local structures are pre-learned from the training samples, and one or several suitable AR models can be adaptively selected for each image patch to regularize the solution space. Then, the image non-local redundancy is obtained by the NL similarity regularization to preserve edges. In the process of computing the sparse representation coefficients, the feature-sign search algorithm is utilized instead of the conventional orthogonal matching pursuit algorithm to improve the accuracy of the sparse coefficients. To restore image details further, a global error compensation model based on weighted guided filter is proposed to realize error compensation for the reconstructed images. Experimental results demonstrate that compared with Bicubic, L1SR, SISR, GR, ANR, NE + LS, NE + NNLS, NE + LLE and A + (16 atoms) methods, the proposed approach has remarkable improvement in peak signal-to-noise ratio, structural similarity and subjective visual perception.

  4. Regularization in global sound equalization based on effort variation

    DEFF Research Database (Denmark)

    Stefanakis, Nick; Sarris, John; Jacobsen, Finn

    2009-01-01

    . Effort variation equalization involves modifying the conventional cost function in sound equalization, which is based on minimizing least-squares reproduction errors, by adding a term that is proportional to the squared deviations between complex source strengths, calculated independently for the sources......Sound equalization in closed spaces can be significantly improved by generating propagating waves that are naturally associated with the geometry, as, for example, plane waves in rectangular enclosures. This paper presents a control approach termed effort variation regularization based on this idea...

  5. Color correction optimization with hue regularization

    Science.gov (United States)

    Zhang, Heng; Liu, Huaping; Quan, Shuxue

    2011-01-01

    Previous work has suggested that observers are capable of judging the quality of an image without any knowledge of the original scene. When no reference is available, observers can extract the apparent objects in an image and compare them with the typical colors of similar objects recalled from their memories. Some generally agreed upon research results indicate that although perfect colorimetric rendering is not conspicuous and color errors can be well tolerated, the appropriate rendition of certain memory colors such as skin, grass, and sky is an important factor in the overall perceived image quality. These colors are appreciated in a fairly consistent manner and are memorized with slightly different hues and higher color saturation. The aim of color correction for a digital color pipeline is to transform the image data from a device dependent color space to a target color space, usually through a color correction matrix which in its most basic form is optimized through linear regressions between the two sets of data in two color spaces in the sense of minimized Euclidean color error. Unfortunately, this method could result in objectionable distortions if the color error biased certain colors undesirably. In this paper, we propose a color correction optimization method with preferred color reproduction in mind through hue regularization and present some experimental results.

  6. Universal regularization prescription for Lovelock AdS gravity

    International Nuclear Information System (INIS)

    Kofinas, Georgios; Olea, Rodrigo

    2007-01-01

    A definite form for the boundary term that produces the finiteness of both the conserved quantities and Euclidean action for any Lovelock gravity with AdS asymptotics is presented. This prescription merely tells even from odd bulk dimensions, regardless the particular theory considered, what is valid even for Einstein-Hilbert and Einstein-Gauss-Bonnet AdS gravity. The boundary term is a given polynomial of the boundary extrinsic and intrinsic curvatures (also referred to as Kounterterms series). Only the coupling constant of the boundary term changes accordingly, such that it always preserves a well-posed variational principle for boundary conditions suitable for asymptotically AdS spaces. The background-independent conserved charges associated to asymptotic symmetries are found. In odd bulk dimensions, this regularization produces a generalized formula for the vacuum energy in Lovelock AdS gravity. The standard entropy for asymptotically AdS black holes is recovered directly from the regularization of the Euclidean action, and not only from the first law of thermodynamics associated to the conserved quantities

  7. Existence, regularity and representation of solutions of time fractional wave equations

    Directory of Open Access Journals (Sweden)

    Valentin Keyantuo

    2017-09-01

    Full Text Available We study the solvability of the fractional order inhomogeneous Cauchy problem $$ \\mathbb{D}_t^\\alpha u(t=Au(t+f(t, \\quad t>0,\\;1<\\alpha\\le 2, $$ where A is a closed linear operator in some Banach space X and $f:[0,\\infty\\to X$ a given function. Operator families associated with this problem are defined and their regularity properties are investigated. In the case where A is a generator of a $\\beta$-times integrated cosine family $(C_\\beta(t$, we derive explicit representations of mild and classical solutions of the above problem in terms of the integrated cosine family. We include applications to elliptic operators with Dirichlet, Neumann or Robin type boundary conditions on $L^p$-spaces and on the space of continuous functions.

  8. Multiview vector-valued manifold regularization for multilabel image classification.

    Science.gov (United States)

    Luo, Yong; Tao, Dacheng; Xu, Chang; Xu, Chao; Liu, Hong; Wen, Yonggang

    2013-05-01

    In computer vision, image datasets used for classification are naturally associated with multiple labels and comprised of multiple views, because each image may contain several objects (e.g., pedestrian, bicycle, and tree) and is properly characterized by multiple visual features (e.g., color, texture, and shape). Currently, available tools ignore either the label relationship or the view complementarily. Motivated by the success of the vector-valued function that constructs matrix-valued kernels to explore the multilabel structure in the output space, we introduce multiview vector-valued manifold regularization (MV(3)MR) to integrate multiple features. MV(3)MR exploits the complementary property of different features and discovers the intrinsic local geometry of the compact support shared by different features under the theme of manifold regularization. We conduct extensive experiments on two challenging, but popular, datasets, PASCAL VOC' 07 and MIR Flickr, and validate the effectiveness of the proposed MV(3)MR for image classification.

  9. Effective field theory dimensional regularization

    International Nuclear Information System (INIS)

    Lehmann, Dirk; Prezeau, Gary

    2002-01-01

    A Lorentz-covariant regularization scheme for effective field theories with an arbitrary number of propagating heavy and light particles is given. This regularization scheme leaves the low-energy analytic structure of Greens functions intact and preserves all the symmetries of the underlying Lagrangian. The power divergences of regularized loop integrals are controlled by the low-energy kinematic variables. Simple diagrammatic rules are derived for the regularization of arbitrary one-loop graphs and the generalization to higher loops is discussed

  10. Effective field theory dimensional regularization

    Science.gov (United States)

    Lehmann, Dirk; Prézeau, Gary

    2002-01-01

    A Lorentz-covariant regularization scheme for effective field theories with an arbitrary number of propagating heavy and light particles is given. This regularization scheme leaves the low-energy analytic structure of Greens functions intact and preserves all the symmetries of the underlying Lagrangian. The power divergences of regularized loop integrals are controlled by the low-energy kinematic variables. Simple diagrammatic rules are derived for the regularization of arbitrary one-loop graphs and the generalization to higher loops is discussed.

  11. Tuning for temporal interval in human apparent motion detection.

    Science.gov (United States)

    Bours, Roger J E; Stuur, Sanne; Lankheet, Martin J M

    2007-01-08

    Detection of apparent motion in random dot patterns requires correlation across time and space. It has been difficult to study the temporal requirements for the correlation step because motion detection also depends on temporal filtering preceding correlation and on integration at the next levels. To specifically study tuning for temporal interval in the correlation step, we performed an experiment in which prefiltering and postintegration were held constant and in which we used a motion stimulus containing coherent motion for a single interval value only. The stimulus consisted of a sparse random dot pattern in which each dot was presented in two frames only, separated by a specified interval. On each frame, half of the dots were refreshed and the other half was a displaced reincarnation of the pattern generated one or several frames earlier. Motion energy statistics in such a stimulus do not vary from frame to frame, and the directional bias in spatiotemporal correlations is similar for different interval settings. We measured coherence thresholds for left-right direction discrimination by varying motion coherence levels in a Quest staircase procedure, as a function of both step size and interval. Results show that highest sensitivity was found for an interval of 17-42 ms, irrespective of viewing distance. The falloff at longer intervals was much sharper than previously described. Tuning for temporal interval was largely, but not completely, independent of step size. The optimal temporal interval slightly decreased with increasing step size. Similarly, the optimal step size decreased with increasing temporal interval.

  12. UNIVERSAL REGULAR AUTONOMOUS ASYNCHRONOUS SYSTEMS: ω-LIMIT SETS, INVARIANCE AND BASINS OF ATTRACTION

    Directory of Open Access Journals (Sweden)

    Serban Vlad

    2011-07-01

    Full Text Available The asynchronous systems are the non-deterministic real timebinarymodels of the asynchronous circuits from electrical engineering.Autonomy means that the circuits and their models have no input.Regularity means analogies with the dynamical systems, thus such systems may be considered to be real time dynamical systems with a’vector field’, Universality refers to the case when the state space of the system is the greatest possible in the sense of theinclusion. The purpose of this paper is that of defining, by analogy with the dynamical systems theory, the omega-limit sets, the invariance and the basins of attraction of the universal regular autonomous asynchronous systems.

  13. Magnetic Resonance Fingerprinting with short relaxation intervals.

    Science.gov (United States)

    Amthor, Thomas; Doneva, Mariya; Koken, Peter; Sommer, Karsten; Meineke, Jakob; Börnert, Peter

    2017-09-01

    The aim of this study was to investigate a technique for improving the performance of Magnetic Resonance Fingerprinting (MRF) in repetitive sampling schemes, in particular for 3D MRF acquisition, by shortening relaxation intervals between MRF pulse train repetitions. A calculation method for MRF dictionaries adapted to short relaxation intervals and non-relaxed initial spin states is presented, based on the concept of stationary fingerprints. The method is applicable to many different k-space sampling schemes in 2D and 3D. For accuracy analysis, T 1 and T 2 values of a phantom are determined by single-slice Cartesian MRF for different relaxation intervals and are compared with quantitative reference measurements. The relevance of slice profile effects is also investigated in this case. To further illustrate the capabilities of the method, an application to in-vivo spiral 3D MRF measurements is demonstrated. The proposed computation method enables accurate parameter estimation even for the shortest relaxation intervals, as investigated for different sampling patterns in 2D and 3D. In 2D Cartesian measurements, we achieved a scan acceleration of more than a factor of two, while maintaining acceptable accuracy: The largest T 1 values of a sample set deviated from their reference values by 0.3% (longest relaxation interval) and 2.4% (shortest relaxation interval). The largest T 2 values showed systematic deviations of up to 10% for all relaxation intervals, which is discussed. The influence of slice profile effects for multislice acquisition is shown to become increasingly relevant for short relaxation intervals. In 3D spiral measurements, a scan time reduction of 36% was achieved, maintaining the quality of in-vivo T1 and T2 maps. Reducing the relaxation interval between MRF sequence repetitions using stationary fingerprint dictionaries is a feasible method to improve the scan efficiency of MRF sequences. The method enables fast implementations of 3D spatially

  14. 75 FR 76006 - Regular Meeting

    Science.gov (United States)

    2010-12-07

    ... FARM CREDIT SYSTEM INSURANCE CORPORATION Regular Meeting AGENCY: Farm Credit System Insurance Corporation Board. ACTION: Regular meeting. SUMMARY: Notice is hereby given of the regular meeting of the Farm Credit System Insurance Corporation Board (Board). Date and Time: The meeting of the Board will be held...

  15. Regular Topographic Patterning of Karst Depressions Suggests Landscape Self-Organization

    Science.gov (United States)

    Quintero, C.; Cohen, M. J.

    2017-12-01

    Thousands of wetland depressions that are commonly host to cypress domes dot the sub-tropical limestone landscape of South Florida. The origin of these depression features has been the topic of debate. Here we build upon the work of previous surveyors of this landscape to analyze the morphology and spatial distribution of depressions on the Big Cypress landscape. We took advantage of the emergence and availability of high resolution Light Direction and Ranging (LiDAR) technology and ArcMap GIS software to analyze the structure and regularity of landscape features with methods unavailable to past surveyors. Six 2.25 km2 LiDAR plots within the preserve were selected for remote analysis and one depression feature within each plot was selected for more intensive sediment and water depth surveying. Depression features on the Big Cypress landscape were found to show strong evidence of regular spatial patterning. Periodicity, a feature of regularly patterned landscapes, is apparent in both Variograms and Radial Spectrum Analyses. Size class distributions of the identified features indicate constrained feature sizes while Average Nearest Neighbor analyses support the inference of dispersed features with non-random spacing. The presence of regular patterning on this landscape strongly implies biotic reinforcement of spatial structure by way of the scale dependent feedback. In characterizing the structure of this wetland landscape we add to the growing body of work dedicated to documenting how water, life and geology may interact to shape the natural landscapes we see today.

  16. General inverse problems for regular variation

    DEFF Research Database (Denmark)

    Damek, Ewa; Mikosch, Thomas Valentin; Rosinski, Jan

    2014-01-01

    Regular variation of distributional tails is known to be preserved by various linear transformations of some random structures. An inverse problem for regular variation aims at understanding whether the regular variation of a transformed random object is caused by regular variation of components ...

  17. Abel transforms with low regularity with applications to x-ray tomography on spherically symmetric manifolds

    Science.gov (United States)

    de Hoop, Maarten V.; Ilmavirta, Joonas

    2017-12-01

    We study ray transforms on spherically symmetric manifolds with a piecewise C1, 1 metric. Assuming the Herglotz condition, the x-ray transform is injective on the space of L 2 functions on such manifolds. We also prove injectivity results for broken ray transforms (with and without periodicity) on such manifolds with a C1, 1 metric. To make these problems tractable in low regularity, we introduce and study a class of generalized Abel transforms and study their properties. This low regularity setting is relevant for geophysical applications.

  18. Feasibility Criteria for Interval Management Operations as Part of Arrival Management Operations

    Science.gov (United States)

    Levitt, Ian M.; Weitz, Lesley A.; Barmore, Bryan E.; Castle, Michael W.

    2014-01-01

    Interval Management (IM) is a future airborne spacing concept that aims to provide more precise inter-aircraft spacing to yield throughput improvements and greater use of fuel efficient trajectories for arrival and approach operations. To participate in an IM operation, an aircraft must be equipped with avionics that provide speeds to achieve and maintain an assigned spacing interval relative to another aircraft. It is not expected that all aircraft will be equipped with the necessary avionics, but rather that IM fits into a larger arrival management concept developed to support the broader mixed-equipage environment. Arrival management concepts are comprised of three parts: a ground-based sequencing and scheduling function to develop an overall arrival strategy, ground-based tools to support the management of aircraft to that schedule, and the IM tools necessary for the IM operation (i.e., ground-based set-up, initiation, and monitoring, and the flight-deck tools to conduct the IM operation). The Federal Aviation Administration is deploying a near-term ground-automation system to support metering operations in the National Airspace System, which falls within the first two components of the arrival management concept. This paper develops a methodology for determining the required delivery precision at controlled meter points for aircraft that are being managed to a schedule and aircraft being managed to a relative spacing interval in order to achieve desired flow rates and adequate separation at the meter points.

  19. On Orthogonal Decomposition of a Sobolev Space

    OpenAIRE

    Lakew, Dejenie A.

    2016-01-01

    The theme of this short article is to investigate an orthogonal decomposition of a Sobolev space and look at some properties of the inner product therein and the distance defined from the inner product. We also determine the dimension of the orthogonal difference space and show the expansion of spaces as their regularity increases.

  20. Assimilating irregularly spaced sparsely observed turbulent signals with hierarchical Bayesian reduced stochastic filters

    International Nuclear Information System (INIS)

    Brown, Kristen A.; Harlim, John

    2013-01-01

    In this paper, we consider a practical filtering approach for assimilating irregularly spaced, sparsely observed turbulent signals through a hierarchical Bayesian reduced stochastic filtering framework. The proposed hierarchical Bayesian approach consists of two steps, blending a data-driven interpolation scheme and the Mean Stochastic Model (MSM) filter. We examine the potential of using the deterministic piecewise linear interpolation scheme and the ordinary kriging scheme in interpolating irregularly spaced raw data to regularly spaced processed data and the importance of dynamical constraint (through MSM) in filtering the processed data on a numerically stiff state estimation problem. In particular, we test this approach on a two-layer quasi-geostrophic model in a two-dimensional domain with a small radius of deformation to mimic ocean turbulence. Our numerical results suggest that the dynamical constraint becomes important when the observation noise variance is large. Second, we find that the filtered estimates with ordinary kriging are superior to those with linear interpolation when observation networks are not too sparse; such robust results are found from numerical simulations with many randomly simulated irregularly spaced observation networks, various observation time intervals, and observation error variances. Third, when the observation network is very sparse, we find that both the kriging and linear interpolations are comparable

  1. Dynamic MRI Using SmooThness Regularization on Manifolds (SToRM).

    Science.gov (United States)

    Poddar, Sunrita; Jacob, Mathews

    2016-04-01

    We introduce a novel algorithm to recover real time dynamic MR images from highly under-sampled k- t space measurements. The proposed scheme models the images in the dynamic dataset as points on a smooth, low dimensional manifold in high dimensional space. We propose to exploit the non-linear and non-local redundancies in the dataset by posing its recovery as a manifold smoothness regularized optimization problem. A navigator acquisition scheme is used to determine the structure of the manifold, or equivalently the associated graph Laplacian matrix. The estimated Laplacian matrix is used to recover the dataset from undersampled measurements. The utility of the proposed scheme is demonstrated by comparisons with state of the art methods in multi-slice real-time cardiac and speech imaging applications.

  2. Predicting restoration of kidney function during CRRT-free intervals

    Directory of Open Access Journals (Sweden)

    Heise Daniel

    2012-01-01

    Full Text Available Abstract Background Renal failure is common in critically ill patients and frequently requires continuous renal replacement therapy (CRRT. CRRT is discontinued at regular intervals for routine changes of the disposable equipment or for replacing clogged filter membrane assemblies. The present study was conducted to determine if the necessity to continue CRRT could be predicted during the CRRT-free period. Materials and methods In the period from 2003 to 2006, 605 patients were treated with CRRT in our ICU. A total of 222 patients with 448 CRRT-free intervals had complete data sets and were used for analysis. Of the total CRRT-free periods, 225 served as an evaluation group. Twenty-nine parameters with an assumed influence on kidney function were analyzed with regard to their potential to predict the restoration of kidney function during the CRRT-free interval. Using univariate analysis and logistic regression, a prospective index was developed and validated in the remaining 223 CRRT-free periods to establish its prognostic strength. Results Only three parameters showed an independent influence on the restoration of kidney function during CRRT-free intervals: the number of previous CRRT cycles (medians in the two outcome groups: 1 vs. 2, the "Sequential Organ Failure Assessment"-score (means in the two outcome groups: 8.3 vs. 9.2 and urinary output after the cessation of CRRT (medians in two outcome groups: 66 ml/h vs. 10 ml/h. The prognostic index, which was calculated from these three variables, showed a satisfactory potential to predict the kidney function during the CRRT-free intervals; Receiver operating characteristic (ROC analysis revealed an area under the curve of 0.798. Conclusion Restoration of kidney function during CRRT-free periods can be predicted with an index calculated from three variables. Prospective trials in other hospitals must clarify whether our results are generally transferable to other patient populations.

  3. From Discrete Space-Time to Minkowski Space: Basic Mechanisms, Methods and Perspectives

    Science.gov (United States)

    Finster, Felix

    This survey article reviews recent results on fermion systems in discrete space-time and corresponding systems in Minkowski space. After a basic introduction to the discrete setting, we explain a mechanism of spontaneous symmetry breaking which leads to the emergence of a discrete causal structure. As methods to study the transition between discrete space-time and Minkowski space, we describe a lattice model for a static and isotropic space-time, outline the analysis of regularization tails of vacuum Dirac sea configurations, and introduce a Lorentz invariant action for the masses of the Dirac seas. We mention the method of the continuum limit, which allows to analyze interacting systems. Open problems are discussed.

  4. Online co-regularized algorithms

    NARCIS (Netherlands)

    Ruijter, T. de; Tsivtsivadze, E.; Heskes, T.

    2012-01-01

    We propose an online co-regularized learning algorithm for classification and regression tasks. We demonstrate that by sequentially co-regularizing prediction functions on unlabeled data points, our algorithm provides improved performance in comparison to supervised methods on several UCI benchmarks

  5. Geometric continuum regularization of quantum field theory

    International Nuclear Information System (INIS)

    Halpern, M.B.

    1989-01-01

    An overview of the continuum regularization program is given. The program is traced from its roots in stochastic quantization, with emphasis on the examples of regularized gauge theory, the regularized general nonlinear sigma model and regularized quantum gravity. In its coordinate-invariant form, the regularization is seen as entirely geometric: only the supermetric on field deformations is regularized, and the prescription provides universal nonperturbative invariant continuum regularization across all quantum field theory. 54 refs

  6. Development of Salivary Cortisol Circadian Rhythm and Reference Intervals in Full-Term Infants.

    Science.gov (United States)

    Ivars, Katrin; Nelson, Nina; Theodorsson, Annette; Theodorsson, Elvar; Ström, Jakob O; Mörelius, Evalotte

    2015-01-01

    Cortisol concentrations in plasma display a circadian rhythm in adults and children older than one year. Earlier studies report divergent results regarding when cortisol circadian rhythm is established. The present study aims to investigate at what age infants develop a circadian rhythm, as well as the possible influences of behavioral regularity and daily life trauma on when the rhythm is established. Furthermore, we determine age-related reference intervals for cortisol concentrations in saliva during the first year of life. 130 healthy full-term infants were included in a prospective, longitudinal study with saliva sampling on two consecutive days, in the morning (07:30-09:30), noon (10:00-12:00) and evening (19:30-21:30), each month from birth until the infant was twelve months old. Information about development of behavioral regularity and potential exposure to trauma was obtained from the parents through the Baby Behavior Questionnaire and the Life Incidence of Traumatic Events checklist. A significant group-level circadian rhythm of salivary cortisol secretion was established at one month, and remained throughout the first year of life, although there was considerable individual variability. No correlation was found between development of cortisol circadian rhythm and the results from either the Baby Behavior Questionnaire or the Life Incidence of Traumatic Events checklist. The study presents salivary cortisol reference intervals for infants during the first twelve months of life. Cortisol circadian rhythm in infants is already established by one month of age, earlier than previous studies have shown. The current study also provides first year age-related reference intervals for salivary cortisol levels in healthy, full-term infants.

  7. Bypassing the Limits of Ll Regularization: Convex Sparse Signal Processing Using Non-Convex Regularization

    Science.gov (United States)

    Parekh, Ankit

    Sparsity has become the basis of some important signal processing methods over the last ten years. Many signal processing problems (e.g., denoising, deconvolution, non-linear component analysis) can be expressed as inverse problems. Sparsity is invoked through the formulation of an inverse problem with suitably designed regularization terms. The regularization terms alone encode sparsity into the problem formulation. Often, the ℓ1 norm is used to induce sparsity, so much so that ℓ1 regularization is considered to be `modern least-squares'. The use of ℓ1 norm, as a sparsity-inducing regularizer, leads to a convex optimization problem, which has several benefits: the absence of extraneous local minima, well developed theory of globally convergent algorithms, even for large-scale problems. Convex regularization via the ℓ1 norm, however, tends to under-estimate the non-zero values of sparse signals. In order to estimate the non-zero values more accurately, non-convex regularization is often favored over convex regularization. However, non-convex regularization generally leads to non-convex optimization, which suffers from numerous issues: convergence may be guaranteed to only a stationary point, problem specific parameters may be difficult to set, and the solution is sensitive to the initialization of the algorithm. The first part of this thesis is aimed toward combining the benefits of non-convex regularization and convex optimization to estimate sparse signals more effectively. To this end, we propose to use parameterized non-convex regularizers with designated non-convexity and provide a range for the non-convex parameter so as to ensure that the objective function is strictly convex. By ensuring convexity of the objective function (sum of data-fidelity and non-convex regularizer), we can make use of a wide variety of convex optimization algorithms to obtain the unique global minimum reliably. The second part of this thesis proposes a non-linear signal

  8. Wave dynamics of regular and chaotic rays

    International Nuclear Information System (INIS)

    McDonald, S.W.

    1983-09-01

    In order to investigate general relationships between waves and rays in chaotic systems, I study the eigenfunctions and spectrum of a simple model, the two-dimensional Helmholtz equation in a stadium boundary, for which the rays are ergodic. Statistical measurements are performed so that the apparent randomness of the stadium modes can be quantitatively contrasted with the familiar regularities observed for the modes in a circular boundary (with integrable rays). The local spatial autocorrelation of the eigenfunctions is constructed in order to indirectly test theoretical predictions for the nature of the Wigner distribution corresponding to chaotic waves. A portion of the large-eigenvalue spectrum is computed and reported in an appendix; the probability distribution of successive level spacings is analyzed and compared with theoretical predictions. The two principal conclusions are: 1) waves associated with chaotic rays may exhibit randomly situated localized regions of high intensity; 2) the Wigner function for these waves may depart significantly from being uniformly distributed over the surface of constant frequency in the ray phase space

  9. Using Tikhonov Regularization for Spatial Projections from CSR Regularized Spherical Harmonic GRACE Solutions

    Science.gov (United States)

    Save, H.; Bettadpur, S. V.

    2013-12-01

    It has been demonstrated before that using Tikhonov regularization produces spherical harmonic solutions from GRACE that have very little residual stripes while capturing all the signal observed by GRACE within the noise level. This paper demonstrates a two-step process and uses Tikhonov regularization to remove the residual stripes in the CSR regularized spherical harmonic coefficients when computing the spatial projections. We discuss methods to produce mass anomaly grids that have no stripe features while satisfying the necessary condition of capturing all observed signal within the GRACE noise level.

  10. Regularized maximum correntropy machine

    KAUST Repository

    Wang, Jim Jing-Yan; Wang, Yunji; Jing, Bing-Yi; Gao, Xin

    2015-01-01

    In this paper we investigate the usage of regularized correntropy framework for learning of classifiers from noisy labels. The class label predictors learned by minimizing transitional loss functions are sensitive to the noisy and outlying labels of training samples, because the transitional loss functions are equally applied to all the samples. To solve this problem, we propose to learn the class label predictors by maximizing the correntropy between the predicted labels and the true labels of the training samples, under the regularized Maximum Correntropy Criteria (MCC) framework. Moreover, we regularize the predictor parameter to control the complexity of the predictor. The learning problem is formulated by an objective function considering the parameter regularization and MCC simultaneously. By optimizing the objective function alternately, we develop a novel predictor learning algorithm. The experiments on two challenging pattern classification tasks show that it significantly outperforms the machines with transitional loss functions.

  11. Regularized maximum correntropy machine

    KAUST Repository

    Wang, Jim Jing-Yan

    2015-02-12

    In this paper we investigate the usage of regularized correntropy framework for learning of classifiers from noisy labels. The class label predictors learned by minimizing transitional loss functions are sensitive to the noisy and outlying labels of training samples, because the transitional loss functions are equally applied to all the samples. To solve this problem, we propose to learn the class label predictors by maximizing the correntropy between the predicted labels and the true labels of the training samples, under the regularized Maximum Correntropy Criteria (MCC) framework. Moreover, we regularize the predictor parameter to control the complexity of the predictor. The learning problem is formulated by an objective function considering the parameter regularization and MCC simultaneously. By optimizing the objective function alternately, we develop a novel predictor learning algorithm. The experiments on two challenging pattern classification tasks show that it significantly outperforms the machines with transitional loss functions.

  12. Application of dimensional regularization to single chain polymer static properties: Conformational space renormalization of polymers. III

    International Nuclear Information System (INIS)

    Oono, Y.; Ohta, T.; Freed, K.F.

    1981-01-01

    A dimensional regularization approach to the renormalization group treatment of polymer excluded volume is formulated in chain conformation space where monomers are specified by their spatial positions and their positions along the chain and the polymers may be taken to be monodisperse. The method utilizes basic scale invariance considerations. First, it is recognized that long wavelength macroscopic descriptions must be well defined in the limit that the minimum atomic or molecular scale L is set to zero. Secondly, the microscopic theory is independent of the conveniently chosen macroscopic scale of length k. The freedom of choice of k is exploited along with the assumed renormalizability of the theory to provide the renormalization group equations which directly imply the universal scaling laws for macroscopic properties. The renormalizability of the model implies the existence of the general relations between the basic macroparameters, such as chain length, excluded volume, etc., and their microscopic counterparts in the microscopic model for the system. These macro--micro relations are defined through the condition that macroscopic quantities be well defined for polymer chains for any spatial dimensionality. The method is illustrated by calculating the end vector distribution function for all values of end vectors R. The evaluation of this distribution function currently requires the use of expansions in e = 4-d. In this case our distribution reduces to known limits for R→0 or infinity. Subsequent papers will present calculations of the polymer coherent scattering function, the monomer spatial distribution function, and concentration dependent properties

  13. on differential operators on w 1,2 space and fredholm operators

    African Journals Online (AJOL)

    A selfadjoint differential operator defined over a closed and bounded interval on Sobolev space which is a dense linear subspace of a Hilbert space over the same interval is considered and shown to be a Fredholm operator with index zero. KEY WORDS: Sobolev space, Hilbert space, dense subspace, Fredholm operator

  14. Stochastic dynamic modeling of regular and slow earthquakes

    Science.gov (United States)

    Aso, N.; Ando, R.; Ide, S.

    2017-12-01

    Both regular and slow earthquakes are slip phenomena on plate boundaries and are simulated by a (quasi-)dynamic modeling [Liu and Rice, 2005]. In these numerical simulations, spatial heterogeneity is usually considered not only for explaining real physical properties but also for evaluating the stability of the calculations or the sensitivity of the results on the condition. However, even though we discretize the model space with small grids, heterogeneity at smaller scales than the grid size is not considered in the models with deterministic governing equations. To evaluate the effect of heterogeneity at the smaller scales we need to consider stochastic interactions between slip and stress in a dynamic modeling. Tidal stress is known to trigger or affect both regular and slow earthquakes [Yabe et al., 2015; Ide et al., 2016], and such an external force with fluctuation can also be considered as a stochastic external force. A healing process of faults may also be stochastic, so we introduce stochastic friction law. In the present study, we propose a stochastic dynamic model to explain both regular and slow earthquakes. We solve mode III problem, which corresponds to the rupture propagation along the strike direction. We use BIEM (boundary integral equation method) scheme to simulate slip evolution, but we add stochastic perturbations in the governing equations, which is usually written in a deterministic manner. As the simplest type of perturbations, we adopt Gaussian deviations in the formulation of the slip-stress kernel, external force, and friction. By increasing the amplitude of perturbations of the slip-stress kernel, we reproduce complicated rupture process of regular earthquakes including unilateral and bilateral ruptures. By perturbing external force, we reproduce slow rupture propagation at a scale of km/day. The slow propagation generated by a combination of fast interaction at S-wave velocity is analogous to the kinetic theory of gasses: thermal

  15. Processing of sub- and supra-second intervals in the primate brain results from the calibration of neuronal oscillators via sensory, motor, and feedback processes

    Science.gov (United States)

    Gupta, Daya S.

    2014-01-01

    The processing of time intervals in the sub- to supra-second range by the brain is critical for the interaction of primates with their surroundings in activities, such as foraging and hunting. For an accurate processing of time intervals by the brain, representation of physical time within neuronal circuits is necessary. I propose that time dimension of the physical surrounding is represented in the brain by different types of neuronal oscillators, generating spikes or spike bursts at regular intervals. The proposed oscillators include the pacemaker neurons, tonic inputs, and synchronized excitation and inhibition of inter-connected neurons. Oscillators, which are built inside various circuits of brain, help to form modular clocks, processing time intervals or other temporal characteristics specific to functions of a circuit. Relative or absolute duration is represented within neuronal oscillators by “neural temporal unit,” defined as the interval between regularly occurring spikes or spike bursts. Oscillator output is processed to produce changes in activities of neurons, named frequency modulator neuron, wired within a separate module, represented by the rate of change in frequency, and frequency of activities, proposed to encode time intervals. Inbuilt oscillators are calibrated by (a) feedback processes, (b) input of time intervals resulting from rhythmic external sensory stimulation, and (c) synchronous effects of feedback processes and evoked sensory activity. A single active clock is proposed per circuit, which is calibrated by one or more mechanisms. Multiple calibration mechanisms, inbuilt oscillators, and the presence of modular connections prevent a complete loss of interval timing functions of the brain. PMID:25136321

  16. EIT image reconstruction with four dimensional regularization.

    Science.gov (United States)

    Dai, Tao; Soleimani, Manuchehr; Adler, Andy

    2008-09-01

    Electrical impedance tomography (EIT) reconstructs internal impedance images of the body from electrical measurements on body surface. The temporal resolution of EIT data can be very high, although the spatial resolution of the images is relatively low. Most EIT reconstruction algorithms calculate images from data frames independently, although data are actually highly correlated especially in high speed EIT systems. This paper proposes a 4-D EIT image reconstruction for functional EIT. The new approach is developed to directly use prior models of the temporal correlations among images and 3-D spatial correlations among image elements. A fast algorithm is also developed to reconstruct the regularized images. Image reconstruction is posed in terms of an augmented image and measurement vector which are concatenated from a specific number of previous and future frames. The reconstruction is then based on an augmented regularization matrix which reflects the a priori constraints on temporal and 3-D spatial correlations of image elements. A temporal factor reflecting the relative strength of the image correlation is objectively calculated from measurement data. Results show that image reconstruction models which account for inter-element correlations, in both space and time, show improved resolution and noise performance, in comparison to simpler image models.

  17. Evaluating Middle School Students' Spatial-scientific Performance in Earth-space Science

    Science.gov (United States)

    Wilhelm, Jennifer; Jackson, C.; Toland, M. D.; Cole, M.; Wilhelm, R. J.

    2013-06-01

    Many astronomical concepts cannot be understood without a developed understanding of four spatial-mathematics domains defined as follows: a) Geometric Spatial Visualization (GSV) - Visualizing the geometric features of a system as it appears above, below, and within the system’s plane; b) Spatial Projection (SP) - Projecting to a different location and visualizing from that global perspective; c) Cardinal Directions (CD) - Distinguishing directions (N, S, E, W) in order to document an object’s vector position in space; and d) Periodic Patterns - (PP) Recognizing occurrences at regular intervals of time and/or space. For this study, differences were examined between groups of sixth grade students’ spatial-scientific development pre/post implementation of an Earth/Space unit. Treatment teachers employed a NASA-based curriculum (Realistic Explorations in Astronomical Learning), while control teachers implemented their regular Earth/Space units. A 2-level hierarchical linear model was used to evaluate student performance on the Lunar Phases Concept Inventory (LPCI) and four spatial-mathematics domains, while controlling for two variables (gender and ethnicity) at the student level and one variable (teaching experience) at the teacher level. Overall LPCI results show pre-test scores predicted post-test scores, boys performed better than girls, and Whites performed better than non-Whites. We also compared experimental and control groups’ by spatial-mathematics domain outcomes. For GSV, it was found that boys, in general, tended to have higher GSV post-scores. For domains CD and SP, no statistically significant differences were observed. PP results show Whites performed better than non-Whites. Also for PP, a significant cross-level interaction term (gender-treatment) was observed, which means differences in control and experimental groups are dependent on students’ gender. These findings can be interpreted as: (a) the experimental girls scored higher than the

  18. Discrete- vs. Continuous-Time Modeling of Unequally Spaced Experience Sampling Method Data

    Directory of Open Access Journals (Sweden)

    Silvia de Haan-Rietdijk

    2017-10-01

    Full Text Available The Experience Sampling Method is a common approach in psychological research for collecting intensive longitudinal data with high ecological validity. One characteristic of ESM data is that it is often unequally spaced, because the measurement intervals within a day are deliberately varied, and measurement continues over several days. This poses a problem for discrete-time (DT modeling approaches, which are based on the assumption that all measurements are equally spaced. Nevertheless, DT approaches such as (vector autoregressive modeling are often used to analyze ESM data, for instance in the context of affective dynamics research. There are equivalent continuous-time (CT models, but they are more difficult to implement. In this paper we take a pragmatic approach and evaluate the practical relevance of the violated model assumption in DT AR(1 and VAR(1 models, for the N = 1 case. We use simulated data under an ESM measurement design to investigate the bias in the parameters of interest under four different model implementations, ranging from the true CT model that accounts for all the exact measurement times, to the crudest possible DT model implementation, where even the nighttime is treated as a regular interval. An analysis of empirical affect data illustrates how the differences between DT and CT modeling can play out in practice. We find that the size and the direction of the bias in DT (VAR models for unequally spaced ESM data depend quite strongly on the true parameter in addition to data characteristics. Our recommendation is to use CT modeling whenever possible, especially now that new software implementations have become available.

  19. Fibonacci-regularization method for solving Cauchy integral equations of the first kind

    Directory of Open Access Journals (Sweden)

    Mohammad Ali Fariborzi Araghi

    2017-09-01

    Full Text Available In this paper, a novel scheme is proposed to solve the first kind Cauchy integral equation over a finite interval. For this purpose, the regularization method is considered. Then, the collocation method with Fibonacci base function is applied to solve the obtained second kind singular integral equation. Also, the error estimate of the proposed scheme is discussed. Finally, some sample Cauchy integral equations stem from the theory of airfoils in fluid mechanics are presented and solved to illustrate the importance and applicability of the given algorithm. The tables in the examples show the efficiency of the method.

  20. Modes on the Move: Interval Cycles and the Emergence of Major-Minor Tonality

    Directory of Open Access Journals (Sweden)

    Matthew Woolhouse

    2011-01-01

    Full Text Available The issue of the emergence of major-minor tonality is addressed by recourse to a novel pitch grouping process, referred to as interval cycle proximity (ICP. An interval cycle is the minimum number of (additive iterations of an interval that are required for octave-related pitches to be re-stated, a property conjectured to be responsible for tonal attraction. It is hypothesised that the actuation of ICP in cognition, possibly in the latter part of the sixteenth century, led to a hierarchy of tonal attraction which favoured certain pitches over others, ostensibly the tonics of the modern major and minor system. An ICP model is described that calculates the level of tonal attraction between adjacent musical elements. The predictions of the model are shown to be consistent with music-theoretic accounts of common practice period tonality, including Piston’s Table of Usual Root Progressions. The development of tonality is illustrated with the historical quotations of commentators from the sixteenth to the eighteenth centuries, and can be characterised as follows. At the beginning of the seventeenth century multiple ‘finals’ were possible, each associated with a different interval configuration (mode. By the end of the seventeenth century, however, only two interval configurations were in regular use: those pertaining to the modern major- minor key system. The implications of this development are discussed with respect interval cycles and their hypothesised effect within music

  1. A Framework for Probabilistic Evaluation of Interval Management Tolerance in the Terminal Radar Control Area

    Science.gov (United States)

    Hercencia-Zapana, Heber; Herencia-Zapana, Heber; Hagen, George E.; Neogi, Natasha

    2012-01-01

    Projections of future traffic in the national airspace show that most of the hub airports and their attendant airspace will need to undergo significant redevelopment and redesign in order to accommodate any significant increase in traffic volume. Even though closely spaced parallel approaches increase throughput into a given airport, controller workload in oversubscribed metroplexes is further taxed by these approaches that require stringent monitoring in a saturated environment. The interval management (IM) concept in the TRACON area is designed to shift some of the operational burden from the control tower to the flight deck, placing the flight crew in charge of implementing the required speed changes to maintain a relative spacing interval. The interval management tolerance is a measure of the allowable deviation from the desired spacing interval for the IM aircraft (and its target aircraft). For this complex task, Formal Methods can help to ensure better design and system implementation. In this paper, we propose a probabilistic framework to quantify the uncertainty and performance associated with the major components of the IM tolerance. The analytical basis for this framework may be used to formalize both correctness and probabilistic system safety claims in a modular fashion at the algorithmic level in a way compatible with several Formal Methods tools.

  2. A spacing trial in Australian Toon ... an interim report

    Science.gov (United States)

    Herbert L. Wick; Robert E. Burgan

    1970-01-01

    Australian toon (Toona australis) was planted at four spacing intervals-6, 8, 10, and 12 feet-in a trial on the island of Hawaii. Measurement of 8-year-old trees showed that spacing interval did not affect diameter or height growth or survival; but basal area per acre decreased as spacing increased. A goal of 400 acceptable and desirable pole-sized...

  3. Regularities of Multifractal Measures

    Indian Academy of Sciences (India)

    First, we prove the decomposition theorem for the regularities of multifractal Hausdorff measure and packing measure in R R d . This decomposition theorem enables us to split a set into regular and irregular parts, so that we can analyze each separately, and recombine them without affecting density properties. Next, we ...

  4. The Acute Effects of Interval-Type Exercise on Glycemic Control in Type 2 Diabetes Subjects: Importance of Interval Length. A Controlled, Counterbalanced, Crossover Study.

    Directory of Open Access Journals (Sweden)

    Ida Jakobsen

    Full Text Available Interval-type exercise is effective for improving glycemic control, but the optimal approach is unknown. The purpose of this study was to determine the importance of the interval length on changes in postprandial glycemic control following a single exercise bout. Twelve subjects with type 2 diabetes completed a cross-over study with three 1-hour interventions performed in a non-randomized but counter-balanced order: 1 Interval walking consisting of repeated cycles of 3 min slow (aiming for 54% of Peak oxygen consumption rate [VO2peak] and 3 min fast (aiming for 89% of VO2peak walking (IW3; 2 Interval walking consisting of repeated cycles of 1 min slow and 1 min fast walking (IW1 and 3 No walking (CON. The exercise interventions were matched with regards to walking speed, and VO2 and heart rate was assessed throughout all interventions. A 4-hour liquid mixed meal tolerance test commenced 30 min after each intervention, with blood samples taken regularly. IW3 and IW1 resulted in comparable mean VO2 and heart rates. Overall mean postprandial blood glucose levels were lower after IW3 compared to CON (10.3±3.0 vs. 11.1±3.3 mmol/L; P 0.05 for both. Conversely blood glucose levels at specific time points during the MMTT differed significantly following both IW3 and IW1 as compared to CON. Our findings support the previously found blood glucose lowering effect of IW3 and suggest that reducing the interval length, while keeping the walking speed and time spend on fast and slow walking constant, does not result in additional improvements.ClinicalTrials.gov NCT02257190.

  5. Adaptive Regularization of Neural Classifiers

    DEFF Research Database (Denmark)

    Andersen, Lars Nonboe; Larsen, Jan; Hansen, Lars Kai

    1997-01-01

    We present a regularization scheme which iteratively adapts the regularization parameters by minimizing the validation error. It is suggested to use the adaptive regularization scheme in conjunction with optimal brain damage pruning to optimize the architecture and to avoid overfitting. Furthermo......, we propose an improved neural classification architecture eliminating an inherent redundancy in the widely used SoftMax classification network. Numerical results demonstrate the viability of the method...

  6. Space proliferation versus space-type dissemination: from semantic issues to political issues

    International Nuclear Information System (INIS)

    Gaillard-Sborowsky, Florence

    2015-01-01

    The space and ballistic capabilities relationships are regularly revisited in forums on international security, in particular about Iran and North Korea cases. The term 'space proliferation' is commonly used by analogy with nuclear proliferation. However, is this analogy relevant? Beyond the semantic aspects, this shift raises political issues that this paper will consider. The study of the assumptions regarding the analysis of nuclear and missile proliferation and their space counterparts will highlight some approximations and presuppositions, such as the amalgam between sounding rockets, launchers and missiles technologies, in order to suggest new thinking of these sensitive issues. (author)

  7. Condition Number Regularized Covariance Estimation.

    Science.gov (United States)

    Won, Joong-Ho; Lim, Johan; Kim, Seung-Jean; Rajaratnam, Bala

    2013-06-01

    Estimation of high-dimensional covariance matrices is known to be a difficult problem, has many applications, and is of current interest to the larger statistics community. In many applications including so-called the "large p small n " setting, the estimate of the covariance matrix is required to be not only invertible, but also well-conditioned. Although many regularization schemes attempt to do this, none of them address the ill-conditioning problem directly. In this paper, we propose a maximum likelihood approach, with the direct goal of obtaining a well-conditioned estimator. No sparsity assumption on either the covariance matrix or its inverse are are imposed, thus making our procedure more widely applicable. We demonstrate that the proposed regularization scheme is computationally efficient, yields a type of Steinian shrinkage estimator, and has a natural Bayesian interpretation. We investigate the theoretical properties of the regularized covariance estimator comprehensively, including its regularization path, and proceed to develop an approach that adaptively determines the level of regularization that is required. Finally, we demonstrate the performance of the regularized estimator in decision-theoretic comparisons and in the financial portfolio optimization setting. The proposed approach has desirable properties, and can serve as a competitive procedure, especially when the sample size is small and when a well-conditioned estimator is required.

  8. Multi-label learning with fuzzy hypergraph regularization for protein subcellular location prediction.

    Science.gov (United States)

    Chen, Jing; Tang, Yuan Yan; Chen, C L Philip; Fang, Bin; Lin, Yuewei; Shang, Zhaowei

    2014-12-01

    Protein subcellular location prediction aims to predict the location where a protein resides within a cell using computational methods. Considering the main limitations of the existing methods, we propose a hierarchical multi-label learning model FHML for both single-location proteins and multi-location proteins. The latent concepts are extracted through feature space decomposition and label space decomposition under the nonnegative data factorization framework. The extracted latent concepts are used as the codebook to indirectly connect the protein features to their annotations. We construct dual fuzzy hypergraphs to capture the intrinsic high-order relations embedded in not only feature space, but also label space. Finally, the subcellular location annotation information is propagated from the labeled proteins to the unlabeled proteins by performing dual fuzzy hypergraph Laplacian regularization. The experimental results on the six protein benchmark datasets demonstrate the superiority of our proposed method by comparing it with the state-of-the-art methods, and illustrate the benefit of exploiting both feature correlations and label correlations.

  9. C1,1 regularity for degenerate elliptic obstacle problems

    Science.gov (United States)

    Daskalopoulos, Panagiota; Feehan, Paul M. N.

    2016-03-01

    The Heston stochastic volatility process is a degenerate diffusion process where the degeneracy in the diffusion coefficient is proportional to the square root of the distance to the boundary of the half-plane. The generator of this process with killing, called the elliptic Heston operator, is a second-order, degenerate-elliptic partial differential operator, where the degeneracy in the operator symbol is proportional to the distance to the boundary of the half-plane. In mathematical finance, solutions to the obstacle problem for the elliptic Heston operator correspond to value functions for perpetual American-style options on the underlying asset. With the aid of weighted Sobolev spaces and weighted Hölder spaces, we establish the optimal C 1 , 1 regularity (up to the boundary of the half-plane) for solutions to obstacle problems for the elliptic Heston operator when the obstacle functions are sufficiently smooth.

  10. Association of regular walking and body mass index on metabolic syndrome among an elderly Korean population.

    Science.gov (United States)

    Kim, Soonyoung; Kim, Dong-Il

    2018-06-01

    Aging is associated with increased body fat and lower lean body mass, which leads to increased prevalence of obesity and metabolic syndrome. This study aimed to investigate the association of regular participation in walking and body mass index (BMI) with metabolic syndrome and its 5 criteria in elderly Koreans. A total of 3554 (male = 1581, female = 1973) elderly subjects (age ≥ 65 years), who participated in the Fifth Korea National Health and Nutrition Examination Survey (KNHANES V) were analyzed in this cross-sectional study. Participation in walking activity, BMI, metabolic syndrome and its 5 criteria; waist circumference (WC), systolic blood pressure (SBP), diastolic blood pressure (DBP), fasting glucose (FG) levels, triglyceride (TG) levels, and high-density lipoprotein cholesterol (HDLC) levels, were measured. Subjects were categorized into four groups based on the duration and regularity of their walks and BMI. In the regular walking (≥30 min of continuous walking a day, on ≥5 days a week) and normal weight (BMI metabolic syndrome was 4.36 times higher (Odds ratio [OR]: 4.36, 95% confidence interval [CI]: 3.37-5.63) in the non-regular walking and overweight group than that of the regular walking and normal weight group after controlling for the influence of age, sex, and smoking status. Moreover, The BMI (β = 0.328, R 2  = 0.152) were more contributing factors than Regular walking (β = -0.011) for metabolic syndrome. In conclusions, regular participation in walking activity and implementing weight control may reduce the incidence rate of metabolic syndrome in elderly Koreans, with weight management serving as the greater influences of the two. Copyright © 2018. Published by Elsevier Inc.

  11. Some regularities in invertebrate succession in different microhabitats on pine stumps

    OpenAIRE

    Franch, Joan

    1989-01-01

    Sixty eight pine stumps felled on known dates from one to sixteen years before the moment of sampling have been studied in the San Juan de la Peña woodland (province of Huesca). Four microhabitats were distinguished: bark, subcortical space, sapwood and heartwood. The object of the study is to compare the invertebrate macrofauna succession of the different microhabitats in order to find regularities among them. The biocenosis has not been completely studied: ipidae, diptera and annelidae are ...

  12. Initiating and maintaining recreational walking: a longitudinal study on the influence of neighborhood green space.

    Science.gov (United States)

    Sugiyama, Takemi; Giles-Corti, Billie; Summers, Jacqui; du Toit, Lorinne; Leslie, Eva; Owen, Neville

    2013-09-01

    This study examined prospective relationships of green space attributes with adults initiating or maintaining recreational walking. Postal surveys were completed by 1036 adults living in Adelaide, Australia, at baseline (two time points in 2003-04) and follow-up (2007-08). Initiating or maintaining recreational walking was determined using self-reported walking frequency. Green space attributes examined were perceived presence, quality, proximity, and the objectively measured area (total and largest) and number of green spaces within a 1.6 km buffer drawn from the center of each study neighborhood. Multilevel regression analyses examined the odds of initiating or maintaining walking separately for each green space attribute. At baseline, participants were categorized into non-regular (n = 395), regular (n = 286), and irregular walkers (n = 313). Among non-regular walkers, 30% had initiated walking, while 70% of regular walkers had maintained walking at follow-up. No green space attributes were associated with initiating walking. However, positive perceptions of the presence of and proximity to green spaces and the total and largest areas of green space were significantly associated with a higher likelihood of walking maintenance over four years. Neighborhood green spaces may not assist adults to initiate walking, but their presence and proximity may facilitate them to maintain recreational walking over time. Copyright © 2013 Elsevier Inc. All rights reserved.

  13. A HEMATOBIOCHEMICAL EVALUATION TO COMPARE THE EFFECTS OF HIGH INTENSITY INTERVAL TRAINING AND AEROBIC EXERCISE TO CONTROL DIABETES MALLITIS AND ITS COMPLICATIONS

    Directory of Open Access Journals (Sweden)

    Muneeb Iqbal

    2016-06-01

    Full Text Available Background: Diabetes has become a very common disease all over the world since last few decades and is now perceived as a global health disorder. Diabetes mellitus is identified on the basis of constant high concentration of blood glucose level and it mainly occurs due to deficiency of the pancreatic hormone insulin. High-intensity interval training (HIIT is an improved form of interval trainings, and exercise strategies which alternate the periods of small intense anaerobic exercise by less-intense regaining periods. The study aimed to compare the hematological parameters associated with diabetes and muscle activity between healthy humans and diabetic type-1 patients when subjected to HIIT and regular aerobic exercises. Methods: A convenience sample of total 60 participants was taken it comprised of thirty healthy individuals taken from the department of Physical Therapy, University of Sargodha, Lahore campus and thirty diabetic type-1 individuals of age 15-30 years taken from Akhuwat health services clinic Township, Lahore. Participants were divided into four groups of fifteen individuals each. Group one was the diabetic HIIT (DH group with diabetic type-1 patients subjected to HIIT. Group two was the diabetic aerobic (DA group with diabetic type-1 patients subjected to regular aerobic exercises. Group three was control High intensity interval training (HH that consisted of fifteen healthy individuals to be subjected to High intensity interval training exercises (HIIT. Group four (HA was the control aerobic group with fifteen healthy individuals of average lifestyles subjected to regular aerobic exercises. Results: Aerobic exercise was found to be more effective in reducing glucose level, lowering exogenous insulin and glycated hemoglobin, however HIIT proved to be more effective in lowering blood cholesterol level and decrease LDL level and increase HDL level. Conclusion: It was concluded that aerobic exercise program in comparison to high

  14. Application of Littlewood-Paley decomposition to the regularity of Boltzmann type kinetic equations

    International Nuclear Information System (INIS)

    EL Safadi, M.

    2007-03-01

    We study the regularity of kinetic equations of Boltzmann type.We use essentially Littlewood-Paley method from harmonic analysis, consisting mainly in working with dyadics annulus. We shall mainly concern with the homogeneous case, where the solution f(t,x,v) depends only on the time t and on the velocities v, while working with realistic and singular cross-sections (non cutoff). In the first part, we study the particular case of Maxwellian molecules. Under this hypothesis, the structure of the Boltzmann operator and his Fourier transform write in a simple form. We show a global C ∞ regularity. Then, we deal with the case of general cross-sections with 'hard potential'. We are interested in the Landau equation which is limit equation to the Boltzmann equation, taking in account grazing collisions. We prove that any weak solution belongs to Schwartz space S. We demonstrate also a similar regularity for the case of Boltzmann equation. Let us note that our method applies directly for all dimensions, and proofs are often simpler compared to other previous ones. Finally, we finish with Boltzmann-Dirac equation. In particular, we adapt the result of regularity obtained in Alexandre, Desvillettes, Wennberg and Villani work, using the dissipation rate connected with Boltzmann-Dirac equation. (author)

  15. High Intensity Interval Training for Maximizing Health Outcomes.

    Science.gov (United States)

    Karlsen, Trine; Aamot, Inger-Lise; Haykowsky, Mark; Rognmo, Øivind

    Regular physical activity and exercise training are important actions to improve cardiorespiratory fitness and maintain health throughout life. There is solid evidence that exercise is an effective preventative strategy against at least 25 medical conditions, including cardiovascular disease, stroke, hypertension, colon and breast cancer, and type 2 diabetes. Traditionally, endurance exercise training (ET) to improve health related outcomes has consisted of low- to moderate ET intensity. However, a growing body of evidence suggests that higher exercise intensities may be superior to moderate intensity for maximizing health outcomes. The primary objective of this review is to discuss how aerobic high-intensity interval training (HIIT) as compared to moderate continuous training may maximize outcomes, and to provide practical advices for successful clinical and home-based HIIT. Copyright © 2017. Published by Elsevier Inc.

  16. Wind Information Uplink to Aircraft Performing Interval Management Operations

    Science.gov (United States)

    Ahmad, Nashat N.; Barmore, Bryan E.; Swieringa, Kurt A.

    2016-01-01

    Interval Management (IM) is an ADS-B-enabled suite of applications that use ground and flight deck capabilities and procedures designed to support the relative spacing of aircraft (Barmore et al., 2004, Murdoch et al. 2009, Barmore 2009, Swieringa et al. 2011; Weitz et al. 2012). Relative spacing refers to managing the position of one aircraft to a time or distance relative to another aircraft, as opposed to a static reference point such as a point over the ground or clock time. This results in improved inter-aircraft spacing precision and is expected to allow aircraft to be spaced closer to the applicable separation standard than current operations. Consequently, if the reduced spacing is used in scheduling, IM can reduce the time interval between the first and last aircraft in an overall arrival flow, resulting in increased throughput. Because IM relies on speed changes to achieve precise spacing, it can reduce costly, low-altitude, vectoring, which increases both efficiency and throughput in capacity-constrained airspace without negatively impacting controller workload and task complexity. This is expected to increase overall system efficiency. The Flight Deck Interval Management (FIM) equipment provides speeds to the flight crew that will deliver them to the achieve-by point at the controller-specified time, i.e., assigned spacing goal, after the target aircraft crosses the achieve-by point (Figure 1.1). Since the IM and target aircraft may not be on the same arrival procedure, the FIM equipment predicts the estimated times of arrival (ETA) for both the IM and target aircraft to the achieve-by point. This involves generating an approximate four-dimensional trajectory for each aircraft. The accuracy of the wind data used to generate those trajectories is critical to the success of the IM operation. There are two main forms of uncertainty in the wind information used by the FIM equipment. The first is the accuracy of the forecast modeling done by the weather

  17. Are Long-Term Chloroquine or Hydroxychloroquine Users Being Checked Regularly for Toxic Maculopathy?

    Science.gov (United States)

    Nika, Melisa; Blachley, Taylor S.; Edwards, Paul; Lee, Paul P.; Stein, Joshua D.

    2014-01-01

    Importance According to evidence-based, expert recommendations, long-term users of chloroquine (CQ) or hydroxychloroquine (HCQ) should undergo regular visits to eye-care providers and diagnostic testing to check for maculopathy. Objective To determine whether patients with rheumatoid arthritis (RA) or systemic lupus erythematosus (SLE) taking CQ or HCQ are regularly visiting eye-care providers and being screened for maculopathy. Setting, Design and Participants Patients with RA or SLE who were continuously enrolled in a particular managed-care network for ≥5 years during 2001-2011 were studied. Patients' amount of CQ/HCQ use in the 5 years since initial RA/SLE diagnosis was calculated, along with their number of eye-care visits and diagnostic tests for maculopathy. Those at high risk for maculopathy were identified. Visits to eye providers and diagnostic testing for maculopathy were assessed for each enrollee over the study period. Logistic regression was performed to assess potential factors associated with regular eye-care-provider visits (≥3 in 5 years) among CQ/HCQ users, including those at greatest risk for maculopathy. Main Outcome Measures Among CQ/HCQ users and those at high risk for toxic maculopathy, the proportions with regular eye-care visits and diagnostic testing, and the likelihood of regular eye-care visits (odds ratios [ORs] with 95% confidence intervals [CI]). Results Among 18,051 beneficiaries with RA or SLE, 6,339 (35.1%) had ≥1 record of HCQ/CQ use and 1,409 (7.8%) used HCQ/CQ for ≥4 years. Among those at high risk for maculopathy, 27.9% lacked regular eye-provider visits, 6.1% had no visits to eye providers, and 34.5% had no diagnostic testing for maculopathy during the 5-year period. Among high-risk patients, each additional month of HCQ/CQ use was associated with a 2.0%-increased likelihood of regular eye care (adjusted OR=1.02, CI=1.01-1.03). High-risk patients whose SLE/RA were managed by rheumatologists had a 77%-increased

  18. Geometry on the space of geometries

    International Nuclear Information System (INIS)

    Christodoulakis, T.; Zanelli, J.

    1988-06-01

    We discuss the geometric structure of the configuration space of pure gravity. This is an infinite dimensional manifold, M, where each point represents one spatial geometry g ij (x). The metric on M is dictated by geometrodynamics, and from it, the Christoffel symbols and Riemann tensor can be found. A ''free geometry'' tracing a geodesic on the manifold describes the time evolution of space in the strong gravity limit. In a regularization previously introduced by the authors, it is found that M does not have the same dimensionality, D, everywhere, and that D is not a scalar, although it is covariantly constant. In this regularization, it is seen that the path integral measure can be absorbed in a renormalization of the cosmological constant. (author). 19 refs

  19. Condition Number Regularized Covariance Estimation*

    Science.gov (United States)

    Won, Joong-Ho; Lim, Johan; Kim, Seung-Jean; Rajaratnam, Bala

    2012-01-01

    Estimation of high-dimensional covariance matrices is known to be a difficult problem, has many applications, and is of current interest to the larger statistics community. In many applications including so-called the “large p small n” setting, the estimate of the covariance matrix is required to be not only invertible, but also well-conditioned. Although many regularization schemes attempt to do this, none of them address the ill-conditioning problem directly. In this paper, we propose a maximum likelihood approach, with the direct goal of obtaining a well-conditioned estimator. No sparsity assumption on either the covariance matrix or its inverse are are imposed, thus making our procedure more widely applicable. We demonstrate that the proposed regularization scheme is computationally efficient, yields a type of Steinian shrinkage estimator, and has a natural Bayesian interpretation. We investigate the theoretical properties of the regularized covariance estimator comprehensively, including its regularization path, and proceed to develop an approach that adaptively determines the level of regularization that is required. Finally, we demonstrate the performance of the regularized estimator in decision-theoretic comparisons and in the financial portfolio optimization setting. The proposed approach has desirable properties, and can serve as a competitive procedure, especially when the sample size is small and when a well-conditioned estimator is required. PMID:23730197

  20. Radioactive means for measuring distance intervals between anomalies in an earth formation

    International Nuclear Information System (INIS)

    Sandier, G.; Nels, J.P.

    1975-01-01

    In earth formation measurements an investigating tool having a first and a second detector at a relatively small spacing from each other and a third detector at a relatively great spacing from the first and second detectors is moved through a borehole having anomalies such as radioactive bullets or casing joints which are relatively uniformly spaced from each other by a distance which is of the order of said great spacing between the third detector and the first and the second detectors. The first and second detectors generate detection signal peaks for the same anomaly within a short time interval, and the third detector generates a detection signal peak for an adjacent anomaly at about the same time. By means of a defined apparatus, electrical signals representing the times of occurrence of the detection signal peaks from the first and second detectors for the same anomaly and the known small spacing between these detectors are used to obtain an electrical signal for the speed of the investigating tool at that time, and at least some of these electrical signals are combined with electrical signals representing the time of occurrence of the detection signal peak from the third detector for an adjacent anomaly and at least one of the known distances between the detectors to thereby obtain an accurate measure of the distance interval between the pair of adjacent anomalies. (U.S.)

  1. Diffusion of charged particles in strong large-scale random and regular magnetic fields

    International Nuclear Information System (INIS)

    Mel'nikov, Yu.P.

    2000-01-01

    The nonlinear collision integral for the Green's function averaged over a random magnetic field is transformed using an iteration procedure taking account of the strong random scattering of particles on the correlation length of the random magnetic field. Under this transformation the regular magnetic field is assumed to be uniform at distances of the order of the correlation length. The single-particle Green's functions of the scattered particles in the presence of a regular magnetic field are investigated. The transport coefficients are calculated taking account of the broadening of the cyclotron and Cherenkov resonances as a result of strong random scattering. The mean-free path lengths parallel and perpendicular to the regular magnetic field are found for a power-law spectrum of the random field. The analytical results obtained are compared with the experimental data on the transport ranges of solar and galactic cosmic rays in the interplanetary magnetic field. As a result, the conditions for the propagation of cosmic rays in the interplanetary space and a more accurate idea of the structure of the interplanetary magnetic field are determined

  2. Convex Interval Games

    NARCIS (Netherlands)

    Alparslan-Gok, S.Z.; Brânzei, R.; Tijs, S.H.

    2008-01-01

    In this paper, convex interval games are introduced and some characterizations are given. Some economic situations leading to convex interval games are discussed. The Weber set and the Shapley value are defined for a suitable class of interval games and their relations with the interval core for

  3. Regular-, irregular-, and pseudo-character processing in Chinese: The regularity effect in normal adult readers

    Directory of Open Access Journals (Sweden)

    Dustin Kai Yan Lau

    2014-03-01

    Full Text Available Background Unlike alphabetic languages, Chinese uses a logographic script. However, the pronunciation of many character’s phonetic radical has the same pronunciation as the character as a whole. These are considered regular characters and can be read through a lexical non-semantic route (Weekes & Chen, 1999. Pseudocharacters are another way to study this non-semantic route. A pseudocharacter is the combination of existing semantic and phonetic radicals in their legal positions resulting in a non-existing character (Ho, Chan, Chung, Lee, & Tsang, 2007. Pseudocharacters can be pronounced by direct derivation from the sound of its phonetic radical. Conversely, if the pronunciation of a character does not follow that of the phonetic radical, it is considered as irregular and can only be correctly read through the lexical-semantic route. The aim of the current investigation was to examine reading aloud in normal adults. We hypothesized that the regularity effect, previously described for alphabetical scripts and acquired dyslexic patients of Chinese (Weekes & Chen, 1999; Wu, Liu, Sun, Chromik, & Zhang, 2014, would also be present in normal adult Chinese readers. Method Participants. Thirty (50% female native Hong Kong Cantonese speakers with a mean age of 19.6 years and a mean education of 12.9 years. Stimuli. Sixty regular-, 60 irregular-, and 60 pseudo-characters (with at least 75% of name agreement in Chinese were matched by initial phoneme, number of strokes and family size. Additionally, regular- and irregular-characters were matched by frequency (low and consistency. Procedure. Each participant was asked to read aloud the stimuli presented on a laptop using the DMDX software. The order of stimuli presentation was randomized. Data analysis. ANOVAs were carried out by participants and items with RTs and errors as dependent variables and type of stimuli (regular-, irregular- and pseudo-character as repeated measures (F1 or between subject

  4. Birth spacing of pregnant women in Nepal: A community-based study

    Directory of Open Access Journals (Sweden)

    Rajendra Karkee

    2016-09-01

    Full Text Available BackgroundOptimal birth spacing has health advantages for both mother and child. In developing countries, shorter birth intervals are common and associated with social, cultural and economic factors, as well as a lack of family planning. This study investigated the first birth interval after marriage and preceding interbirth interval in Nepal.MethodsA community-based prospective cohort study was conducted in the Kaski district of Nepal. Information on birth spacing, demographic and obstetric characteristics was obtained from 701 pregnant women using a structured questionnaire. Logistic regression analyses were performed to ascertain factors associated with short birth spacing.ResultsAbout 39% of primiparous women gave their first child birth within one year of marriage and 23% of multiparous women had short preceding interbirth intervals (<24 months. The average birth spacing among the multiparous group was 44.9 (SD 21.8 months. Overall, short birth spacing appeared to be inversely associated with advancing maternal age.For the multiparous group, Janajati and lower caste women, and those whose newborn was female, were more likely to have short birth spacing.ConclusionsThe preceding interbirth interval was relatively long in the Kaski district of Nepal and tended to be associated with maternal age, caste, and sex of newborn infant. Optimal birth spacing programs should target Janajati and lower caste women, along with promotion of gender equality in society.

  5. The right time to learn: mechanisms and optimization of spaced learning

    Science.gov (United States)

    Smolen, Paul; Zhang, Yili; Byrne, John H.

    2016-01-01

    For many types of learning, spaced training, which involves repeated long inter-trial intervals, leads to more robust memory formation than does massed training, which involves short or no intervals. Several cognitive theories have been proposed to explain this superiority, but only recently have data begun to delineate the underlying cellular and molecular mechanisms of spaced training, and we review these theories and data here. Computational models of the implicated signalling cascades have predicted that spaced training with irregular inter-trial intervals can enhance learning. This strategy of using models to predict optimal spaced training protocols, combined with pharmacotherapy, suggests novel ways to rescue impaired synaptic plasticity and learning. PMID:26806627

  6. Regularity effect in prospective memory during aging

    Directory of Open Access Journals (Sweden)

    Geoffrey Blondelle

    2016-10-01

    Full Text Available Background: Regularity effect can affect performance in prospective memory (PM, but little is known on the cognitive processes linked to this effect. Moreover, its impacts with regard to aging remain unknown. To our knowledge, this study is the first to examine regularity effect in PM in a lifespan perspective, with a sample of young, intermediate, and older adults. Objective and design: Our study examined the regularity effect in PM in three groups of participants: 28 young adults (18–30, 16 intermediate adults (40–55, and 25 older adults (65–80. The task, adapted from the Virtual Week, was designed to manipulate the regularity of the various activities of daily life that were to be recalled (regular repeated activities vs. irregular non-repeated activities. We examine the role of several cognitive functions including certain dimensions of executive functions (planning, inhibition, shifting, and binding, short-term memory, and retrospective episodic memory to identify those involved in PM, according to regularity and age. Results: A mixed-design ANOVA showed a main effect of task regularity and an interaction between age and regularity: an age-related difference in PM performances was found for irregular activities (older < young, but not for regular activities. All participants recalled more regular activities than irregular ones with no age effect. It appeared that recalling of regular activities only involved planning for both intermediate and older adults, while recalling of irregular ones were linked to planning, inhibition, short-term memory, binding, and retrospective episodic memory. Conclusion: Taken together, our data suggest that planning capacities seem to play a major role in remembering to perform intended actions with advancing age. Furthermore, the age-PM-paradox may be attenuated when the experimental design is adapted by implementing a familiar context through the use of activities of daily living. The clinical

  7. J-regular rings with injectivities

    OpenAIRE

    Shen, Liang

    2010-01-01

    A ring $R$ is called a J-regular ring if R/J(R) is von Neumann regular, where J(R) is the Jacobson radical of R. It is proved that if R is J-regular, then (i) R is right n-injective if and only if every homomorphism from an $n$-generated small right ideal of $R$ to $R_{R}$ can be extended to one from $R_{R}$ to $R_{R}$; (ii) R is right FP-injective if and only if R is right (J, R)-FP-injective. Some known results are improved.

  8. Chiral Thirring–Wess model with Faddeevian regularization

    International Nuclear Information System (INIS)

    Rahaman, Anisur

    2015-01-01

    Replacing vector type of interaction of the Thirring–Wess model by the chiral type a new model is presented which is termed here as chiral Thirring–Wess model. Ambiguity parameters of regularization are so chosen that the model falls into the Faddeevian class. The resulting Faddeevian class of model in general does not possess Lorentz invariance. However we can exploit the arbitrariness admissible in the ambiguity parameters to relate the quantum mechanically generated ambiguity parameters with the classical parameter involved in the masslike term of the gauge field which helps to maintain physical Lorentz invariance instead of the absence of manifestly Lorentz covariance of the model. The phase space structure and the theoretical spectrum of this class of model have been determined through Dirac’s method of quantization of constraint system

  9. Discriminative Elastic-Net Regularized Linear Regression.

    Science.gov (United States)

    Zhang, Zheng; Lai, Zhihui; Xu, Yong; Shao, Ling; Wu, Jian; Xie, Guo-Sen

    2017-03-01

    In this paper, we aim at learning compact and discriminative linear regression models. Linear regression has been widely used in different problems. However, most of the existing linear regression methods exploit the conventional zero-one matrix as the regression targets, which greatly narrows the flexibility of the regression model. Another major limitation of these methods is that the learned projection matrix fails to precisely project the image features to the target space due to their weak discriminative capability. To this end, we present an elastic-net regularized linear regression (ENLR) framework, and develop two robust linear regression models which possess the following special characteristics. First, our methods exploit two particular strategies to enlarge the margins of different classes by relaxing the strict binary targets into a more feasible variable matrix. Second, a robust elastic-net regularization of singular values is introduced to enhance the compactness and effectiveness of the learned projection matrix. Third, the resulting optimization problem of ENLR has a closed-form solution in each iteration, which can be solved efficiently. Finally, rather than directly exploiting the projection matrix for recognition, our methods employ the transformed features as the new discriminate representations to make final image classification. Compared with the traditional linear regression model and some of its variants, our method is much more accurate in image classification. Extensive experiments conducted on publicly available data sets well demonstrate that the proposed framework can outperform the state-of-the-art methods. The MATLAB codes of our methods can be available at http://www.yongxu.org/lunwen.html.

  10. Iterative Regularization with Minimum-Residual Methods

    DEFF Research Database (Denmark)

    Jensen, Toke Koldborg; Hansen, Per Christian

    2007-01-01

    subspaces. We provide a combination of theory and numerical examples, and our analysis confirms the experience that MINRES and MR-II can work as general regularization methods. We also demonstrate theoretically and experimentally that the same is not true, in general, for GMRES and RRGMRES their success......We study the regularization properties of iterative minimum-residual methods applied to discrete ill-posed problems. In these methods, the projection onto the underlying Krylov subspace acts as a regularizer, and the emphasis of this work is on the role played by the basis vectors of these Krylov...... as regularization methods is highly problem dependent....

  11. Iterative regularization with minimum-residual methods

    DEFF Research Database (Denmark)

    Jensen, Toke Koldborg; Hansen, Per Christian

    2006-01-01

    subspaces. We provide a combination of theory and numerical examples, and our analysis confirms the experience that MINRES and MR-II can work as general regularization methods. We also demonstrate theoretically and experimentally that the same is not true, in general, for GMRES and RRGMRES - their success......We study the regularization properties of iterative minimum-residual methods applied to discrete ill-posed problems. In these methods, the projection onto the underlying Krylov subspace acts as a regularizer, and the emphasis of this work is on the role played by the basis vectors of these Krylov...... as regularization methods is highly problem dependent....

  12. Prevalence of and factors associated with regular khat chewing among university students in Ethiopia

    Directory of Open Access Journals (Sweden)

    Astatkie A

    2015-02-01

    Full Text Available Ayalew Astatkie,1 Meaza Demissie,2 Yemane Berhane,2 Alemayehu Worku2,3 1School of Public and Environmental Health, College of Medicine and Health Sciences, Hawassa University, Hawassa, Ethiopia; 2Addis Continental Institute of Public Health, Addis Ababa, Ethiopia; 3School of Public Health, College of Health Sciences, Addis Ababa University, Addis Ababa, Ethiopia Purpose: Khat (Catha edulis is commonly chewed for its psychostimulant and euphorigenic effects in Africa and the Arabian Peninsula. Students use it to help them study for long hours especially during the period of examination. However, how regularly khat is chewed among university students and its associated factors are not well documented. In this article we report on the prevalence of and factors associated with regular khat chewing among university students in Ethiopia. Methods: We did a cross-sectional study from May 20, 2014 to June 23, 2014 on a sample of 1,255 regular students recruited from all campuses of Hawassa University, southern Ethiopia. The data were collected using self-administered questionnaires. We analyzed the data to identify factors associated with current regular khat chewing using complex sample adjusted logistic regression analysis. Results: The prevalence of current regular khat chewing was 10.5% (95% confidence interval [CI]: 6.1%–14.9%. After controlling for sex, religion, year of study, having a father who chews khat, cigarette smoking and alcohol drinking in the adjusted logistic regression model, living off-campus in rented houses as compared to living in the university dormitory (adjusted odds ratio [95% CI] =8.09 [1.56–42.01], and having friends who chew khat (adjusted odds ratio [95% CI] =4.62 [1.98–10.74] were found to significantly increase the odds of current regular khat use. Conclusion: Students living outside the university campus in rented houses compared to those living in dormitory and those with khat chewing peers are more likely to use

  13. Multiple graph regularized protein domain ranking.

    Science.gov (United States)

    Wang, Jim Jing-Yan; Bensmail, Halima; Gao, Xin

    2012-11-19

    Protein domain ranking is a fundamental task in structural biology. Most protein domain ranking methods rely on the pairwise comparison of protein domains while neglecting the global manifold structure of the protein domain database. Recently, graph regularized ranking that exploits the global structure of the graph defined by the pairwise similarities has been proposed. However, the existing graph regularized ranking methods are very sensitive to the choice of the graph model and parameters, and this remains a difficult problem for most of the protein domain ranking methods. To tackle this problem, we have developed the Multiple Graph regularized Ranking algorithm, MultiG-Rank. Instead of using a single graph to regularize the ranking scores, MultiG-Rank approximates the intrinsic manifold of protein domain distribution by combining multiple initial graphs for the regularization. Graph weights are learned with ranking scores jointly and automatically, by alternately minimizing an objective function in an iterative algorithm. Experimental results on a subset of the ASTRAL SCOP protein domain database demonstrate that MultiG-Rank achieves a better ranking performance than single graph regularized ranking methods and pairwise similarity based ranking methods. The problem of graph model and parameter selection in graph regularized protein domain ranking can be solved effectively by combining multiple graphs. This aspect of generalization introduces a new frontier in applying multiple graphs to solving protein domain ranking applications.

  14. Association between regular physical exercise and depressive symptoms mediated through social support and resilience in Japanese company workers: a cross-sectional study.

    Science.gov (United States)

    Yoshikawa, Eisho; Nishi, Daisuke; Matsuoka, Yutaka J

    2016-07-12

    Regular physical exercise has been reported to reduce depressive symptoms. Several lines of evidence suggest that physical exercise may prevent depression by promoting social support or resilience, which is the ability to adapt to challenging life conditions. The aim of this study was to compare depressive symptoms, social support, and resilience between Japanese company workers who engaged in regular physical exercise and workers who did not exercise regularly. We also investigated whether regular physical exercise has an indirect association with depressive symptoms through social support and resilience. Participants were 715 Japanese employees at six worksites. Depressive symptoms were assessed with the Center for Epidemiologic Studies Depression (CES-D) scale, social support with the short version of the Social Support Questionnaire (SSQ), and resilience with the 14-item Resilience Scale (RS-14). A self-report questionnaire, which was extracted from the Japanese version of the Health-Promoting Lifestyle Profile, was used to assess whether participants engage in regular physical exercise, defined as more than 20 min, three or more times per week. The group differences in CES-D, SSQ, and RS-14 scores were investigated by using analysis of covariance (ANCOVA). Mediation analysis was conducted by using Preacher and Hayes' bootstrap script to assess whether regular physical exercise is associated with depressive symptoms indirectly through resilience and social support. The SSQ Number score (F = 4.82, p = 0.03), SSQ Satisfaction score (F = 6.68, p = 0.01), and RS-14 score (F = 6.01, p = 0.01) were significantly higher in the group with regular physical exercise (n = 83) than in the group without regular physical exercise (n = 632) after adjusting for age, education, marital status, and job status. The difference in CES-D score was not significant (F = 2.90, p = 0.09). Bootstrapping revealed significant negative indirect

  15. [Incidence of long (short) PR interval in electrocardiogram among healthy people in Changsha and its clinical significance].

    Science.gov (United States)

    Liu, Liping; Lin, Ping; Xu, Yi; Wu, Lijia; Zou, Runmei; Xie, Zhenwu; Wang, Cheng

    2016-04-01

    To analyze the incidence of long (short) PR interval in electrocardiogram among healthy people in Changsha and the clinical significance.
 Twelve-lead body surface electrocardiogram was taken to measure the heart rates and PR intervals from 4 025 healthy individuals (age range from 6 min after birth to 83 years old) who performed physical examination from Jan, 1993 to Dec, 2012 in the Second Xiangya Hospital, Central South University. Statistics were analyzed by SPSS 16.0.
 The total incidence of short PR interval was 19.65% (791/4 025). The age group from birth to 13 years old had a higher incidence than the other age groups (χ2=432, PPR intervals was 3.58% (144/4 025). The 1 year-old group had the highest incidence (6.74%), which decreased with the increase of age. The lowest incidence of long PR intervals occurred in the age group from 14-17 years old, which gradually increased after 50 years old. There were no significant differences in long (short) PR intervals between the gender (P>0.05).
 The incidence of long (short) PR intervals varies in different age groups of healthy people. The incidences of long (short) PR intervals in children before 10 years old are higher than those in adults, especially the short PR intervals, as a result of the heart rate affected by childhood autonomic nervous function and the change in atrial volume with age. Adults have long (short) PR interval should be regularly followed-up to prevent cardiovascular events.

  16. Quantum magnification of classical sub-Planck phase space features

    International Nuclear Information System (INIS)

    Hensinger, W.K.; Heckenberg, N.; Rubinsztein-Dunlop, H.; Delande, D.

    2002-01-01

    Full text: To understand the relationship between quantum mechanics and classical physics a crucial question to be answered is how distinct classical dynamical phase space features translate into the quantum picture. This problem becomes even more interesting if these phase space features occupy a much smaller volume than ℎ in a phase space spanned by two non-commuting variables such as position and momentum. The question whether phase space structures in quantum mechanics associated with sub-Planck scales have physical signatures has recently evoked a lot of discussion. Here we will show that sub-Planck classical dynamical phase space structures, for example regions of regular motion, can give rise to states whose phase space representation is of size ℎ or larger. This is illustrated using period-1 regions of regular motion (modes of oscillatory motion of a particle in a modulated well) whose volume is distinctly smaller than Planck's constant. They are magnified in the quantum picture and appear as states whose phase space representation is of size h or larger. Cold atoms provide an ideal test bed to probe such fundamental aspects of quantum and classical dynamics. In the experiment a Bose-Einstein condensate is loaded into a far detuned optical lattice. The lattice depth is modulated resulting in the emergence of regions of regular motion surrounded by chaotic motion in the phase space spanned by position and momentum of the atoms along the standing wave. Sub-Planck scaled phase space features in the classical phase space are magnified and appear as distinct broad peaks in the atomic momentum distribution. The corresponding quantum analysis shows states of size Ti which can be associated with much smaller classical dynamical phase space features. This effect may considered as the dynamical equivalent of the Goldstone and Jaffe theorem which predicts the existence of at least one bound state at a bend in a two or three dimensional spatial potential

  17. Higher derivative regularization and chiral anomaly

    International Nuclear Information System (INIS)

    Nagahama, Yoshinori.

    1985-02-01

    A higher derivative regularization which automatically leads to the consistent chiral anomaly is analyzed in detail. It explicitly breaks all the local gauge symmetry but preserves global chiral symmetry and leads to the chirally symmetric consistent anomaly. This regularization thus clarifies the physics content contained in the consistent anomaly. We also briefly comment on the application of this higher derivative regularization to massless QED. (author)

  18. Generalisation for regular black holes on general relativity to f(R) gravity

    Energy Technology Data Exchange (ETDEWEB)

    Rodrigues, Manuel E. [Universidade Federal do Para Campus Universitario de Abaetetuba, Faculdade de Ciencias Exatas e Tecnologia, Abaetetuba, Para (Brazil); Universidade Federal do Para, Faculdade de Fisica, PPGF, Belem, Para (Brazil); Fabris, Julio C. [Universidade Federal do Espirito Santo, Vitoria, ES (Brazil); National Research Nuclear University MEPhI, Moscow (Russian Federation); Junior, Ednaldo L.B. [Universidade Federal do Para, Faculdade de Fisica, PPGF, Belem, Para (Brazil); Universidade Federal do Para, Campus Universitario de Tucurui, Faculdade de Engenharia da Computacao, Tucurui, Para (Brazil); Marques, Glauber T. [Universidade Federal Rural da Amazonia ICIBE - LASIC, Belem, PA (Brazil)

    2016-05-15

    IIn this paper, we determine regular black hole solutions using a very general f(R) theory, coupled to a nonlinear electromagnetic field given by a Lagrangian L{sub NED}. The functions f(R) and L{sub NED} are in principle left unspecified. Instead, the model is constructed through a choice of the mass function M(r) presented in the metric coefficients. Solutions which have a regular behaviour of the geometric invariants are found. These solutions have two horizons, the event horizon and the Cauchy horizon. All energy conditions are satisfied in the whole space-time, except the strong energy condition (SEC), which is violated near the Cauchy horizon.We present also a new theorem related to the energy conditions in f(R) gravity, re-obtaining the well-known conditions in the context of general relativity when the geometry of the solution is the same. (orig.)

  19. Ito's formula in UMD Banach spaces and regularity of solution of the Zakai equation

    NARCIS (Netherlands)

    Brzezniak, Z.; Van Neerven, J.M.A.M.; Veraar, M.C.; Weis, L.

    2008-01-01

    Using the theory of stochastic integration for processes with values in a UMD Banach space developed recently by the authors, an Itô formula is proved which is applied to prove the existence of strong solutions for a class of stochastic evolution equations in UMD Banach spaces. The abstract results

  20. Multiple graph regularized protein domain ranking

    KAUST Repository

    Wang, Jim Jing-Yan

    2012-11-19

    Background: Protein domain ranking is a fundamental task in structural biology. Most protein domain ranking methods rely on the pairwise comparison of protein domains while neglecting the global manifold structure of the protein domain database. Recently, graph regularized ranking that exploits the global structure of the graph defined by the pairwise similarities has been proposed. However, the existing graph regularized ranking methods are very sensitive to the choice of the graph model and parameters, and this remains a difficult problem for most of the protein domain ranking methods.Results: To tackle this problem, we have developed the Multiple Graph regularized Ranking algorithm, MultiG-Rank. Instead of using a single graph to regularize the ranking scores, MultiG-Rank approximates the intrinsic manifold of protein domain distribution by combining multiple initial graphs for the regularization. Graph weights are learned with ranking scores jointly and automatically, by alternately minimizing an objective function in an iterative algorithm. Experimental results on a subset of the ASTRAL SCOP protein domain database demonstrate that MultiG-Rank achieves a better ranking performance than single graph regularized ranking methods and pairwise similarity based ranking methods.Conclusion: The problem of graph model and parameter selection in graph regularized protein domain ranking can be solved effectively by combining multiple graphs. This aspect of generalization introduces a new frontier in applying multiple graphs to solving protein domain ranking applications. 2012 Wang et al; licensee BioMed Central Ltd.

  1. Multiple graph regularized protein domain ranking

    KAUST Repository

    Wang, Jim Jing-Yan; Bensmail, Halima; Gao, Xin

    2012-01-01

    Background: Protein domain ranking is a fundamental task in structural biology. Most protein domain ranking methods rely on the pairwise comparison of protein domains while neglecting the global manifold structure of the protein domain database. Recently, graph regularized ranking that exploits the global structure of the graph defined by the pairwise similarities has been proposed. However, the existing graph regularized ranking methods are very sensitive to the choice of the graph model and parameters, and this remains a difficult problem for most of the protein domain ranking methods.Results: To tackle this problem, we have developed the Multiple Graph regularized Ranking algorithm, MultiG-Rank. Instead of using a single graph to regularize the ranking scores, MultiG-Rank approximates the intrinsic manifold of protein domain distribution by combining multiple initial graphs for the regularization. Graph weights are learned with ranking scores jointly and automatically, by alternately minimizing an objective function in an iterative algorithm. Experimental results on a subset of the ASTRAL SCOP protein domain database demonstrate that MultiG-Rank achieves a better ranking performance than single graph regularized ranking methods and pairwise similarity based ranking methods.Conclusion: The problem of graph model and parameter selection in graph regularized protein domain ranking can be solved effectively by combining multiple graphs. This aspect of generalization introduces a new frontier in applying multiple graphs to solving protein domain ranking applications. 2012 Wang et al; licensee BioMed Central Ltd.

  2. Multiple graph regularized protein domain ranking

    Directory of Open Access Journals (Sweden)

    Wang Jim

    2012-11-01

    Full Text Available Abstract Background Protein domain ranking is a fundamental task in structural biology. Most protein domain ranking methods rely on the pairwise comparison of protein domains while neglecting the global manifold structure of the protein domain database. Recently, graph regularized ranking that exploits the global structure of the graph defined by the pairwise similarities has been proposed. However, the existing graph regularized ranking methods are very sensitive to the choice of the graph model and parameters, and this remains a difficult problem for most of the protein domain ranking methods. Results To tackle this problem, we have developed the Multiple Graph regularized Ranking algorithm, MultiG-Rank. Instead of using a single graph to regularize the ranking scores, MultiG-Rank approximates the intrinsic manifold of protein domain distribution by combining multiple initial graphs for the regularization. Graph weights are learned with ranking scores jointly and automatically, by alternately minimizing an objective function in an iterative algorithm. Experimental results on a subset of the ASTRAL SCOP protein domain database demonstrate that MultiG-Rank achieves a better ranking performance than single graph regularized ranking methods and pairwise similarity based ranking methods. Conclusion The problem of graph model and parameter selection in graph regularized protein domain ranking can be solved effectively by combining multiple graphs. This aspect of generalization introduces a new frontier in applying multiple graphs to solving protein domain ranking applications.

  3. 75 FR 53966 - Regular Meeting

    Science.gov (United States)

    2010-09-02

    ... FARM CREDIT SYSTEM INSURANCE CORPORATION Regular Meeting AGENCY: Farm Credit System Insurance Corporation Board. SUMMARY: Notice is hereby given of the regular meeting of the Farm Credit System Insurance Corporation Board (Board). DATE AND TIME: The meeting of the Board will be held at the offices of the Farm...

  4. Impulsive sounds change European seabass swimming patterns: Influence of pulse repetition interval

    International Nuclear Information System (INIS)

    Neo, Y.Y.; Ufkes, E.; Kastelein, R.A.; Winter, H.V.; Cate, C. ten; Slabbekoorn, H.

    2015-01-01

    Highlights: • We exposed impulsive sounds of different repetition intervals to European seabass. • Immediate behavioural changes mirrored previous indoor & outdoor studies. • Repetition intervals influenced the impacts differentially but not the recovery. • Sound temporal patterns may be more important than some standard metrics. - Abstract: Seismic shootings and offshore pile-driving are regularly performed, emitting significant amounts of noise that may negatively affect fish behaviour. The pulse repetition interval (PRI) of these impulsive sounds may vary considerably and influence the behavioural impact and recovery. Here, we tested the effect of four PRIs (0.5–4.0 s) on European seabass swimming patterns in an outdoor basin. At the onset of the sound exposures, the fish swam faster and dived deeper in tighter shoals. PRI affected the immediate and delayed behavioural changes but not the recovery time. Our study highlights that (1) the behavioural changes of captive European seabass were consistent with previous indoor and outdoor studies; (2) PRI could influence behavioural impact differentially, which may have management implications; (3) some acoustic metrics, e.g. SEL cum , may have limited predictive power to assess the strength of behavioural impacts of noise. Noise impact assessments need to consider the contribution of sound temporal structure

  5. Work and family life of childrearing women workers in Japan: comparison of non-regular employees with short working hours, non-regular employees with long working hours, and regular employees.

    Science.gov (United States)

    Seto, Masako; Morimoto, Kanehisa; Maruyama, Soichiro

    2006-05-01

    This study assessed the working and family life characteristics, and the degree of domestic and work strain of female workers with different employment statuses and weekly working hours who are rearing children. Participants were the mothers of preschoolers in a large Japanese city. We classified the women into three groups according to the hours they worked and their employment conditions. The three groups were: non-regular employees working less than 30 h a week (n=136); non-regular employees working 30 h or more per week (n=141); and regular employees working 30 h or more a week (n=184). We compared among the groups the subjective values of work, financial difficulties, childcare and housework burdens, psychological effects, and strains such as work and family strain, work-family conflict, and work dissatisfaction. Regular employees were more likely to report job pressures and inflexible work schedules and to experience more strain related to work and family than non-regular employees. Non-regular employees were more likely to be facing financial difficulties. In particular, non-regular employees working longer hours tended to encounter socioeconomic difficulties and often lacked support from family and friends. Female workers with children may have different social backgrounds and different stressors according to their working hours and work status.

  6. Dose domain regularization of MLC leaf patterns for highly complex IMRT plans

    Energy Technology Data Exchange (ETDEWEB)

    Nguyen, Dan; Yu, Victoria Y.; Ruan, Dan; Cao, Minsong; Low, Daniel A.; Sheng, Ke, E-mail: ksheng@mednet.ucla.edu [Department of Radiation Oncology, University of California Los Angeles, Los Angeles, California 90095 (United States); O’Connor, Daniel [Department of Mathematics, University of California Los Angeles, Los Angeles, California 90095 (United States)

    2015-04-15

    Purpose: The advent of automated beam orientation and fluence optimization enables more complex intensity modulated radiation therapy (IMRT) planning using an increasing number of fields to exploit the expanded solution space. This has created a challenge in converting complex fluences to robust multileaf collimator (MLC) segments for delivery. A novel method to regularize the fluence map and simplify MLC segments is introduced to maximize delivery efficiency, accuracy, and plan quality. Methods: In this work, we implemented a novel approach to regularize optimized fluences in the dose domain. The treatment planning problem was formulated in an optimization framework to minimize the segmentation-induced dose distribution degradation subject to a total variation regularization to encourage piecewise smoothness in fluence maps. The optimization problem was solved using a first-order primal-dual algorithm known as the Chambolle-Pock algorithm. Plans for 2 GBM, 2 head and neck, and 2 lung patients were created using 20 automatically selected and optimized noncoplanar beams. The fluence was first regularized using Chambolle-Pock and then stratified into equal steps, and the MLC segments were calculated using a previously described level reducing method. Isolated apertures with sizes smaller than preset thresholds of 1–3 bixels, which are square units of an IMRT fluence map from MLC discretization, were removed from the MLC segments. Performance of the dose domain regularized (DDR) fluences was compared to direct stratification and direct MLC segmentation (DMS) of the fluences using level reduction without dose domain fluence regularization. Results: For all six cases, the DDR method increased the average planning target volume dose homogeneity (D95/D5) from 0.814 to 0.878 while maintaining equivalent dose to organs at risk (OARs). Regularized fluences were more robust to MLC sequencing, particularly to the stratification and small aperture removal. The maximum and

  7. Birth Spacing of Pregnant Women in Nepal: A Community-Based Study.

    Science.gov (United States)

    Karkee, Rajendra; Lee, Andy H

    2016-01-01

    Optimal birth spacing has health advantages for both mother and child. In developing countries, shorter birth intervals are common and associated with social, cultural, and economic factors, as well as a lack of family planning. This study investigated the first birth interval after marriage and preceding interbirth interval in Nepal. A community-based prospective cohort study was conducted in the Kaski district of Nepal. Information on birth spacing, demographic, and obstetric characteristics was obtained from 701 pregnant women using a structured questionnaire. Logistic regression analyses were performed to ascertain factors associated with short birth spacing. About 39% of primiparous women gave their first child birth within 1 year of marriage and 23% of multiparous women had short preceding interbirth intervals (gender equality in society.

  8. Dangerous Spaces: Threatening Sites for Social Justice

    Science.gov (United States)

    Schostak, John

    2012-01-01

    There is nothing natural about space as it is understood here. Spacing is an act that constructs relationships, intervals, separations and thus boundaries. The earth has no territories other than those imagined and enforced through acts of territorialisation. A city has its private spaces closed to open access and open spaces that are inscribed…

  9. Incremental projection approach of regularization for inverse problems

    Energy Technology Data Exchange (ETDEWEB)

    Souopgui, Innocent, E-mail: innocent.souopgui@usm.edu [The University of Southern Mississippi, Department of Marine Science (United States); Ngodock, Hans E., E-mail: hans.ngodock@nrlssc.navy.mil [Naval Research Laboratory (United States); Vidard, Arthur, E-mail: arthur.vidard@imag.fr; Le Dimet, François-Xavier, E-mail: ledimet@imag.fr [Laboratoire Jean Kuntzmann (France)

    2016-10-15

    This paper presents an alternative approach to the regularized least squares solution of ill-posed inverse problems. Instead of solving a minimization problem with an objective function composed of a data term and a regularization term, the regularization information is used to define a projection onto a convex subspace of regularized candidate solutions. The objective function is modified to include the projection of each iterate in the place of the regularization. Numerical experiments based on the problem of motion estimation for geophysical fluid images, show the improvement of the proposed method compared with regularization methods. For the presented test case, the incremental projection method uses 7 times less computation time than the regularization method, to reach the same error target. Moreover, at convergence, the incremental projection is two order of magnitude more accurate than the regularization method.

  10. A Regularized Linear Dynamical System Framework for Multivariate Time Series Analysis.

    Science.gov (United States)

    Liu, Zitao; Hauskrecht, Milos

    2015-01-01

    Linear Dynamical System (LDS) is an elegant mathematical framework for modeling and learning Multivariate Time Series (MTS). However, in general, it is difficult to set the dimension of an LDS's hidden state space. A small number of hidden states may not be able to model the complexities of a MTS, while a large number of hidden states can lead to overfitting. In this paper, we study learning methods that impose various regularization penalties on the transition matrix of the LDS model and propose a regularized LDS learning framework (rLDS) which aims to (1) automatically shut down LDSs' spurious and unnecessary dimensions, and consequently, address the problem of choosing the optimal number of hidden states; (2) prevent the overfitting problem given a small amount of MTS data; and (3) support accurate MTS forecasting. To learn the regularized LDS from data we incorporate a second order cone program and a generalized gradient descent method into the Maximum a Posteriori framework and use Expectation Maximization to obtain a low-rank transition matrix of the LDS model. We propose two priors for modeling the matrix which lead to two instances of our rLDS. We show that our rLDS is able to recover well the intrinsic dimensionality of the time series dynamics and it improves the predictive performance when compared to baselines on both synthetic and real-world MTS datasets.

  11. Regularization by Functions of Bounded Variation and Applications to Image Enhancement

    International Nuclear Information System (INIS)

    Casas, E.; Kunisch, K.; Pola, C.

    1999-01-01

    Optimization problems regularized by bounded variation seminorms are analyzed. The optimality system is obtained and finite-dimensional approximations of bounded variation function spaces as well as of the optimization problems are studied. It is demonstrated that the choice of the vector norm in the definition of the bounded variation seminorm is of special importance for approximating subspaces consisting of piecewise constant functions. Algorithms based on a primal-dual framework that exploit the structure of these nondifferentiable optimization problems are proposed. Numerical examples are given for denoising of blocky images with very high noise

  12. Adaptive regularization

    DEFF Research Database (Denmark)

    Hansen, Lars Kai; Rasmussen, Carl Edward; Svarer, C.

    1994-01-01

    Regularization, e.g., in the form of weight decay, is important for training and optimization of neural network architectures. In this work the authors provide a tool based on asymptotic sampling theory, for iterative estimation of weight decay parameters. The basic idea is to do a gradient desce...

  13. Regularizing portfolio optimization

    International Nuclear Information System (INIS)

    Still, Susanne; Kondor, Imre

    2010-01-01

    The optimization of large portfolios displays an inherent instability due to estimation error. This poses a fundamental problem, because solutions that are not stable under sample fluctuations may look optimal for a given sample, but are, in effect, very far from optimal with respect to the average risk. In this paper, we approach the problem from the point of view of statistical learning theory. The occurrence of the instability is intimately related to over-fitting, which can be avoided using known regularization methods. We show how regularized portfolio optimization with the expected shortfall as a risk measure is related to support vector regression. The budget constraint dictates a modification. We present the resulting optimization problem and discuss the solution. The L2 norm of the weight vector is used as a regularizer, which corresponds to a diversification 'pressure'. This means that diversification, besides counteracting downward fluctuations in some assets by upward fluctuations in others, is also crucial because it improves the stability of the solution. The approach we provide here allows for the simultaneous treatment of optimization and diversification in one framework that enables the investor to trade off between the two, depending on the size of the available dataset.

  14. Regularizing portfolio optimization

    Science.gov (United States)

    Still, Susanne; Kondor, Imre

    2010-07-01

    The optimization of large portfolios displays an inherent instability due to estimation error. This poses a fundamental problem, because solutions that are not stable under sample fluctuations may look optimal for a given sample, but are, in effect, very far from optimal with respect to the average risk. In this paper, we approach the problem from the point of view of statistical learning theory. The occurrence of the instability is intimately related to over-fitting, which can be avoided using known regularization methods. We show how regularized portfolio optimization with the expected shortfall as a risk measure is related to support vector regression. The budget constraint dictates a modification. We present the resulting optimization problem and discuss the solution. The L2 norm of the weight vector is used as a regularizer, which corresponds to a diversification 'pressure'. This means that diversification, besides counteracting downward fluctuations in some assets by upward fluctuations in others, is also crucial because it improves the stability of the solution. The approach we provide here allows for the simultaneous treatment of optimization and diversification in one framework that enables the investor to trade off between the two, depending on the size of the available dataset.

  15. Space Mathematics, A Resource for Teachers Outlining Supplementary Space-Related Problems in Mathematics.

    Science.gov (United States)

    Reynolds, Thomas D.; And Others

    This compilation of 138 problems illustrating applications of high school mathematics to various aspects of space science is intended as a resource from which the teacher may select questions to supplement his regular course. None of the problems require a knowledge of calculus or physics, and solutions are presented along with the problem…

  16. Tessellating the Sphere with Regular Polygons

    Science.gov (United States)

    Soto-Johnson, Hortensia; Bechthold, Dawn

    2004-01-01

    Tessellations in the Euclidean plane and regular polygons that tessellate the sphere are reviewed. The regular polygons that can possibly tesellate the sphere are spherical triangles, squares and pentagons.

  17. Regular exercise behaviour and intention and symptoms of anxiety and depression in coronary heart disease patients across Europe: Results from the EUROASPIRE III survey.

    Science.gov (United States)

    Prugger, Christof; Wellmann, Jürgen; Heidrich, Jan; De Bacquer, Dirk; De Smedt, Delphine; De Backer, Guy; Reiner, Željko; Empana, Jean-Philippe; Fras, Zlatko; Gaita, Dan; Jennings, Catriona; Kotseva, Kornelia; Wood, David; Keil, Ulrich

    2017-01-01

    Regular exercise lowers the risk of cardiovascular death in coronary heart disease (CHD) patients. We aimed to investigate regular exercise behaviour and intention in relation to symptoms of anxiety and depression in CHD patients across Europe. This study was based on a multicentre cross-sectional survey. In the EUROpean Action on Secondary and Primary Prevention through Intervention to Reduce Events (EUROASPIRE) III survey, 8966 CHD patients patients exercised or intended to exercise regularly was assessed using the Stages of Change questionnaire in 8330 patients. Symptoms of anxiety and depression were evaluated using the Hospital Anxiety and Depression Scale. Total physical activity was measured by the International Physical Activity Questionnaire in patients from a subset of 14 countries. Overall, 50.3% of patients were not intending to exercise regularly, 15.9% were intending to exercise regularly, and 33.8% were exercising regularly. Patients with severe symptoms of depression less frequently exercised regularly than patients with symptoms in the normal range (20.2%, 95% confidence interval (CI) 14.8-26.8 vs 36.7%, 95% CI 29.8-44.2). Among patients not exercising regularly, patients with severe symptoms of depression were less likely to have an intention to exercise regularly (odds ratio 0.62, 95% CI 0.46-0.85). Symptoms of anxiety did not affect regular exercise intention. In sensitivity analysis, results were consistent when adjusting for total physical activity. Lower frequency of regular exercise and decreased likelihood of exercise intention were observed in CHD patients with severe depressive symptoms. Severe symptoms of depression may preclude CHD patients from performing regular exercise. © The European Society of Cardiology 2016.

  18. Accretion onto some well-known regular black holes

    International Nuclear Information System (INIS)

    Jawad, Abdul; Shahzad, M.U.

    2016-01-01

    In this work, we discuss the accretion onto static spherically symmetric regular black holes for specific choices of the equation of state parameter. The underlying regular black holes are charged regular black holes using the Fermi-Dirac distribution, logistic distribution, nonlinear electrodynamics, respectively, and Kehagias-Sftesos asymptotically flat regular black holes. We obtain the critical radius, critical speed, and squared sound speed during the accretion process near the regular black holes. We also study the behavior of radial velocity, energy density, and the rate of change of the mass for each of the regular black holes. (orig.)

  19. Accretion onto some well-known regular black holes

    Energy Technology Data Exchange (ETDEWEB)

    Jawad, Abdul; Shahzad, M.U. [COMSATS Institute of Information Technology, Department of Mathematics, Lahore (Pakistan)

    2016-03-15

    In this work, we discuss the accretion onto static spherically symmetric regular black holes for specific choices of the equation of state parameter. The underlying regular black holes are charged regular black holes using the Fermi-Dirac distribution, logistic distribution, nonlinear electrodynamics, respectively, and Kehagias-Sftesos asymptotically flat regular black holes. We obtain the critical radius, critical speed, and squared sound speed during the accretion process near the regular black holes. We also study the behavior of radial velocity, energy density, and the rate of change of the mass for each of the regular black holes. (orig.)

  20. Accretion onto some well-known regular black holes

    Science.gov (United States)

    Jawad, Abdul; Shahzad, M. Umair

    2016-03-01

    In this work, we discuss the accretion onto static spherically symmetric regular black holes for specific choices of the equation of state parameter. The underlying regular black holes are charged regular black holes using the Fermi-Dirac distribution, logistic distribution, nonlinear electrodynamics, respectively, and Kehagias-Sftesos asymptotically flat regular black holes. We obtain the critical radius, critical speed, and squared sound speed during the accretion process near the regular black holes. We also study the behavior of radial velocity, energy density, and the rate of change of the mass for each of the regular black holes.

  1. A variational regularization of Abel transform for GPS radio occultation

    Directory of Open Access Journals (Sweden)

    T.-K. Wee

    2018-04-01

    Full Text Available In the Global Positioning System (GPS radio occultation (RO technique, the inverse Abel transform of measured bending angle (Abel inversion, hereafter AI is the standard means of deriving the refractivity. While concise and straightforward to apply, the AI accumulates and propagates the measurement error downward. The measurement error propagation is detrimental to the refractivity in lower altitudes. In particular, it builds up negative refractivity bias in the tropical lower troposphere. An alternative to AI is the numerical inversion of the forward Abel transform, which does not incur the integration of error-possessing measurement and thus precludes the error propagation. The variational regularization (VR proposed in this study approximates the inversion of the forward Abel transform by an optimization problem in which the regularized solution describes the measurement as closely as possible within the measurement's considered accuracy. The optimization problem is then solved iteratively by means of the adjoint technique. VR is formulated with error covariance matrices, which permit a rigorous incorporation of prior information on measurement error characteristics and the solution's desired behavior into the regularization. VR holds the control variable in the measurement space to take advantage of the posterior height determination and to negate the measurement error due to the mismodeling of the refractional radius. The advantages of having the solution and the measurement in the same space are elaborated using a purposely corrupted synthetic sounding with a known true solution. The competency of VR relative to AI is validated with a large number of actual RO soundings. The comparison to nearby radiosonde observations shows that VR attains considerably smaller random and systematic errors compared to AI. A noteworthy finding is that in the heights and areas that the measurement bias is supposedly small, VR follows AI very closely in the

  2. A variational regularization of Abel transform for GPS radio occultation

    Science.gov (United States)

    Wee, Tae-Kwon

    2018-04-01

    In the Global Positioning System (GPS) radio occultation (RO) technique, the inverse Abel transform of measured bending angle (Abel inversion, hereafter AI) is the standard means of deriving the refractivity. While concise and straightforward to apply, the AI accumulates and propagates the measurement error downward. The measurement error propagation is detrimental to the refractivity in lower altitudes. In particular, it builds up negative refractivity bias in the tropical lower troposphere. An alternative to AI is the numerical inversion of the forward Abel transform, which does not incur the integration of error-possessing measurement and thus precludes the error propagation. The variational regularization (VR) proposed in this study approximates the inversion of the forward Abel transform by an optimization problem in which the regularized solution describes the measurement as closely as possible within the measurement's considered accuracy. The optimization problem is then solved iteratively by means of the adjoint technique. VR is formulated with error covariance matrices, which permit a rigorous incorporation of prior information on measurement error characteristics and the solution's desired behavior into the regularization. VR holds the control variable in the measurement space to take advantage of the posterior height determination and to negate the measurement error due to the mismodeling of the refractional radius. The advantages of having the solution and the measurement in the same space are elaborated using a purposely corrupted synthetic sounding with a known true solution. The competency of VR relative to AI is validated with a large number of actual RO soundings. The comparison to nearby radiosonde observations shows that VR attains considerably smaller random and systematic errors compared to AI. A noteworthy finding is that in the heights and areas that the measurement bias is supposedly small, VR follows AI very closely in the mean refractivity

  3. Computation of robustly stabilizing PID controllers for interval systems.

    Science.gov (United States)

    Matušů, Radek; Prokop, Roman

    2016-01-01

    The paper is focused on the computation of all possible robustly stabilizing Proportional-Integral-Derivative (PID) controllers for plants with interval uncertainty. The main idea of the proposed method is based on Tan's (et al.) technique for calculation of (nominally) stabilizing PI and PID controllers or robustly stabilizing PI controllers by means of plotting the stability boundary locus in either P-I plane or P-I-D space. Refinement of the existing method by consideration of 16 segment plants instead of 16 Kharitonov plants provides an elegant and efficient tool for finding all robustly stabilizing PID controllers for an interval system. The validity and relatively effortless application of presented theoretical concepts are demonstrated through a computation and simulation example in which the uncertain mathematical model of an experimental oblique wing aircraft is robustly stabilized.

  4. Recommended Changes to Interval Management to Achieve Operational Implementation

    Science.gov (United States)

    Baxley, Brian; Swieringa, Kurt; Roper, Roy; Hubbs, Clay; Goess, Paul; Shay, Richard

    2017-01-01

    A 19-day flight test of an Interval Management (IM) avionics prototype was conducted in Washington State using three aircraft to precisely achieve and maintain a spacing interval behind the preceding aircraft. NASA contracted with Boeing, Honeywell, and United Airlines to build this prototype, and then worked closely with them, the FAA, and other industry partners to test this prototype in flight. Four different IM operation types were investigated during this test in the en route, arrival, and final approach phases of flight. Many of the IM operations met or exceeded the design goals established prior to the test. However, there were issues discovered throughout the flight test, including the rate and magnitude of IM commanded speed changes and the difference between expected and actual aircraft deceleration rates.

  5. Programming with Intervals

    Science.gov (United States)

    Matsakis, Nicholas D.; Gross, Thomas R.

    Intervals are a new, higher-level primitive for parallel programming with which programmers directly construct the program schedule. Programs using intervals can be statically analyzed to ensure that they do not deadlock or contain data races. In this paper, we demonstrate the flexibility of intervals by showing how to use them to emulate common parallel control-flow constructs like barriers and signals, as well as higher-level patterns such as bounded-buffer producer-consumer. We have implemented intervals as a publicly available library for Java and Scala.

  6. Factors associated with regular consumption of obesogenic foods: National School-Based Student Health Hurvey, 2012

    Directory of Open Access Journals (Sweden)

    Giovana LONGO-SILVA

    Full Text Available ABSTRACT Objective: To investigate the frequency of consumption of obesogenic foods among adolescents and its association with sociodemographic, family, behavioral, and environmental variables. Methods: Secondary data from the National School-Based Student Health Hurvey were analyzed from a representative sample of 9th grade Brazilian students (high school. A self-administered questionnaire, organized into thematic blocks, was used. The dependent variables were the consumption of deep fried snacks, packaged snacks, sugar candies, and soft drinks; consumption frequency for the seven days preceding the study was analyzed. Bivariate analysis was carried out to determine the empirical relationship between the regular consumption of these foods (≥3 days/week with sociodemographic, family, behavioral, and school structural variables. p-value <0.20 was used as the criterion for initial inclusion in the multivariate logistic analysis, which was conducted using the "Enter" method, and the results were expressed as adjusted odds ratios with 95% confidence interval and p<0.05 indicating a statistically significance. Results: Regular food consumption ranged from 27.17% to 65.96%. The variables female gender, mobile phone ownership, Internet access at home, tobacco use, alcohol consumption, regular physical activity, eating while watching television or studying, watching television for at least 2 hours a day, and not willing to lose weight were associated in the final logistic models of all foods analyzed. Conclusion: It was concluded that fried snacks, packaged snacks, sugar candies, and soft drinks are regularly consumed by adolescents and that such consumption was associated with the sociodemographic, family, behavioral, and school structural variables.

  7. A Comparative Study of Interval Management Control Law Capabilities

    Science.gov (United States)

    Barmore, Bryan E.; Smith, Colin L.; Palmer, Susan O.; Abbott, Terence S.

    2012-01-01

    This paper presents a new tool designed to allow for rapid development and testing of different control algorithms for airborne spacing. This tool, Interval Management Modeling and Spacing Tool (IM MAST), is a fast-time, low-fidelity tool created to model the approach of aircraft to a runway, with a focus on their interactions with each other. Errors can be induced between pairs of aircraft by varying initial positions, winds, speed profiles, and altitude profiles. Results to-date show that only a few of the algorithms tested had poor behavior in the arrival and approach environment. The majority of the algorithms showed only minimal variation in performance under the test conditions. Trajectory-based algorithms showed high susceptibility to wind forecast errors, while performing marginally better than the other algorithms under other conditions. Trajectory-based algorithms have a sizable advantage, however, of being able to perform relative spacing operations between aircraft on different arrival routes and flight profiles without employing ghosting. methods. This comes at the higher cost of substantially increased complexity, however. Additionally, it was shown that earlier initiation of relative spacing operations provided more time for corrections to be made without any significant problems in the spacing operation itself. Initiating spacing farther out, however, would require more of the aircraft to begin spacing before they merge onto a common route.

  8. Trajectories of problem video gaming among adult regular gamers: an 18-month longitudinal study.

    Science.gov (United States)

    King, Daniel L; Delfabbro, Paul H; Griffiths, Mark D

    2013-01-01

    A three-wave, longitudinal study examined the long-term trajectory of problem gaming symptoms among adult regular video gamers. Potential changes in problem gaming status were assessed at two intervals using an online survey over an 18-month period. Participants (N=117) were recruited by an advertisement posted on the public forums of multiple Australian video game-related websites. Inclusion criteria were being of adult age and having a video gaming history of at least 1 hour of gaming every week over the past 3 months. Two groups of adult video gamers were identified: those players who did (N=37) and those who did not (N=80) identify as having a serious gaming problem at the initial survey intake. The results showed that regular gamers who self-identified as having a video gaming problem at baseline reported more severe problem gaming symptoms than normal gamers, at all time points. However, both groups experienced a significant decline in problem gaming symptoms over an 18-month period, controlling for age, video gaming activity, and psychopathological symptoms.

  9. A regularized approach for geodesic-based semisupervised multimanifold learning.

    Science.gov (United States)

    Fan, Mingyu; Zhang, Xiaoqin; Lin, Zhouchen; Zhang, Zhongfei; Bao, Hujun

    2014-05-01

    Geodesic distance, as an essential measurement for data dissimilarity, has been successfully used in manifold learning. However, most geodesic distance-based manifold learning algorithms have two limitations when applied to classification: 1) class information is rarely used in computing the geodesic distances between data points on manifolds and 2) little attention has been paid to building an explicit dimension reduction mapping for extracting the discriminative information hidden in the geodesic distances. In this paper, we regard geodesic distance as a kind of kernel, which maps data from linearly inseparable space to linear separable distance space. In doing this, a new semisupervised manifold learning algorithm, namely regularized geodesic feature learning algorithm, is proposed. The method consists of three techniques: a semisupervised graph construction method, replacement of original data points with feature vectors which are built by geodesic distances, and a new semisupervised dimension reduction method for feature vectors. Experiments on the MNIST, USPS handwritten digit data sets, MIT CBCL face versus nonface data set, and an intelligent traffic data set show the effectiveness of the proposed algorithm.

  10. Temporal regularity of the environment drives time perception

    OpenAIRE

    van Rijn, H; Rhodes, D; Di Luca, M

    2016-01-01

    It’s reasonable to assume that a regularly paced sequence should be perceived as regular, but here we show that perceived regularity depends on the context in which the sequence is embedded. We presented one group of participants with perceptually regularly paced sequences, and another group of participants with mostly irregularly paced sequences (75% irregular, 25% regular). The timing of the final stimulus in each sequence could be var- ied. In one experiment, we asked whether the last stim...

  11. Approximate message passing for nonconvex sparse regularization with stability and asymptotic analysis

    Science.gov (United States)

    Sakata, Ayaka; Xu, Yingying

    2018-03-01

    We analyse a linear regression problem with nonconvex regularization called smoothly clipped absolute deviation (SCAD) under an overcomplete Gaussian basis for Gaussian random data. We propose an approximate message passing (AMP) algorithm considering nonconvex regularization, namely SCAD-AMP, and analytically show that the stability condition corresponds to the de Almeida-Thouless condition in spin glass literature. Through asymptotic analysis, we show the correspondence between the density evolution of SCAD-AMP and the replica symmetric (RS) solution. Numerical experiments confirm that for a sufficiently large system size, SCAD-AMP achieves the optimal performance predicted by the replica method. Through replica analysis, a phase transition between replica symmetric and replica symmetry breaking (RSB) region is found in the parameter space of SCAD. The appearance of the RS region for a nonconvex penalty is a significant advantage that indicates the region of smooth landscape of the optimization problem. Furthermore, we analytically show that the statistical representation performance of the SCAD penalty is better than that of \

  12. Continuous time modelling with individually varying time intervals for oscillating and non-oscillating processes.

    Science.gov (United States)

    Voelkle, Manuel C; Oud, Johan H L

    2013-02-01

    When designing longitudinal studies, researchers often aim at equal intervals. In practice, however, this goal is hardly ever met, with different time intervals between assessment waves and different time intervals between individuals being more the rule than the exception. One of the reasons for the introduction of continuous time models by means of structural equation modelling has been to deal with irregularly spaced assessment waves (e.g., Oud & Delsing, 2010). In the present paper we extend the approach to individually varying time intervals for oscillating and non-oscillating processes. In addition, we show not only that equal intervals are unnecessary but also that it can be advantageous to use unequal sampling intervals, in particular when the sampling rate is low. Two examples are provided to support our arguments. In the first example we compare a continuous time model of a bivariate coupled process with varying time intervals to a standard discrete time model to illustrate the importance of accounting for the exact time intervals. In the second example the effect of different sampling intervals on estimating a damped linear oscillator is investigated by means of a Monte Carlo simulation. We conclude that it is important to account for individually varying time intervals, and encourage researchers to conceive of longitudinal studies with different time intervals within and between individuals as an opportunity rather than a problem. © 2012 The British Psychological Society.

  13. The uniqueness of the regularization procedure

    International Nuclear Information System (INIS)

    Brzezowski, S.

    1981-01-01

    On the grounds of the BPHZ procedure, the criteria of correct regularization in perturbation calculations of QFT are given, together with the prescription for dividing the regularized formulas into the finite and infinite parts. (author)

  14. Strong self-coupling expansion in the lattice-regularized standard SU(2) Higgs model

    International Nuclear Information System (INIS)

    Decker, K.; Weisz, P.; Montvay, I.

    1985-11-01

    Expectation values at an arbitrary point of the 3-dimensional coupling parameter space in the lattice-regularized SU(2) Higgs-model with a doublet scalar field are expressed by a series of expectation values at infinite self-coupling (lambda=infinite). Questions of convergence of this 'strong self-coupling expansion' (SSCE) are investigated. The SSCE is a potentially useful tool for the study of the lambda-dependence at any value (zero or non-zero) of the bare gauge coupling. (orig.)

  15. Strong self-coupling expansion in the lattice-regularized standard SU(2) Higgs model

    International Nuclear Information System (INIS)

    Decker, K.; Weisz, P.

    1986-01-01

    Expectation values at an arbitrary point of the 3-dimensional coupling parameter space in the lattice-regularized SU(2) Higgs model with a doublet scalar field are expressed by a series of expectation values at infinite self-coupling (lambda=infinite). Questions of convergence of this ''strong self-coupling expansion'' (SSCE) are investigated. The SSCE is a potentially useful tool for the study of the lambda-dependence at any value (zero or non-zero) of the bare gauge coupling. (orig.)

  16. Joint Segmentation and Shape Regularization with a Generalized Forward Backward Algorithm.

    Science.gov (United States)

    Stefanoiu, Anca; Weinmann, Andreas; Storath, Martin; Navab, Nassir; Baust, Maximilian

    2016-05-11

    This paper presents a method for the simultaneous segmentation and regularization of a series of shapes from a corresponding sequence of images. Such series arise as time series of 2D images when considering video data, or as stacks of 2D images obtained by slicewise tomographic reconstruction. We first derive a model where the regularization of the shape signal is achieved by a total variation prior on the shape manifold. The method employs a modified Kendall shape space to facilitate explicit computations together with the concept of Sobolev gradients. For the proposed model, we derive an efficient and computationally accessible splitting scheme. Using a generalized forward-backward approach, our algorithm treats the total variation atoms of the splitting via proximal mappings, whereas the data terms are dealt with by gradient descent. The potential of the proposed method is demonstrated on various application examples dealing with 3D data. We explain how to extend the proposed combined approach to shape fields which, for instance, arise in the context of 3D+t imaging modalities, and show an application in this setup as well.

  17. Coupling regularizes individual units in noisy populations

    International Nuclear Information System (INIS)

    Ly Cheng; Ermentrout, G. Bard

    2010-01-01

    The regularity of a noisy system can modulate in various ways. It is well known that coupling in a population can lower the variability of the entire network; the collective activity is more regular. Here, we show that diffusive (reciprocal) coupling of two simple Ornstein-Uhlenbeck (O-U) processes can regularize the individual, even when it is coupled to a noisier process. In cellular networks, the regularity of individual cells is important when a select few play a significant role. The regularizing effect of coupling surprisingly applies also to general nonlinear noisy oscillators. However, unlike with the O-U process, coupling-induced regularity is robust to different kinds of coupling. With two coupled noisy oscillators, we derive an asymptotic formula assuming weak noise and coupling for the variance of the period (i.e., spike times) that accurately captures this effect. Moreover, we find that reciprocal coupling can regularize the individual period of higher dimensional oscillators such as the Morris-Lecar and Brusselator models, even when coupled to noisier oscillators. Coupling can have a counterintuitive and beneficial effect on noisy systems. These results have implications for the role of connectivity with noisy oscillators and the modulation of variability of individual oscillators.

  18. Learning regularization parameters for general-form Tikhonov

    International Nuclear Information System (INIS)

    Chung, Julianne; Español, Malena I

    2017-01-01

    Computing regularization parameters for general-form Tikhonov regularization can be an expensive and difficult task, especially if multiple parameters or many solutions need to be computed in real time. In this work, we assume training data is available and describe an efficient learning approach for computing regularization parameters that can be used for a large set of problems. We consider an empirical Bayes risk minimization framework for finding regularization parameters that minimize average errors for the training data. We first extend methods from Chung et al (2011 SIAM J. Sci. Comput. 33 3132–52) to the general-form Tikhonov problem. Then we develop a learning approach for multi-parameter Tikhonov problems, for the case where all involved matrices are simultaneously diagonalizable. For problems where this is not the case, we describe an approach to compute near-optimal regularization parameters by using operator approximations for the original problem. Finally, we propose a new class of regularizing filters, where solutions correspond to multi-parameter Tikhonov solutions, that requires less data than previously proposed optimal error filters, avoids the generalized SVD, and allows flexibility and novelty in the choice of regularization matrices. Numerical results for 1D and 2D examples using different norms on the errors show the effectiveness of our methods. (paper)

  19. 5 CFR 551.421 - Regular working hours.

    Science.gov (United States)

    2010-01-01

    ... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false Regular working hours. 551.421 Section... Activities § 551.421 Regular working hours. (a) Under the Act there is no requirement that a Federal employee... distinction based on whether the activity is performed by an employee during regular working hours or outside...

  20. Regular extensions of some classes of grammars

    NARCIS (Netherlands)

    Nijholt, Antinus

    Culik and Cohen introduced the class of LR-regular grammars, an extension of the LR(k) grammars. In this report we consider the analogous extension of the LL(k) grammers, called the LL-regular grammars. The relations of this class of grammars to other classes of grammars are shown. Every LL-regular

  1. Differential regularization of a non-relativistic anyon model

    International Nuclear Information System (INIS)

    Freedman, D.Z.; Rius, N.

    1993-07-01

    Differential regularization is applied to a field theory of a non-relativistic charged boson field φ with λ(φ * φ) 2 self-interaction and coupling to a statistics-changing 0(1) Chern-Simons gauge field. Renormalized configuration-space amplitudes for all diagrams contributing to the φ * φ * φφ 4-point function, which is the only primitively divergent Green's function, are obtained up to 3-loop order. The renormalization group equations are explicitly checked, and the scheme dependence of the β-function is investigated. If the renormalization scheme is fixed to agree with a previous 1-loop calculation, the 2- and 3-loop contributions to β(λ, e) vanish, and β(λ, ε) itself vanishes when the ''self-dual'' condition relating λ to the gauge coupling e is imposed. (author). 12 refs, 1 fig

  2. Quantum mechanics on Laakso spaces

    Science.gov (United States)

    Kauffman, Christopher J.; Kesler, Robert M.; Parshall, Amanda G.; Stamey, Evelyn A.; Steinhurst, Benjamin A.

    2012-04-01

    We first review the spectrum of the Laplacian operator on a general Laakso space before considering modified Hamiltonians for the infinite square well, parabola, and Coulomb potentials. Additionally, we compute the spectrum for the Laplacian and its multiplicities when certain regions of a Laakso space are compressed or stretched and calculate the Casimir force experienced by two uncharged conducting plates by imposing physically relevant boundary conditions and then analytically regularizing the resulting zeta function. Lastly, we derive a general formula for the spectral zeta function and its derivative for Laakso spaces with strict self-similar structure before listing explicit spectral values for some special cases

  3. Near-Regular Structure Discovery Using Linear Programming

    KAUST Repository

    Huang, Qixing

    2014-06-02

    Near-regular structures are common in manmade and natural objects. Algorithmic detection of such regularity greatly facilitates our understanding of shape structures, leads to compact encoding of input geometries, and enables efficient generation and manipulation of complex patterns on both acquired and synthesized objects. Such regularity manifests itself both in the repetition of certain geometric elements, as well as in the structured arrangement of the elements. We cast the regularity detection problem as an optimization and efficiently solve it using linear programming techniques. Our optimization has a discrete aspect, that is, the connectivity relationships among the elements, as well as a continuous aspect, namely the locations of the elements of interest. Both these aspects are captured by our near-regular structure extraction framework, which alternates between discrete and continuous optimizations. We demonstrate the effectiveness of our framework on a variety of problems including near-regular structure extraction, structure-preserving pattern manipulation, and markerless correspondence detection. Robustness results with respect to geometric and topological noise are presented on synthesized, real-world, and also benchmark datasets. © 2014 ACM.

  4. Association between regular physical exercise and depressive symptoms mediated through social support and resilience in Japanese company workers: a cross-sectional study

    Directory of Open Access Journals (Sweden)

    Eisho Yoshikawa

    2016-07-01

    Full Text Available Abstract Background Regular physical exercise has been reported to reduce depressive symptoms. Several lines of evidence suggest that physical exercise may prevent depression by promoting social support or resilience, which is the ability to adapt to challenging life conditions. The aim of this study was to compare depressive symptoms, social support, and resilience between Japanese company workers who engaged in regular physical exercise and workers who did not exercise regularly. We also investigated whether regular physical exercise has an indirect association with depressive symptoms through social support and resilience. Methods Participants were 715 Japanese employees at six worksites. Depressive symptoms were assessed with the Center for Epidemiologic Studies Depression (CES-D scale, social support with the short version of the Social Support Questionnaire (SSQ, and resilience with the 14-item Resilience Scale (RS-14. A self-report questionnaire, which was extracted from the Japanese version of the Health-Promoting Lifestyle Profile, was used to assess whether participants engage in regular physical exercise, defined as more than 20 min, three or more times per week. The group differences in CES-D, SSQ, and RS-14 scores were investigated by using analysis of covariance (ANCOVA. Mediation analysis was conducted by using Preacher and Hayes’ bootstrap script to assess whether regular physical exercise is associated with depressive symptoms indirectly through resilience and social support. Results The SSQ Number score (F = 4.82, p = 0.03, SSQ Satisfaction score (F = 6.68, p = 0.01, and RS-14 score (F = 6.01, p = 0.01 were significantly higher in the group with regular physical exercise (n = 83 than in the group without regular physical exercise (n = 632 after adjusting for age, education, marital status, and job status. The difference in CES-D score was not significant (F = 2.90, p = 0

  5. Regularized κ-distributions with non-diverging moments

    Science.gov (United States)

    Scherer, K.; Fichtner, H.; Lazar, M.

    2017-12-01

    For various plasma applications the so-called (non-relativistic) κ-distribution is widely used to reproduce and interpret the suprathermal particle populations exhibiting a power-law distribution in velocity or energy. Despite its reputation the standard κ-distribution as a concept is still disputable, mainly due to the velocity moments M l which make a macroscopic characterization possible, but whose existence is restricted only to low orders l definition of the κ-distribution itself is conditioned by the existence of the moment of order l = 2 (i.e., kinetic temperature) satisfied only for κ > 3/2 . In order to resolve these critical limitations we introduce the regularized κ-distribution with non-diverging moments. For the evaluation of all velocity moments a general analytical expression is provided enabling a significant step towards a macroscopic (fluid-like) description of space plasmas, and, in general, any system of κ-distributed particles.

  6. Regular Expression Matching and Operational Semantics

    Directory of Open Access Journals (Sweden)

    Asiri Rathnayake

    2011-08-01

    Full Text Available Many programming languages and tools, ranging from grep to the Java String library, contain regular expression matchers. Rather than first translating a regular expression into a deterministic finite automaton, such implementations typically match the regular expression on the fly. Thus they can be seen as virtual machines interpreting the regular expression much as if it were a program with some non-deterministic constructs such as the Kleene star. We formalize this implementation technique for regular expression matching using operational semantics. Specifically, we derive a series of abstract machines, moving from the abstract definition of matching to increasingly realistic machines. First a continuation is added to the operational semantics to describe what remains to be matched after the current expression. Next, we represent the expression as a data structure using pointers, which enables redundant searches to be eliminated via testing for pointer equality. From there, we arrive both at Thompson's lockstep construction and a machine that performs some operations in parallel, suitable for implementation on a large number of cores, such as a GPU. We formalize the parallel machine using process algebra and report some preliminary experiments with an implementation on a graphics processor using CUDA.

  7. Tetravalent one-regular graphs of order 4p2

    DEFF Research Database (Denmark)

    Feng, Yan-Quan; Kutnar, Klavdija; Marusic, Dragan

    2014-01-01

    A graph is one-regular if its automorphism group acts regularly on the set of its arcs. In this paper tetravalent one-regular graphs of order 4p2, where p is a prime, are classified.......A graph is one-regular if its automorphism group acts regularly on the set of its arcs. In this paper tetravalent one-regular graphs of order 4p2, where p is a prime, are classified....

  8. Regularization and error assignment to unfolded distributions

    CERN Document Server

    Zech, Gunter

    2011-01-01

    The commonly used approach to present unfolded data only in graphical formwith the diagonal error depending on the regularization strength is unsatisfac-tory. It does not permit the adjustment of parameters of theories, the exclusionof theories that are admitted by the observed data and does not allow the com-bination of data from different experiments. We propose fixing the regulariza-tion strength by a p-value criterion, indicating the experimental uncertaintiesindependent of the regularization and publishing the unfolded data in additionwithout regularization. These considerations are illustrated with three differentunfolding and smoothing approaches applied to a toy example.

  9. Joint interval reliability for Markov systems with an application in transmission line reliability

    International Nuclear Information System (INIS)

    Csenki, Attila

    2007-01-01

    We consider Markov reliability models whose finite state space is partitioned into the set of up states U and the set of down states D . Given a collection of k disjoint time intervals I l =[t l ,t l +x l ], l=1,...,k, the joint interval reliability is defined as the probability of the system being in U for all time instances in I 1 union ... union I k . A closed form expression is derived here for the joint interval reliability for this class of models. The result is applied to power transmission lines in a two-state fluctuating environment. We use the Linux versions of the free packages Maxima and Scilab in our implementation for symbolic and numerical work, respectively

  10. Graph Regularized Meta-path Based Transductive Regression in Heterogeneous Information Network.

    Science.gov (United States)

    Wan, Mengting; Ouyang, Yunbo; Kaplan, Lance; Han, Jiawei

    2015-01-01

    A number of real-world networks are heterogeneous information networks, which are composed of different types of nodes and links. Numerical prediction in heterogeneous information networks is a challenging but significant area because network based information for unlabeled objects is usually limited to make precise estimations. In this paper, we consider a graph regularized meta-path based transductive regression model ( Grempt ), which combines the principal philosophies of typical graph-based transductive classification methods and transductive regression models designed for homogeneous networks. The computation of our method is time and space efficient and the precision of our model can be verified by numerical experiments.

  11. On the interplay of basis smoothness and specific range conditions occurring in sparsity regularization

    International Nuclear Information System (INIS)

    Anzengruber, Stephan W; Hofmann, Bernd; Ramlau, Ronny

    2013-01-01

    The convergence rates results in ℓ 1 -regularization when the sparsity assumption is narrowly missed, presented by Burger et al (2013 Inverse Problems 29 025013), are based on a crucial condition which requires that all basis elements belong to the range of the adjoint of the forward operator. Partly it was conjectured that such a condition is very restrictive. In this context, we study sparsity-promoting varieties of Tikhonov regularization for linear ill-posed problems with respect to an orthonormal basis in a separable Hilbert space using ℓ 1 and sublinear penalty terms. In particular, we show that the corresponding range condition is always satisfied for all basis elements if the problems are well-posed in a certain weaker topology and the basis elements are chosen appropriately related to an associated Gelfand triple. The Radon transform, Symm’s integral equation and linear integral operators of Volterra type are examples for such behaviour, which allows us to apply convergence rates results for non-sparse solutions, and we further extend these results also to the case of non-convex ℓ q -regularization with 0 < q < 1. (paper)

  12. Interpregnancy intervals: impact of postpartum contraceptive effectiveness and coverage.

    Science.gov (United States)

    Thiel de Bocanegra, Heike; Chang, Richard; Howell, Mike; Darney, Philip

    2014-04-01

    The purpose of this study was to determine the use of contraceptive methods, which was defined by effectiveness, length of coverage, and their association with short interpregnancy intervals, when controlling for provider type and client demographics. We identified a cohort of 117,644 women from the 2008 California Birth Statistical Master file with second or higher order birth and at least 1 Medicaid (Family Planning, Access, Care, and Treatment [Family PACT] program or Medi-Cal) claim within 18 months after index birth. We explored the effect of contraceptive method provision on the odds of having an optimal interpregnancy interval and controlled for covariates. The average length of contraceptive coverage was 3.81 months (SD = 4.84). Most women received user-dependent hormonal contraceptives as their most effective contraceptive method (55%; n = 65,103 women) and one-third (33%; n = 39,090 women) had no contraceptive claim. Women who used long-acting reversible contraceptive methods had 3.89 times the odds and women who used user-dependent hormonal methods had 1.89 times the odds of achieving an optimal birth interval compared with women who used barrier methods only; women with no method had 0.66 times the odds. When user-dependent methods are considered, the odds of having an optimal birth interval increased for each additional month of contraceptive coverage by 8% (odds ratio, 1.08; 95% confidence interval, 1.08-1.09). Women who were seen by Family PACT or by both Family PACT and Medi-Cal providers had significantly higher odds of optimal birth intervals compared with women who were served by Medi-Cal only. To achieve optimal birth spacing and ultimately to improve birth outcomes, attention should be given to contraceptive counseling and access to contraceptive methods in the postpartum period. Copyright © 2014 Mosby, Inc. All rights reserved.

  13. Higher order total variation regularization for EIT reconstruction.

    Science.gov (United States)

    Gong, Bo; Schullcke, Benjamin; Krueger-Ziolek, Sabine; Zhang, Fan; Mueller-Lisse, Ullrich; Moeller, Knut

    2018-01-08

    Electrical impedance tomography (EIT) attempts to reveal the conductivity distribution of a domain based on the electrical boundary condition. This is an ill-posed inverse problem; its solution is very unstable. Total variation (TV) regularization is one of the techniques commonly employed to stabilize reconstructions. However, it is well known that TV regularization induces staircase effects, which are not realistic in clinical applications. To reduce such artifacts, modified TV regularization terms considering a higher order differential operator were developed in several previous studies. One of them is called total generalized variation (TGV) regularization. TGV regularization has been successively applied in image processing in a regular grid context. In this study, we adapted TGV regularization to the finite element model (FEM) framework for EIT reconstruction. Reconstructions using simulation and clinical data were performed. First results indicate that, in comparison to TV regularization, TGV regularization promotes more realistic images. Graphical abstract Reconstructed conductivity changes located on selected vertical lines. For each of the reconstructed images as well as the ground truth image, conductivity changes located along the selected left and right vertical lines are plotted. In these plots, the notation GT in the legend stands for ground truth, TV stands for total variation method, and TGV stands for total generalized variation method. Reconstructed conductivity distributions from the GREIT algorithm are also demonstrated.

  14. Application of Turchin's method of statistical regularization

    Science.gov (United States)

    Zelenyi, Mikhail; Poliakova, Mariia; Nozik, Alexander; Khudyakov, Alexey

    2018-04-01

    During analysis of experimental data, one usually needs to restore a signal after it has been convoluted with some kind of apparatus function. According to Hadamard's definition this problem is ill-posed and requires regularization to provide sensible results. In this article we describe an implementation of the Turchin's method of statistical regularization based on the Bayesian approach to the regularization strategy.

  15. Regularization based on steering parameterized Gaussian filters and a Bhattacharyya distance functional

    Science.gov (United States)

    Lopes, Emerson P.

    2001-08-01

    Template regularization embeds the problem of class separability. In the machine vision perspective, this problem is critical when a textural classification procedure is applied to non-stationary pattern mosaic images. These applications often present low accuracy performance due to disturbance of the classifiers produced by exogenous or endogenous signal regularity perturbations. Natural scene imaging, where the images present certain degree of homogeneity in terms of texture element size or shape (primitives) shows a variety of behaviors, especially varying the preferential spatial directionality. The space-time image pattern characterization is only solved if classification procedures are designed considering the most robust tools within a parallel and hardware perspective. The results to be compared in this paper are obtained using a framework based on multi-resolution, frame and hypothesis approach. Two strategies for the bank of Gabor filters applications are considered: adaptive strategy using the KL transform and fix configuration strategy. The regularization under discussion is accomplished in the pyramid building system instance. The filterings are steering Gaussians controlled by free parameters which are adjusted in accordance with a feedback process driven by hints obtained from sequence of frames interaction functionals pos-processed in the training process and including classification of training set samples as examples. Besides these adjustments there is continuous input data sensitive adaptiveness. The experimental result assessments are focused on two basic issues: Bhattacharyya distance as pattern characterization feature and the combination of KL transform as feature selection and adaptive criterion with the regularization of the pattern Bhattacharyya distance functional (BDF) behavior, using the BDF state separability and symmetry as the main indicators of an optimum framework parameter configuration.

  16. On the regularized fermionic projector of the vacuum

    Science.gov (United States)

    Finster, Felix

    2008-03-01

    We construct families of fermionic projectors with spherically symmetric regularization, which satisfy the condition of a distributional MP-product. The method is to analyze regularization tails with a power law or logarithmic scaling in composite expressions in the fermionic projector. The resulting regularizations break the Lorentz symmetry and give rise to a multilayer structure of the fermionic projector near the light cone. Furthermore, we construct regularizations which go beyond the distributional MP-product in that they yield additional distributional contributions supported at the origin. The remaining freedom for the regularization parameters and the consequences for the normalization of the fermionic states are discussed.

  17. On the regularized fermionic projector of the vacuum

    International Nuclear Information System (INIS)

    Finster, Felix

    2008-01-01

    We construct families of fermionic projectors with spherically symmetric regularization, which satisfy the condition of a distributional MP-product. The method is to analyze regularization tails with a power law or logarithmic scaling in composite expressions in the fermionic projector. The resulting regularizations break the Lorentz symmetry and give rise to a multilayer structure of the fermionic projector near the light cone. Furthermore, we construct regularizations which go beyond the distributional MP-product in that they yield additional distributional contributions supported at the origin. The remaining freedom for the regularization parameters and the consequences for the normalization of the fermionic states are discussed

  18. Quantum effects in non-maximally symmetric spaces

    International Nuclear Information System (INIS)

    Shen, T.C.

    1985-01-01

    Non-Maximally symmetric spaces provide a more general background to explore the relation between the geometry of the manifold and the quantum fields defined in the manifold than those with maximally symmetric spaces. A static Taub universe is used to study the effect of curvature anisotropy on the spontaneous symmetry breaking of a self-interacting scalar field. The one-loop effective potential on a λphi 4 field with arbitrary coupling xi is computed by zeta function regularization. For massless minimal coupled scalar fields, first order phase transitions can occur. Keeping the shape invariant but decreasing the curvature radius of the universe induces symmetry breaking. If the curvature radius is held constant, increasing deformation can restore the symmetry. Studies on the higher-dimensional Kaluza-Klein theories are also focused on the deformation effect. Using the dimensional regularization, the effective potential of the free scalar fields in M 4 x T/sup N/ and M 4 x (Taub) 3 spaces are obtained. The stability criterions for the static solutions of the self-consistent Einstein equations are derived. Stable solutions of the M 4 x S/sup N/ topology do not exist. With the Taub space as the internal space, the gauge coupling constants of SU(2), and U(1) can be determined geometrically. The weak angle is therefore predicted by geometry in this model

  19. Reversibility and the structure of the local state space

    International Nuclear Information System (INIS)

    Al-Safi, Sabri W; Richens, Jonathan

    2015-01-01

    The richness of quantum theory’s reversible dynamics is one of its unique operational characteristics, with recent results suggesting deep links between the theory’s reversible dynamics, its local state space and the degree of non-locality it permits. We explore the delicate interplay between these features, demonstrating that reversibility places strong constraints on both the local and global state space. Firstly, we show that all reversible dynamics are trivial (composed of local transformations and permutations of subsytems) in maximally non-local theories whose local state spaces satisfy a dichotomy criterion; this applies to a range of operational models that have previously been studied, such as d-dimensional ‘hyperballs’ and almost all regular polytope systems. By separately deriving a similar result for odd-sided polygons, we show that classical systems are the only regular polytope state spaces whose maximally non-local composites allow for non-trivial reversible dynamics. Secondly, we show that non-trivial reversible dynamics do exist in maximally non-local theories whose state spaces are reducible into two or more smaller spaces. We conjecture that this is a necessary condition for the existence of such dynamics, but that reversible entanglement generation remains impossible even in this scenario. (paper)

  20. Regularization modeling for large-eddy simulation

    NARCIS (Netherlands)

    Geurts, Bernardus J.; Holm, D.D.

    2003-01-01

    A new modeling approach for large-eddy simulation (LES) is obtained by combining a "regularization principle" with an explicit filter and its inversion. This regularization approach allows a systematic derivation of the implied subgrid model, which resolves the closure problem. The central role of

  1. Spatially-Variant Tikhonov Regularization for Double-Difference Waveform Inversion

    Energy Technology Data Exchange (ETDEWEB)

    Lin, Youzuo [Los Alamos National Laboratory; Huang, Lianjie [Los Alamos National Laboratory; Zhang, Zhigang [Los Alamos National Laboratory

    2011-01-01

    Double-difference waveform inversion is a potential tool for quantitative monitoring for geologic carbon storage. It jointly inverts time-lapse seismic data for changes in reservoir geophysical properties. Due to the ill-posedness of waveform inversion, it is a great challenge to obtain reservoir changes accurately and efficiently, particularly when using time-lapse seismic reflection data. Regularization techniques can be utilized to address the issue of ill-posedness. The regularization parameter controls the smoothness of inversion results. A constant regularization parameter is normally used in waveform inversion, and an optimal regularization parameter has to be selected. The resulting inversion results are a trade off among regions with different smoothness or noise levels; therefore the images are either over regularized in some regions while under regularized in the others. In this paper, we employ a spatially-variant parameter in the Tikhonov regularization scheme used in double-difference waveform tomography to improve the inversion accuracy and robustness. We compare the results obtained using a spatially-variant parameter with those obtained using a constant regularization parameter and those produced without any regularization. We observe that, utilizing a spatially-variant regularization scheme, the target regions are well reconstructed while the noise is reduced in the other regions. We show that the spatially-variant regularization scheme provides the flexibility to regularize local regions based on the a priori information without increasing computational costs and the computer memory requirement.

  2. Multiplicity distributions in small phase-space domains in central nucleus-nucleus collisions

    International Nuclear Information System (INIS)

    Baechler, J.; Hoffmann, M.; Runge, K.; Schmoetten, E.; Bartke, J.; Gladysz, E.; Kowalski, M.; Stefanski, P.; Bialkowska, H.; Bock, R.; Brockmann, R.; Sandoval, A.; Buncic, P.; Ferenc, D.; Kadija, K.; Ljubicic, A. Jr.; Vranic, D.; Chase, S.I.; Harris, J.W.; Odyniec, G.; Pugh, H.G.; Rai, G.; Teitelbaum, L.; Tonse, S.; Derado, I.; Eckardt, V.; Gebauer, H.J.; Rauch, W.; Schmitz, N.; Seyboth, P.; Seyerlein, J.; Vesztergombi, G.; Eschke, J.; Heck, W.; Kabana, S.; Kuehmichel, A.; Lahanas, M.; Lee, Y.; Le Vine, M.; Margetis, S.; Renfordt, R.; Roehrich, D.; Rothard, H.; Schmidt, E.; Schneider, I.; Stock, R.; Stroebele, H.; Wenig, S.; Fleischmann, B.; Fuchs, M.; Gazdzicki, M.; Kosiec, J.; Skrzypczak, E.; Keidel, R.; Piper, A.; Puehlhofer, F.; Nappi, E.; Posa, F.; Paic, G.; Panagiotou, A.D.; Petridis, A.; Vassileiadis, G.; Pfenning, J.; Wosiek, B.

    1992-10-01

    Multiplicity distributions of negatively charged particles have been studied in restricted phase space intervals for central S + S, O + Au and S + Au collisions at 200 GeV/nucleon. It is shown that multiplicity distributions are well described by a negative binomial form irrespectively of the size and dimensionality of phase space domain. A clan structure analysis reveals interesting similarities between complex nuclear collisions and a simple partonic shower. The lognormal distribution agrees reasonably well with the multiplicity data in large domains, but fails in the case of small intervals. No universal scaling function was found to describe the shape of multiplicity distributions in phase space intervals of varying size. (orig.)

  3. Espaçamento de plantio e intervalos de colheita na biomassa e no óleo essencial de gerânio Plant spacing and harvest intervals on biomass and essential oil of geranium

    Directory of Open Access Journals (Sweden)

    Arie F Blank

    2012-12-01

    Full Text Available O gerânio (Pelargonium graveolens é uma erva aromática e seu óleo essencial é comumente utilizado na constituição de fármacos e cosméticos no mundo. A planta adapta-se bem às condições climáticas do nordeste brasileiro, mas são escassos os dados sobre seu rendimento quantitativo e qualitativo. O objetivo do trabalho foi avaliar a influência de espaçamento de plantas e intervalos de colheita em gerânio na produção de biomassa e de óleo essencial. Foram testados três espaçamentos de plantio (50x50, 50x60 e 50x80 cm e três intervalos de colheita (8, 12 e 16 semanas. Os maiores valores totais de massa fresca e seca de folhas e caules (2679,04 g m-2; 424,62 g m-2; 1035,08 g m-2; 136,85 g m-2, respectivamente e rendimento de óleo essencial (7,56 mL m-2, que são características de interesse direto para o mercado, foram obtidos em intervalo de colheita de oito semanas e no espaçamento de 50x50 cm. Nas demais variáveis analisadas, houve pouca diferenciação entre os tratamentos, contudo, períodos longos de colheita são menos produtivos, pois houve queda nos valores de todas as variáveis nas ultimas colheitas de cada intervalo. Assim, melhores resultados de quantidade e qualidade para o gerânio foram obtidos com intervalo de 8 semanas e no espaçamento 50x50 cm.Geranium (Pelargonium graveolens is an aromatic herb and its essential oil is commonly used in the creation of drugs and cosmetics worldwide. The plant is well adapted to the climatic conditions of the Brazilian northeast, but there are few data on its quantitative and qualitative yield. The aim of this study was to evaluate the influence of plant spacing and harvest intervals of geranium on the production of biomass and essential oil. Three different plant spacings (50x50, 50x60 and 50x80 cm and three harvest intervals (8, 12 and 16 weeks were tested. The highest values of total fresh and dry weight of leaves and stems (2679.04 g m-2; 424.62 g m-2; 1035.08 g m-2

  4. From recreational to regular drug use

    DEFF Research Database (Denmark)

    Järvinen, Margaretha; Ravn, Signe

    2011-01-01

    This article analyses the process of going from recreational use to regular and problematic use of illegal drugs. We present a model containing six career contingencies relevant for young people’s progress from recreational to regular drug use: the closing of social networks, changes in forms...

  5. Online evolution reconstruction from a single measurement record with random time intervals for quantum communication

    Science.gov (United States)

    Zhou, Hua; Su, Yang; Wang, Rong; Zhu, Yong; Shen, Huiping; Pu, Tao; Wu, Chuanxin; Zhao, Jiyong; Zhang, Baofu; Xu, Zhiyong

    2017-10-01

    Online reconstruction of a time-variant quantum state from the encoding/decoding results of quantum communication is addressed by developing a method of evolution reconstruction from a single measurement record with random time intervals. A time-variant two-dimensional state is reconstructed on the basis of recovering its expectation value functions of three nonorthogonal projectors from a random single measurement record, which is composed from the discarded qubits of the six-state protocol. The simulated results prove that our method is robust to typical metro quantum channels. Our work extends the Fourier-based method of evolution reconstruction from the version for a regular single measurement record with equal time intervals to a unified one, which can be applied to arbitrary single measurement records. The proposed protocol of evolution reconstruction runs concurrently with the one of quantum communication, which can facilitate the online quantum tomography.

  6. A wavelet-based regularized reconstruction algorithm for SENSE parallel MRI with applications to neuroimaging

    International Nuclear Information System (INIS)

    Chaari, L.; Pesquet, J.Ch.; Chaari, L.; Ciuciu, Ph.; Benazza-Benyahia, A.

    2011-01-01

    To reduce scanning time and/or improve spatial/temporal resolution in some Magnetic Resonance Imaging (MRI) applications, parallel MRI acquisition techniques with multiple coils acquisition have emerged since the early 1990's as powerful imaging methods that allow a faster acquisition process. In these techniques, the full FOV image has to be reconstructed from the resulting acquired under sampled k-space data. To this end, several reconstruction techniques have been proposed such as the widely-used Sensitivity Encoding (SENSE) method. However, the reconstructed image generally presents artifacts when perturbations occur in both the measured data and the estimated coil sensitivity profiles. In this paper, we aim at achieving accurate image reconstruction under degraded experimental conditions (low magnetic field and high reduction factor), in which neither the SENSE method nor the Tikhonov regularization in the image domain give convincing results. To this end, we present a novel method for SENSE-based reconstruction which proceeds with regularization in the complex wavelet domain by promoting sparsity. The proposed approach relies on a fast algorithm that enables the minimization of regularized non-differentiable criteria including more general penalties than a classical l 1 term. To further enhance the reconstructed image quality, local convex constraints are added to the regularization process. In vivo human brain experiments carried out on Gradient-Echo (GRE) anatomical and Echo Planar Imaging (EPI) functional MRI data at 1.5 T indicate that our algorithm provides reconstructed images with reduced artifacts for high reduction factors. (authors)

  7. Chosen interval methods for solving linear interval systems with special type of matrix

    Science.gov (United States)

    Szyszka, Barbara

    2013-10-01

    The paper is devoted to chosen direct interval methods for solving linear interval systems with special type of matrix. This kind of matrix: band matrix with a parameter, from finite difference problem is obtained. Such linear systems occur while solving one dimensional wave equation (Partial Differential Equations of hyperbolic type) by using the central difference interval method of the second order. Interval methods are constructed so as the errors of method are enclosed in obtained results, therefore presented linear interval systems contain elements that determining the errors of difference method. The chosen direct algorithms have been applied for solving linear systems because they have no errors of method. All calculations were performed in floating-point interval arithmetic.

  8. Regular variation on measure chains

    Czech Academy of Sciences Publication Activity Database

    Řehák, Pavel; Vitovec, J.

    2010-01-01

    Roč. 72, č. 1 (2010), s. 439-448 ISSN 0362-546X R&D Projects: GA AV ČR KJB100190701 Institutional research plan: CEZ:AV0Z10190503 Keywords : regularly varying function * regularly varying sequence * measure chain * time scale * embedding theorem * representation theorem * second order dynamic equation * asymptotic properties Subject RIV: BA - General Mathematics Impact factor: 1.279, year: 2010 http://www.sciencedirect.com/science/article/pii/S0362546X09008475

  9. New regular black hole solutions

    International Nuclear Information System (INIS)

    Lemos, Jose P. S.; Zanchin, Vilson T.

    2011-01-01

    In the present work we consider general relativity coupled to Maxwell's electromagnetism and charged matter. Under the assumption of spherical symmetry, there is a particular class of solutions that correspond to regular charged black holes whose interior region is de Sitter, the exterior region is Reissner-Nordstroem and there is a charged thin-layer in-between the two. The main physical and geometrical properties of such charged regular black holes are analyzed.

  10. On geodesics in low regularity

    Science.gov (United States)

    Sämann, Clemens; Steinbauer, Roland

    2018-02-01

    We consider geodesics in both Riemannian and Lorentzian manifolds with metrics of low regularity. We discuss existence of extremal curves for continuous metrics and present several old and new examples that highlight their subtle interrelation with solutions of the geodesic equations. Then we turn to the initial value problem for geodesics for locally Lipschitz continuous metrics and generalize recent results on existence, regularity and uniqueness of solutions in the sense of Filippov.

  11. Recall intervals and time used for examination and prevention by dentists in child dental care in Denmark, Iceland, Norway and Sweden in 1996 and 2014

    DEFF Research Database (Denmark)

    Wang, N J; Petersen, P E; Sveinsdóttir, E G

    2018-01-01

    OBJECTIVE: The purpose of the present study was to explore intervals between regular dental examination and the time dentists spent for examination and preventive dental care of children in 1996 and 2014. PARTICIPANTS AND METHODS: In Denmark, Norway and Sweden, random samples of dentists working...... examinations in three of the four countries in 2014 than in 1996. CONCLUSIONS: This study of trends in dental care delivered by dentists during recent decades showed moves towards extended recall intervals and preventive care individualized according to caries risk. In addition, extending intervals could...... dentists used ample time delivering preventive care to children. Dentists reported spending significantly more time providing preventive care for caries risk children than for other children both in 1996 and 2014. Concurrent with extended intervals, dentists reported spending longer performing routine...

  12. Joint interval reliability for Markov systems with an application in transmission line reliability

    Energy Technology Data Exchange (ETDEWEB)

    Csenki, Attila [School of Computing and Mathematics, University of Bradford, Bradford, West Yorkshire, BD7 1DP (United Kingdom)]. E-mail: a.csenki@bradford.ac.uk

    2007-06-15

    We consider Markov reliability models whose finite state space is partitioned into the set of up states {sub U} and the set of down states {sub D}. Given a collection of k disjoint time intervals I{sub l}=[t{sub l},t{sub l}+x{sub l}], l=1,...,k, the joint interval reliability is defined as the probability of the system being in {sub U} for all time instances in I{sub 1} union ... union I{sub k}. A closed form expression is derived here for the joint interval reliability for this class of models. The result is applied to power transmission lines in a two-state fluctuating environment. We use the Linux versions of the free packages Maxima and Scilab in our implementation for symbolic and numerical work, respectively.

  13. Supersymmetric black holes with lens-space topology.

    Science.gov (United States)

    Kunduri, Hari K; Lucietti, James

    2014-11-21

    We present a new supersymmetric, asymptotically flat, black hole solution to five-dimensional supergravity. It is regular on and outside an event horizon of lens-space topology L(2,1). It is the first example of an asymptotically flat black hole with lens-space topology. The solution is characterized by a charge, two angular momenta, and a magnetic flux through a noncontractible disk region ending on the horizon, with one constraint relating these.

  14. Laplacian manifold regularization method for fluorescence molecular tomography

    Science.gov (United States)

    He, Xuelei; Wang, Xiaodong; Yi, Huangjian; Chen, Yanrong; Zhang, Xu; Yu, Jingjing; He, Xiaowei

    2017-04-01

    Sparse regularization methods have been widely used in fluorescence molecular tomography (FMT) for stable three-dimensional reconstruction. Generally, ℓ1-regularization-based methods allow for utilizing the sparsity nature of the target distribution. However, in addition to sparsity, the spatial structure information should be exploited as well. A joint ℓ1 and Laplacian manifold regularization model is proposed to improve the reconstruction performance, and two algorithms (with and without Barzilai-Borwein strategy) are presented to solve the regularization model. Numerical studies and in vivo experiment demonstrate that the proposed Gradient projection-resolved Laplacian manifold regularization method for the joint model performed better than the comparative algorithm for ℓ1 minimization method in both spatial aggregation and location accuracy.

  15. Learning Sparse Visual Representations with Leaky Capped Norm Regularizers

    OpenAIRE

    Wangni, Jianqiao; Lin, Dahua

    2017-01-01

    Sparsity inducing regularization is an important part for learning over-complete visual representations. Despite the popularity of $\\ell_1$ regularization, in this paper, we investigate the usage of non-convex regularizations in this problem. Our contribution consists of three parts. First, we propose the leaky capped norm regularization (LCNR), which allows model weights below a certain threshold to be regularized more strongly as opposed to those above, therefore imposes strong sparsity and...

  16. Adaptive regularization of noisy linear inverse problems

    DEFF Research Database (Denmark)

    Hansen, Lars Kai; Madsen, Kristoffer Hougaard; Lehn-Schiøler, Tue

    2006-01-01

    In the Bayesian modeling framework there is a close relation between regularization and the prior distribution over parameters. For prior distributions in the exponential family, we show that the optimal hyper-parameter, i.e., the optimal strength of regularization, satisfies a simple relation: T......: The expectation of the regularization function, i.e., takes the same value in the posterior and prior distribution. We present three examples: two simulations, and application in fMRI neuroimaging....

  17. Multiple Kernel Learning for adaptive graph regularized nonnegative matrix factorization

    KAUST Repository

    Wang, Jim Jing-Yan; AbdulJabbar, Mustafa Abdulmajeed

    2012-01-01

    Nonnegative Matrix Factorization (NMF) has been continuously evolving in several areas like pattern recognition and information retrieval methods. It factorizes a matrix into a product of 2 low-rank non-negative matrices that will define parts-based, and linear representation of non-negative data. Recently, Graph regularized NMF (GrNMF) is proposed to find a compact representation, which uncovers the hidden semantics and simultaneously respects the intrinsic geometric structure. In GNMF, an affinity graph is constructed from the original data space to encode the geometrical information. In this paper, we propose a novel idea which engages a Multiple Kernel Learning approach into refining the graph structure that reflects the factorization of the matrix and the new data space. The GrNMF is improved by utilizing the graph refined by the kernel learning, and then a novel kernel learning method is introduced under the GrNMF framework. Our approach shows encouraging results of the proposed algorithm in comparison to the state-of-the-art clustering algorithms like NMF, GrNMF, SVD etc.

  18. A Weak S-Implication on Interval Sets%区间集上的弱S-蕴涵

    Institute of Scientific and Technical Information of China (English)

    薛占熬; 程惠茹; 黄海松; 李跃军

    2013-01-01

      区间集是表达模糊信息的一种有效方法,蕴涵算子的研究是逻辑研究的关键,在区间集上构造蕴涵具有重要的研究价值。在区间集上构造了一个新的区间集弱S-蕴涵,证明了其正则性、单调性以及其他一些重要性质。最后证明了由它可构造剩余格,这为区间集的逻辑系统的建立提供了理论基础。%The theory of interval sets provides an effective method for expressing fuzzy information. Researching implication operators is a key step for fuzzy logic system. Thus it is very important to construct implication operators on interval sets. This paper constructs a new implication operator of interval sets, named as weak S-implication. Meanwhile, it proves the regularity, monotonicity and other important properties of weak S-implication. Finally, it proves that the residuated lattice is constituted under the S-implication operator on interval sets. This research provides a theoretical foundation for the logic system of interval sets.

  19. Exclusion of children with intellectual disabilities from regular ...

    African Journals Online (AJOL)

    Study investigated why teachers exclude children with intellectual disability from the regular classrooms in Nigeria. Participants were, 169 regular teachers randomly selected from Oyo and Ogun states. Questionnaire was used to collect data result revealed that 57.4% regular teachers could not cope with children with ID ...

  20. Continuous-time interval model identification of blood glucose dynamics for type 1 diabetes

    Science.gov (United States)

    Kirchsteiger, Harald; Johansson, Rolf; Renard, Eric; del Re, Luigi

    2014-07-01

    While good physiological models of the glucose metabolism in type 1 diabetic patients are well known, their parameterisation is difficult. The high intra-patient variability observed is a further major obstacle. This holds for data-based models too, so that no good patient-specific models are available. Against this background, this paper proposes the use of interval models to cover the different metabolic conditions. The control-oriented models contain a carbohydrate and insulin sensitivity factor to be used for insulin bolus calculators directly. Available clinical measurements were sampled on an irregular schedule which prompts the use of continuous-time identification, also for the direct estimation of the clinically interpretable factors mentioned above. An identification method is derived and applied to real data from 28 diabetic patients. Model estimation was done on a clinical data-set, whereas validation results shown were done on an out-of-clinic, everyday life data-set. The results show that the interval model approach allows a much more regular estimation of the parameters and avoids physiologically incompatible parameter estimates.

  1. On infinite regular and chiral maps

    OpenAIRE

    Arredondo, John A.; Valdez, Camilo Ramírez y Ferrán

    2015-01-01

    We prove that infinite regular and chiral maps take place on surfaces with at most one end. Moreover, we prove that an infinite regular or chiral map on an orientable surface with genus can only be realized on the Loch Ness monster, that is, the topological surface of infinite genus with one end.

  2. 29 CFR 779.18 - Regular rate.

    Science.gov (United States)

    2010-07-01

    ... employee under subsection (a) or in excess of the employee's normal working hours or regular working hours... Relating to Labor (Continued) WAGE AND HOUR DIVISION, DEPARTMENT OF LABOR STATEMENTS OF GENERAL POLICY OR... not less than one and one-half times their regular rates of pay. Section 7(e) of the Act defines...

  3. Continuum regularized Yang-Mills theory

    International Nuclear Information System (INIS)

    Sadun, L.A.

    1987-01-01

    Using the machinery of stochastic quantization, Z. Bern, M. B. Halpern, C. Taubes and I recently proposed a continuum regularization technique for quantum field theory. This regularization may be implemented by applying a regulator to either the (d + 1)-dimensional Parisi-Wu Langevin equation or, equivalently, to the d-dimensional second order Schwinger-Dyson (SD) equations. This technique is non-perturbative, respects all gauge and Lorentz symmetries, and is consistent with a ghost-free gauge fixing (Zwanziger's). This thesis is a detailed study of this regulator, and of regularized Yang-Mills theory, using both perturbative and non-perturbative techniques. The perturbative analysis comes first. The mechanism of stochastic quantization is reviewed, and a perturbative expansion based on second-order SD equations is developed. A diagrammatic method (SD diagrams) for evaluating terms of this expansion is developed. We apply the continuum regulator to a scalar field theory. Using SD diagrams, we show that all Green functions can be rendered finite to all orders in perturbation theory. Even non-renormalizable theories can be regularized. The continuum regulator is then applied to Yang-Mills theory, in conjunction with Zwanziger's gauge fixing. A perturbative expansion of the regulator is incorporated into the diagrammatic method. It is hoped that the techniques discussed in this thesis will contribute to the construction of a renormalized Yang-Mills theory is 3 and 4 dimensions

  4. Critical spaces for quasilinear parabolic evolution equations and applications

    Science.gov (United States)

    Prüss, Jan; Simonett, Gieri; Wilke, Mathias

    2018-02-01

    We present a comprehensive theory of critical spaces for the broad class of quasilinear parabolic evolution equations. The approach is based on maximal Lp-regularity in time-weighted function spaces. It is shown that our notion of critical spaces coincides with the concept of scaling invariant spaces in case that the underlying partial differential equation enjoys a scaling invariance. Applications to the vorticity equations for the Navier-Stokes problem, convection-diffusion equations, the Nernst-Planck-Poisson equations in electro-chemistry, chemotaxis equations, the MHD equations, and some other well-known parabolic equations are given.

  5. Regularity for 3D Navier-Stokes equations in terms of two components of the vorticity

    Directory of Open Access Journals (Sweden)

    Sadek Gala

    2010-10-01

    Full Text Available We establish regularity conditions for the 3D Navier-Stokes equation via two components of the vorticity vector. It is known that if a Leray-Hopf weak solution $u$ satisfies $$ ilde{omega}in L^{2/(2-r}(0,T;L^{3/r}(mathbb{R}^3quad hbox{with }0regularity of Leray-Hopf weak solution $u$ under each of the following two (weaker conditions: $$displaylines{ ilde{omega}in L^{2/(2-r}(0,T;dot {mathcal{M}}_{2, 3/r}(mathbb{R}^3quad hbox{for }0space. Since $L^{3/r}(mathbb{R}^3$ is a proper subspace of $dot {mathcal{M}}_{2,3/r}(mathbb{R}^3$, our regularity criterion improves the results in Chae-Choe [5].

  6. The effects of thioridazine on the QTc interval - cardiovascular safety in a South African setting

    Directory of Open Access Journals (Sweden)

    Cathlene Seller

    2005-09-01

    Full Text Available Background. Thioridazine has long been used as a first-line antipsychotic in South Africa without any apparent problems. Recently the American Food and Drug Administration (FDA and Novartis have warned of potentially lethal arrhythmias that may result from the use of thioridazine. Abnormal QT-interval prolongation on the electrocardiogram (ECG seems to be the most reliable indicator of risk of arrhythmias, such as torsade de pointes and ventricular fibrillation. Objective. The purpose of this study was to determine whether these warnings are of clinical relevance in a setting where there are already a limited number of antipsychotics available. Method. Thirty hospitalised subjects who required switching from a high-potency to a low-potency antipsychotic were included. All subjects were commenced on thioridazine 300 mg per day and had an ECG 1 week after initiation and 48 hours after each dose adjustment. QTc was determined using Bazett’s formula. Results. Thioridazine induced a significant increase (p = 0.0001 in QTc interval from baseline values of 400.6 (± 27.3 milliseconds to 429.1 (± 44.2 milliseconds. The QTc interval increased to above 450 milliseconds in 7 subjects (23% and thioridazine was discontinued in 2 subjects because of a QTc interval greater or equal to 500 milliseconds. Conclusion. Thioridazine caused a significant, although asymptomatic, increase in QTc interval in almost one-quarter of subjects who received the medication as second-line treatment. Thioridazine should no longer be used as a first-line treatment and if used it should be accompanied by regular ECG monitoring.

  7. Optimization of the Reconstruction Interval in Neurovascular 4D-CTA Imaging

    Science.gov (United States)

    Hoogenboom, T.C.H.; van Beurden, R.M.J.; van Teylingen, B.; Schenk, B.; Willems, P.W.A.

    2012-01-01

    Summary Time resolved whole brain CT angiography (4D-CTA) is a novel imaging technology providing information regarding blood flow. One of the factors that influence the diagnostic value of this examination is the temporal resolution, which is affected by the gantry rotation speed during acquisition and the reconstruction interval during post-processing. Post-processing determines the time spacing between two reconstructed volumes and, unlike rotation speed, does not affect radiation burden. The data sets of six patients who underwent a cranial 4D-CTA were used for this study. Raw data was acquired using a 320-slice scanner with a rotation speed of 2 Hz. The arterial to venous passage of an intravenous contrast bolus was captured during a 15 s continuous scan. The raw data was reconstructed using four different reconstruction-intervals: 0.2, 0.3, 0.5 and 1.0 s. The results were rated by two observers using a standardized score sheet. The appearance of each lesion was rated correctly in all readings. Scoring for quality of temporal resolution revealed a stepwise improvement from the 1.0 s interval to the 0.3 s interval, while no discernable improvement was noted between the 0.3 s and 0.2 s interval. An increase in temporal resolution may improve the diagnostic quality of cranial 4D-CTA. Using a rotation speed of 0.5 s, the optimal reconstruction interval appears to be 0.3 s, beyond which, changes can no longer be discerned. PMID:23217631

  8. Periodicity in the autocorrelation function as a mechanism for regularly occurring zero crossings or extreme values of a Gaussian process.

    Science.gov (United States)

    Wilson, Lorna R M; Hopcraft, Keith I

    2017-12-01

    The problem of zero crossings is of great historical prevalence and promises extensive application. The challenge is to establish precisely how the autocorrelation function or power spectrum of a one-dimensional continuous random process determines the density function of the intervals between the zero crossings of that process. This paper investigates the case where periodicities are incorporated into the autocorrelation function of a smooth process. Numerical simulations, and statistics about the number of crossings in a fixed interval, reveal that in this case the zero crossings segue between a random and deterministic point process depending on the relative time scales of the periodic and nonperiodic components of the autocorrelation function. By considering the Laplace transform of the density function, we show that incorporating correlation between successive intervals is essential to obtaining accurate results for the interval variance. The same method enables prediction of the density function tail in some regions, and we suggest approaches for extending this to cover all regions. In an ever-more complex world, the potential applications for this scale of regularity in a random process are far reaching and powerful.

  9. Periodicity in the autocorrelation function as a mechanism for regularly occurring zero crossings or extreme values of a Gaussian process

    Science.gov (United States)

    Wilson, Lorna R. M.; Hopcraft, Keith I.

    2017-12-01

    The problem of zero crossings is of great historical prevalence and promises extensive application. The challenge is to establish precisely how the autocorrelation function or power spectrum of a one-dimensional continuous random process determines the density function of the intervals between the zero crossings of that process. This paper investigates the case where periodicities are incorporated into the autocorrelation function of a smooth process. Numerical simulations, and statistics about the number of crossings in a fixed interval, reveal that in this case the zero crossings segue between a random and deterministic point process depending on the relative time scales of the periodic and nonperiodic components of the autocorrelation function. By considering the Laplace transform of the density function, we show that incorporating correlation between successive intervals is essential to obtaining accurate results for the interval variance. The same method enables prediction of the density function tail in some regions, and we suggest approaches for extending this to cover all regions. In an ever-more complex world, the potential applications for this scale of regularity in a random process are far reaching and powerful.

  10. Proximity Queries between Interval-Based CSG Octrees

    International Nuclear Information System (INIS)

    Dyllong, Eva; Grimm, Cornelius

    2007-01-01

    This short paper is concerned with a new algorithm for collision and distance calculation between CSG octrees, a generalization of an octree model created from a Constructive Solid Geometry (CSG) object. The data structure uses interval arithmetic and allows us to extend the tests for classifying points in space as inside, on the boundary, or outside a CSG object to entire sections of the space at once. Tree nodes with additional information about relevant parts of the CSG object are introduced in order to reduce the depth of the required subdivision. The new data structure reduces the input complexity and enables us to reconstruct the CSG object. We present an efficient algorithm for computing the distance between CSG objects encoded by the new data structure. The distance algorithm is based on a distance algorithm for classical octrees but, additionally, it utilizes an elaborated sort sequence and differentiated handling of pairs of octree nodes to enhance its efficiency. Experimental results indicate that, in comparison to common octrees, the new representation has advantages in the field of proximity query

  11. Regularity effect in prospective memory during aging

    OpenAIRE

    Blondelle, Geoffrey; Hainselin, Mathieu; Gounden, Yannick; Heurley, Laurent; Voisin, Hélène; Megalakaki, Olga; Bressous, Estelle; Quaglino, Véronique

    2016-01-01

    Background: Regularity effect can affect performance in prospective memory (PM), but little is known on the cognitive processes linked to this effect. Moreover, its impacts with regard to aging remain unknown. To our knowledge, this study is the first to examine regularity effect in PM in a lifespan perspective, with a sample of young, intermediate, and older adults.Objective and design: Our study examined the regularity effect in PM in three groups of participants: 28 young adults (18–30), 1...

  12. Does interval laparoscopic sterilisation influence the risk of lower genital tract infections and menstrual abnormalities

    Directory of Open Access Journals (Sweden)

    G Kistan

    2017-09-01

    Full Text Available Background. Tubal sterilisation is a safe, accessible and effective contraceptive method. There is a paucity of data regarding the risk of genital tract infections and menstrual abnormalities post sterilisation in Durban, South Africa.Objectives. To evaluate the risk of lower genital tract infections (LGTIs and menstrual abnormalities following interval laparoscopic sterilisation.Methods. A prospective cohort study of 225 women undergoing sterilisation between August 2012 and April 2013, with follow-up 1 year post procedure, was conducted at King Dinuzulu Hospital, Durban.Results. Following sterilisation, LGTIs were increased only in women with a history of infection pre sterilisation (odds ratio 6.7; 95% CI 2.2 - 20.9; p=0.002. There was no significant risk of HIV acquisition post sterilisation. In women who had not used contraception or used barrier methods pre sterilisation, we found no significant change in menstrual patterns post sterilisation. There was an increase in menstrual bleeding and dysmenorrhoea post sterilisation among previous combined oral contraceptive users. Among women with amenorrhoea on injectable contraception pre sterilisation, 73.8% reported return to regular menses and 26.2% reported abnormal uterine bleeding post sterilisation. Among injectable contraceptive users with regular menses pre sterilisation, 71.4% reported no change in menses post sterilisation and 28.6% reported abnormal uterine bleeding post sterilisation.Conclusion. In women undergoing interval tubal sterilisation, the risk of LGTIs was only increased in those women with a history. Menstrual abnormalities post sterilisation were more likely in women who used steroid contraception prior to sterilisation.

  13. Data Qulaification Report Flowong Interval Data for Use On the Yucca Mountain Project

    Energy Technology Data Exchange (ETDEWEB)

    C.R. Wilson; T.A. Grant

    2000-08-03

    generalized use and can be appropriately used in a wide variety of applications, so long as consideration is given to accuracy, precision and representativeness of the data for an intended use in a technical product. This qualification report addresses the specific flowing interval data sets selected to support the flowing interval studies in the Analysis and Modeling Report (AMR) ''Probability Distribution of Flowing Interval Spacing'' (Kuzio 1999). This AMR presents an analysis of the spacing of flowing intervals in the saturated volcanic rocks beneath Yucca Mountain. Based on a preponderance of evidence, the flowing interval data used in the AMR are recommended to be qualified for inclusion in technical products in support of the Site Recommendation for generalized uses as described in this report.

  14. 20 CFR 226.14 - Employee regular annuity rate.

    Science.gov (United States)

    2010-04-01

    ... 20 Employees' Benefits 1 2010-04-01 2010-04-01 false Employee regular annuity rate. 226.14 Section... COMPUTING EMPLOYEE, SPOUSE, AND DIVORCED SPOUSE ANNUITIES Computing an Employee Annuity § 226.14 Employee regular annuity rate. The regular annuity rate payable to the employee is the total of the employee tier I...

  15. On interval and cyclic interval edge colorings of (3,5)-biregular graphs

    DEFF Research Database (Denmark)

    Casselgren, Carl Johan; Petrosyan, Petros; Toft, Bjarne

    2017-01-01

    A proper edge coloring f of a graph G with colors 1,2,3,…,t is called an interval coloring if the colors on the edges incident to every vertex of G form an interval of integers. The coloring f is cyclic interval if for every vertex v of G, the colors on the edges incident to v either form an inte...

  16. Point interactions of the dipole type defined through a three-parametric power regularization

    International Nuclear Information System (INIS)

    Zolotaryuk, A V

    2010-01-01

    A family of point interactions of the dipole type is studied in one dimension using a regularization by rectangles in the form of a barrier and a well separated by a finite distance. The rectangles and the distance are parametrized by a squeezing parameter ε → 0 with three powers μ, ν and τ describing the squeezing rates for the barrier, the well and the distance, respectively. This parametrization allows us to construct a whole family of point potentials of the dipole type including some other point interactions, such as e.g. δ-potentials. Varying the power τ, it is possible to obtain in the zero-range limit the following two cases: (i) the limiting δ'-potential is opaque (the conventional result obtained earlier by some authors) or (ii) this potential admits a resonant tunneling (the opposite result obtained recently by other authors). The structure of resonances (if any) also depends on a regularizing sequence. The sets of the {μ, ν, τ}-space where a non-zero (resonant or non-resonant) transmission occurs are found. For all these cases in the zero-range limit the transfer matrix is shown to be with real parameters χ and g depending on a regularizing sequence. Those cases when χ ≠ 1 and g ≠ 0 mean that the corresponding δ'-potential is accompanied by an effective δ-potential.

  17. Decay property of regularity-loss type of solutions in elastic solids with voids

    KAUST Repository

    Said-Houari, Belkacem; Messaoudi, Salim A.

    2013-01-01

    In this article, we consider two porous systems of nonclassical thermoelasticity in the whole real line. We discuss the long-time behaviour of the solutions in the presence of a strong damping acting, together with the heat effect, on the elastic equation and establish several decay results. Those decay results are shown to be very slow and of regularity-loss type. Some improvements of the decay rates have also been given, provided that the initial data belong to some weighted spaces. © 2013 Copyright Taylor and Francis Group, LLC.

  18. Decay property of regularity-loss type of solutions in elastic solids with voids

    KAUST Repository

    Said-Houari, Belkacem

    2013-12-01

    In this article, we consider two porous systems of nonclassical thermoelasticity in the whole real line. We discuss the long-time behaviour of the solutions in the presence of a strong damping acting, together with the heat effect, on the elastic equation and establish several decay results. Those decay results are shown to be very slow and of regularity-loss type. Some improvements of the decay rates have also been given, provided that the initial data belong to some weighted spaces. © 2013 Copyright Taylor and Francis Group, LLC.

  19. Regular algebra and finite machines

    CERN Document Server

    Conway, John Horton

    2012-01-01

    World-famous mathematician John H. Conway based this classic text on a 1966 course he taught at Cambridge University. Geared toward graduate students of mathematics, it will also prove a valuable guide to researchers and professional mathematicians.His topics cover Moore's theory of experiments, Kleene's theory of regular events and expressions, Kleene algebras, the differential calculus of events, factors and the factor matrix, and the theory of operators. Additional subjects include event classes and operator classes, some regulator algebras, context-free languages, communicative regular alg

  20. Stickiness in Hamiltonian systems: From sharply divided to hierarchical phase space

    Science.gov (United States)

    Altmann, Eduardo G.; Motter, Adilson E.; Kantz, Holger

    2006-02-01

    We investigate the dynamics of chaotic trajectories in simple yet physically important Hamiltonian systems with nonhierarchical borders between regular and chaotic regions with positive measures. We show that the stickiness to the border of the regular regions in systems with such a sharply divided phase space occurs through one-parameter families of marginally unstable periodic orbits and is characterized by an exponent γ=2 for the asymptotic power-law decay of the distribution of recurrence times. Generic perturbations lead to systems with hierarchical phase space, where the stickiness is apparently enhanced due to the presence of infinitely many regular islands and Cantori. In this case, we show that the distribution of recurrence times can be composed of a sum of exponentials or a sum of power laws, depending on the relative contribution of the primary and secondary structures of the hierarchy. Numerical verification of our main results are provided for area-preserving maps, mushroom billiards, and the newly defined magnetic mushroom billiards.

  1. ℓ1/2-norm regularized nonnegative low-rank and sparse affinity graph for remote sensing image segmentation

    Science.gov (United States)

    Tian, Shu; Zhang, Ye; Yan, Yiming; Su, Nan

    2016-10-01

    Segmentation of real-world remote sensing images is a challenge due to the complex texture information with high heterogeneity. Thus, graph-based image segmentation methods have been attracting great attention in the field of remote sensing. However, most of the traditional graph-based approaches fail to capture the intrinsic structure of the feature space and are sensitive to noises. A ℓ-norm regularization-based graph segmentation method is proposed to segment remote sensing images. First, we use the occlusion of the random texture model (ORTM) to extract the local histogram features. Then, a ℓ-norm regularized low-rank and sparse representation (LNNLRS) is implemented to construct a ℓ-regularized nonnegative low-rank and sparse graph (LNNLRS-graph), by the union of feature subspaces. Moreover, the LNNLRS-graph has a high ability to discriminate the manifold intrinsic structure of highly homogeneous texture information. Meanwhile, the LNNLRS representation takes advantage of the low-rank and sparse characteristics to remove the noises and corrupted data. Last, we introduce the LNNLRS-graph into the graph regularization nonnegative matrix factorization to enhance the segmentation accuracy. The experimental results using remote sensing images show that when compared to five state-of-the-art image segmentation methods, the proposed method achieves more accurate segmentation results.

  2. 39 CFR 6.1 - Regular meetings, annual meeting.

    Science.gov (United States)

    2010-07-01

    ... 39 Postal Service 1 2010-07-01 2010-07-01 false Regular meetings, annual meeting. 6.1 Section 6.1 Postal Service UNITED STATES POSTAL SERVICE THE BOARD OF GOVERNORS OF THE U.S. POSTAL SERVICE MEETINGS (ARTICLE VI) § 6.1 Regular meetings, annual meeting. The Board shall meet regularly on a schedule...

  3. High-Intensity Interval Training Attenuates Insulin Resistance Induced by Sleep Deprivation in Healthy Males

    Directory of Open Access Journals (Sweden)

    Jorge F. T. de Souza

    2017-12-01

    Full Text Available Introduction: Sleep deprivation can impair several physiological systems and recently, new evidence has pointed to the relationship between a lack of sleep and carbohydrate metabolism, consequently resulting in insulin resistance. To minimize this effect, High-Intensity Interval Training (HIIT is emerging as a potential strategy.Objective: The aim of this study was to investigate the effects of HIIT on insulin resistance induced by sleep deprivation.Method: Eleven healthy male volunteers were recruited, aged 18–35 years, who declared taking 7–8 h sleep per night. All volunteers were submitted to four different conditions: a single night of regular sleep (RS condition, 24 h of total sleep deprivation (SD condition, HIIT training followed by regular sleep (HIIT+RS condition, and HIIT training followed by 24 h of total sleep deprivation (HIIT+SD condition. They performed six training sessions over 2 weeks and each session consisted of 8–12 × 60 s intervals at 100% of peak power output. In each experimental condition, tests for glucose, insulin, cortisol, free fatty acids, and insulin sensitivity, measured by oral glucose tolerance test (OGTT, were performed.Results: Sleep deprivation increased glycaemia and insulin levels, as well as the area under the curve. Furthermore, an increase in free fatty acids concentrations and basal metabolism was observed. There were no differences in the concentrations of cortisol. However, HIIT before 24 h of sleep deprivation attenuated the increase of glucose, insulin, and free fatty acids.Conclusion: Twenty-four hours of sleep deprivation resulted in acute insulin resistance. However, HIIT is an effective strategy to minimize the deleterious effects promoted by this condition.

  4. High-Intensity Interval Training Attenuates Insulin Resistance Induced by Sleep Deprivation in Healthy Males.

    Science.gov (United States)

    de Souza, Jorge F T; Dáttilo, Murilo; de Mello, Marco T; Tufik, Sergio; Antunes, Hanna K M

    2017-01-01

    Introduction: Sleep deprivation can impair several physiological systems and recently, new evidence has pointed to the relationship between a lack of sleep and carbohydrate metabolism, consequently resulting in insulin resistance. To minimize this effect, High-Intensity Interval Training (HIIT) is emerging as a potential strategy. Objective: The aim of this study was to investigate the effects of HIIT on insulin resistance induced by sleep deprivation. Method: Eleven healthy male volunteers were recruited, aged 18-35 years, who declared taking 7-8 h sleep per night. All volunteers were submitted to four different conditions: a single night of regular sleep (RS condition), 24 h of total sleep deprivation ( SD condition), HIIT training followed by regular sleep (HIIT+RS condition), and HIIT training followed by 24 h of total sleep deprivation (HIIT+ SD condition). They performed six training sessions over 2 weeks and each session consisted of 8-12 × 60 s intervals at 100% of peak power output. In each experimental condition, tests for glucose, insulin, cortisol, free fatty acids, and insulin sensitivity, measured by oral glucose tolerance test (OGTT), were performed. Results: Sleep deprivation increased glycaemia and insulin levels, as well as the area under the curve. Furthermore, an increase in free fatty acids concentrations and basal metabolism was observed. There were no differences in the concentrations of cortisol. However, HIIT before 24 h of sleep deprivation attenuated the increase of glucose, insulin, and free fatty acids. Conclusion: Twenty-four hours of sleep deprivation resulted in acute insulin resistance. However, HIIT is an effective strategy to minimize the deleterious effects promoted by this condition.

  5. High-Intensity Interval Training Attenuates Insulin Resistance Induced by Sleep Deprivation in Healthy Males

    Science.gov (United States)

    de Souza, Jorge F. T.; Dáttilo, Murilo; de Mello, Marco T.; Tufik, Sergio; Antunes, Hanna K. M.

    2017-01-01

    Introduction: Sleep deprivation can impair several physiological systems and recently, new evidence has pointed to the relationship between a lack of sleep and carbohydrate metabolism, consequently resulting in insulin resistance. To minimize this effect, High-Intensity Interval Training (HIIT) is emerging as a potential strategy. Objective: The aim of this study was to investigate the effects of HIIT on insulin resistance induced by sleep deprivation. Method: Eleven healthy male volunteers were recruited, aged 18–35 years, who declared taking 7–8 h sleep per night. All volunteers were submitted to four different conditions: a single night of regular sleep (RS condition), 24 h of total sleep deprivation (SD condition), HIIT training followed by regular sleep (HIIT+RS condition), and HIIT training followed by 24 h of total sleep deprivation (HIIT+SD condition). They performed six training sessions over 2 weeks and each session consisted of 8–12 × 60 s intervals at 100% of peak power output. In each experimental condition, tests for glucose, insulin, cortisol, free fatty acids, and insulin sensitivity, measured by oral glucose tolerance test (OGTT), were performed. Results: Sleep deprivation increased glycaemia and insulin levels, as well as the area under the curve. Furthermore, an increase in free fatty acids concentrations and basal metabolism was observed. There were no differences in the concentrations of cortisol. However, HIIT before 24 h of sleep deprivation attenuated the increase of glucose, insulin, and free fatty acids. Conclusion: Twenty-four hours of sleep deprivation resulted in acute insulin resistance. However, HIIT is an effective strategy to minimize the deleterious effects promoted by this condition. PMID:29270126

  6. Analysis in Banach spaces

    CERN Document Server

    Hytönen, Tuomas; Veraar, Mark; Weis, Lutz

    The present volume develops the theory of integration in Banach spaces, martingales and UMD spaces, and culminates in a treatment of the Hilbert transform, Littlewood-Paley theory and the vector-valued Mihlin multiplier theorem. Over the past fifteen years, motivated by regularity problems in evolution equations, there has been tremendous progress in the analysis of Banach space-valued functions and processes. The contents of this extensive and powerful toolbox have been mostly scattered around in research papers and lecture notes. Collecting this diverse body of material into a unified and accessible presentation fills a gap in the existing literature. The principal audience that we have in mind consists of researchers who need and use Analysis in Banach Spaces as a tool for studying problems in partial differential equations, harmonic analysis, and stochastic analysis. Self-contained and offering complete proofs, this work is accessible to graduate students and researchers with a background in functional an...

  7. Ombud's corner: space invaders

    CERN Multimedia

    Sudeshna Datta-Cockerill

    2015-01-01

    When normal communication breaks down and there is no sharing anymore, office-mates can become ‘space invaders’. Very often, the situation can be resolved effectively by taking just a few simple steps...   The lack of office space at CERN is a permanent issue that the various departments regularly have to address. As a result, very often this precious space where we spend the entire day has to be shared with other colleagues. Office-mates may come from different backgrounds and cultures and may have very different habits and behaviours; they may also have different activities during the day, sometimes requiring unusual, (perhaps even strange?) interactions with the space they occupy; finally, their presence might be irregular, making it very difficult for us to establish a stable relationship. Mark and Claire share an office as well as some professional activities. In the beginning, the relationship seems to work normally but, over time, the communication between them ste...

  8. Space-Time Discrete KPZ Equation

    Science.gov (United States)

    Cannizzaro, G.; Matetski, K.

    2018-03-01

    We study a general family of space-time discretizations of the KPZ equation and show that they converge to its solution. The approach we follow makes use of basic elements of the theory of regularity structures (Hairer in Invent Math 198(2):269-504, 2014) as well as its discrete counterpart (Hairer and Matetski in Discretizations of rough stochastic PDEs, 2015. arXiv:1511.06937). Since the discretization is in both space and time and we allow non-standard discretization for the product, the methods mentioned above have to be suitably modified in order to accommodate the structure of the models under study.

  9. The lucid interval associated with epidural bleeding: evolving understanding.

    Science.gov (United States)

    Ganz, Jeremy C

    2013-04-01

    The aim of this paper was to elucidate the evolution of our understanding of the term "lucid interval." A number of texts were reviewed to assess their suitability for analysis. The primary requirement was that the text contain detailed descriptions of a series of patients. Details of the clinical course, the findings and timing of surgery, and, when relevant, the time of death and postmortem findings were required. Books written by Henri-François Le Dran, Percival Pott, and James Hill fulfilled these criteria. Surgical findings included the presence and type of fractures, changes in the bone, separation of periosteum, malodorous or purulent material, tense brain, and hematoma. Postmortem findings supplemented and/or complemented the surgical findings. The courses of the patients were then tabulated, and the correlation between different clinical and operative findings was thereby determined. Our understanding of a lucid interval began in the early 18th century with the work of Henri-François Le Dran and Percival Pott in London. They did not, however, demonstrate an interval without symptoms between trauma and deterioration in patients with epidural hematomas (EDHs). The interval they described was longer than usually expected with EDHs and occurred exclusively in patients who had a posttraumatic infection. In 1751, James Hill, from Dumfries, Scotland, described the first hematoma-related lucid interval in a patient with a subdural hematoma. The first case of a lucid interval associated with an EDH was described by John Abernethy. In the 19th century, Jonathan Hutchinson and Walter Jacobson described the interval as it is known today, in cases of EDH. The most recent work on the topic came from studies in Cincinnati and Oslo, where it was demonstrated that bleeding can separate dura mater and that hemorrhage into the epidural space can be shunted out via the veins. This shunting could delay the accumulation of a hematoma and thus the rise in intracranial pressure

  10. Circadian profile of QT interval and QT interval variability in 172 healthy volunteers

    DEFF Research Database (Denmark)

    Bonnemeier, Hendrik; Wiegand, Uwe K H; Braasch, Wiebke

    2003-01-01

    of sleep. QT and R-R intervals revealed a characteristic day-night-pattern. Diurnal profiles of QT interval variability exhibited a significant increase in the morning hours (6-9 AM; P ... lower at day- and nighttime. Aging was associated with an increase of QT interval mainly at daytime and a significant shift of the T wave apex towards the end of the T wave. The circadian profile of ventricular repolarization is strongly related to the mean R-R interval, however, there are significant...

  11. Two intervals Rényi entanglement entropy of compact free boson on torus

    International Nuclear Information System (INIS)

    Liu, Feihu; Liu, Xiao

    2016-01-01

    We compute the N=2 Rényi entanglement entropy of two intervals at equal time in a circle, for the theory of a 2D compact complex free scalar at finite temperature. This is carried out by performing functional integral on a genus 3 ramified cover of the torus, wherein the quantum part of the integral is captured by the four point function of twist fields on the worldsheet torus, and the classical piece is given by summing over winding modes of the genus 3 surface onto the target space torus. The final result is given in terms of a product of theta functions and certain multi-dimensional theta functions. We demonstrate the T-duality invariance of the result. We also study its low temperature limit. In the case in which the size of the intervals and of their separation are much smaller than the whole system, our result is in exact agreement with the known result for two intervals on an infinite system at zero temperature http://dx.doi.org/10.1088/1742-5468/2009/11/P11001. In the case in which the separation between the two intervals is much smaller than the interval length, the leading thermal corrections take the same universal form as proposed in http://dx.doi.org/10.1103/PhysRevLett.112.171603, http://dx.doi.org/10.1103/PhysRevD.91.105013 for Rényi entanglement entropy of a single interval.

  12. A regularized stationary mean-field game

    KAUST Repository

    Yang, Xianjin

    2016-01-01

    In the thesis, we discuss the existence and numerical approximations of solutions of a regularized mean-field game with a low-order regularization. In the first part, we prove a priori estimates and use the continuation method to obtain the existence of a solution with a positive density. Finally, we introduce the monotone flow method and solve the system numerically.

  13. A regularized stationary mean-field game

    KAUST Repository

    Yang, Xianjin

    2016-04-19

    In the thesis, we discuss the existence and numerical approximations of solutions of a regularized mean-field game with a low-order regularization. In the first part, we prove a priori estimates and use the continuation method to obtain the existence of a solution with a positive density. Finally, we introduce the monotone flow method and solve the system numerically.

  14. Automating InDesign with Regular Expressions

    CERN Document Server

    Kahrel, Peter

    2006-01-01

    If you need to make automated changes to InDesign documents beyond what basic search and replace can handle, you need regular expressions, and a bit of scripting to make them work. This Short Cut explains both how to write regular expressions, so you can find and replace the right things, and how to use them in InDesign specifically.

  15. Optimal behaviour can violate the principle of regularity.

    Science.gov (United States)

    Trimmer, Pete C

    2013-07-22

    Understanding decisions is a fundamental aim of behavioural ecology, psychology and economics. The regularity axiom of utility theory holds that a preference between options should be maintained when other options are made available. Empirical studies have shown that animals violate regularity but this has not been understood from a theoretical perspective, such decisions have therefore been labelled as irrational. Here, I use models of state-dependent behaviour to demonstrate that choices can violate regularity even when behavioural strategies are optimal. I also show that the range of conditions over which regularity should be violated can be larger when options do not always persist into the future. Consequently, utility theory--based on axioms, including transitivity, regularity and the independence of irrelevant alternatives--is undermined, because even alternatives that are never chosen by an animal (in its current state) can be relevant to a decision.

  16. Regular Breakfast and Blood Lead Levels among Preschool Children

    Directory of Open Access Journals (Sweden)

    Needleman Herbert

    2011-04-01

    Full Text Available Abstract Background Previous studies have shown that fasting increases lead absorption in the gastrointestinal tract of adults. Regular meals/snacks are recommended as a nutritional intervention for lead poisoning in children, but epidemiological evidence of links between fasting and blood lead levels (B-Pb is rare. The purpose of this study was to examine the association between eating a regular breakfast and B-Pb among children using data from the China Jintan Child Cohort Study. Methods Parents completed a questionnaire regarding children's breakfast-eating habit (regular or not, demographics, and food frequency. Whole blood samples were collected from 1,344 children for the measurements of B-Pb and micronutrients (iron, copper, zinc, calcium, and magnesium. B-Pb and other measures were compared between children with and without regular breakfast. Linear regression modeling was used to evaluate the association between regular breakfast and log-transformed B-Pb. The association between regular breakfast and risk of lead poisoning (B-Pb≥10 μg/dL was examined using logistic regression modeling. Results Median B-Pb among children who ate breakfast regularly and those who did not eat breakfast regularly were 6.1 μg/dL and 7.2 μg/dL, respectively. Eating breakfast was also associated with greater zinc blood levels. Adjusting for other relevant factors, the linear regression model revealed that eating breakfast regularly was significantly associated with lower B-Pb (beta = -0.10 units of log-transformed B-Pb compared with children who did not eat breakfast regularly, p = 0.02. Conclusion The present study provides some initial human data supporting the notion that eating a regular breakfast might reduce B-Pb in young children. To our knowledge, this is the first human study exploring the association between breakfast frequency and B-Pb in young children.

  17. On the equivalence of different regularization methods

    International Nuclear Information System (INIS)

    Brzezowski, S.

    1985-01-01

    The R-circunflex-operation preceded by the regularization procedure is discussed. Some arguments are given, according to which the results may depend on the method of regularization, introduced in order to avoid divergences in perturbation calculations. 10 refs. (author)

  18. Birth spacing and fertility limitation: a behavioral analysis of a nineteenth century frontier population.

    Science.gov (United States)

    Anderton, D L; Bean, L L

    1985-05-01

    Our analysis of changing birth interval distributions over the course of a fertility transition from natural to controlled fertility has examined three closely related propositions. First, within both natural fertility populations (identified at the aggregate level) and cohorts following the onset of fertility limitation, we hypothesized that substantial groups of women with long birth intervals across the individually specified childbearing careers could be identified. That is, even during periods when fertility behavior at the aggregate level is consistent with a natural fertility regime, birth intervals at all parities are inversely related to completed family size. Our tabular analysis enables us to conclude that birth spacing patterns are parity dependent; there is stability in CEB-parity specific mean and birth interval variance over the entire transition. Our evidence does not suggest that the early group of women limiting and spacing births was marked by infecundity. Secondly, the transition appears to be associated with an increasingly larger proportion of women shifting to the same spacing schedules associated with smaller families in earlier cohorts. Thirdly, variations in birth spacing by age of marriage indicate that changes in birth intervals over time are at least indirectly associated with age of marriage, indicating an additional compositional effect. The evidence we have presented on spacing behavior does not negate the argument that parity-dependent stopping behavior was a powerful factor in the fertility transition. Our data also provide evidence of attempts to truncate childbearing. Specifically, the smaller the completed family size, the longer the ultimate birth interval; and ultimate birth intervals increase across cohorts controlling CEB and parity. But spacing appears to represent an additional strategy of fertility limitation. Thus, it may be necessary to distinguish spacing and stopping behavior if one wishes to clarify behavioral

  19. Regular-chaos transition of the energy spectrum and electromagnetic transition intensities in 44V nucleus using the framework of the nuclear shell model

    International Nuclear Information System (INIS)

    Hamoudi, A.K.; Abdul Majeed Al-Rahmani, A.

    2012-01-01

    The spectral fluctuations and the statistics of electromagnetic transition intensities and electromagnetic moments in 44 V nucleus are studied by the framework of the interacting shell model, using the FPD6 as a realistic effective interaction in the isospin formalism for 4 particles move in the fp-model space with a 40 Ca core. To look for a regular-chaos transition in 44 V nucleus, we perform shell model calculations using various interaction strengths β to the off-diagonal matrix elements of the FPD6. The nearest-neighbors level spacing distribution P(s) and the distribution of electromagnetic transition intensities [such as, B(M1) and B(E2) transitions] are found to have a regular dynamic at β=0, a chaotic dynamic at β⩾0.3 and an intermediate situation at 0 3 statistic we have found a regular dynamic at β=0, a chaotic dynamic at β⩾0.4 and an intermediate situation at 0<β<0.4. It is also found that the statistics of the squares of M1 and E2 moments, which are consistent with a Porter-Thomas distribution, have no dependence on the interaction strength β.

  20. Bounded Perturbation Regularization for Linear Least Squares Estimation

    KAUST Repository

    Ballal, Tarig

    2017-10-18

    This paper addresses the problem of selecting the regularization parameter for linear least-squares estimation. We propose a new technique called bounded perturbation regularization (BPR). In the proposed BPR method, a perturbation with a bounded norm is allowed into the linear transformation matrix to improve the singular-value structure. Following this, the problem is formulated as a min-max optimization problem. Next, the min-max problem is converted to an equivalent minimization problem to estimate the unknown vector quantity. The solution of the minimization problem is shown to converge to that of the ℓ2 -regularized least squares problem, with the unknown regularizer related to the norm bound of the introduced perturbation through a nonlinear constraint. A procedure is proposed that combines the constraint equation with the mean squared error (MSE) criterion to develop an approximately optimal regularization parameter selection algorithm. Both direct and indirect applications of the proposed method are considered. Comparisons with different Tikhonov regularization parameter selection methods, as well as with other relevant methods, are carried out. Numerical results demonstrate that the proposed method provides significant improvement over state-of-the-art methods.

  1. Ad Hoc Physical Hilbert Spaces in Quantum Mechanics

    Czech Academy of Sciences Publication Activity Database

    Fernandez, F. M.; Garcia, J.; Semorádová, Iveta; Znojil, Miloslav

    2015-01-01

    Roč. 54, č. 12 (2015), s. 4187-4203 ISSN 0020-7748 Institutional support: RVO:61389005 Keywords : quantum mechanics * physical Hilbert spaces * ad hoc inner product * singular potentials regularized * low lying energies Subject RIV: BE - Theoretical Physics Impact factor: 1.041, year: 2015

  2. MRI reconstruction with joint global regularization and transform learning.

    Science.gov (United States)

    Tanc, A Korhan; Eksioglu, Ender M

    2016-10-01

    Sparsity based regularization has been a popular approach to remedy the measurement scarcity in image reconstruction. Recently, sparsifying transforms learned from image patches have been utilized as an effective regularizer for the Magnetic Resonance Imaging (MRI) reconstruction. Here, we infuse additional global regularization terms to the patch-based transform learning. We develop an algorithm to solve the resulting novel cost function, which includes both patchwise and global regularization terms. Extensive simulation results indicate that the introduced mixed approach has improved MRI reconstruction performance, when compared to the algorithms which use either of the patchwise transform learning or global regularization terms alone. Copyright © 2016 Elsevier Ltd. All rights reserved.

  3. How to Reclaim Space?

    DEFF Research Database (Denmark)

    Blasco, Maribel

    2016-01-01

    multitasking and very little autonomy. So far, very little attention has been paid to this problem in responsible management education scholarship. So far, we know little about how to build productive learning spaces into curricula. Inspired by the Japanese concept of ma, meaning ‘inbetweenness’, ‘interval...

  4. Dynamical tunneling in systems with a mixed phase space

    International Nuclear Information System (INIS)

    Loeck, Steffen

    2010-01-01

    Tunneling is one of the most prominent features of quantum mechanics. While the tunneling process in one-dimensional integrable systems is well understood, its quantitative prediction for systems with a mixed phase space is a long-standing open challenge. In such systems regions of regular and chaotic dynamics coexist in phase space, which are classically separated but quantum mechanically coupled by the process of dynamical tunneling. We derive a prediction of dynamical tunneling rates which describe the decay of states localized inside the regular region towards the so-called chaotic sea. This approach uses a fictitious integrable system which mimics the dynamics inside the regular domain and extends it into the chaotic region. Excellent agreement with numerical data is found for kicked systems, billiards, and optical microcavities, if nonlinear resonances are negligible. Semiclassically, however, such nonlinear resonance chains dominate the tunneling process. Hence, we combine our approach with an improved resonance-assisted tunneling theory and derive a unified prediction which is valid from the quantum to the semiclassical regime. We obtain results which show a drastically improved accuracy of several orders of magnitude compared to previous studies. (orig.)

  5. Dynamical tunneling in systems with a mixed phase space

    Energy Technology Data Exchange (ETDEWEB)

    Loeck, Steffen

    2010-04-22

    Tunneling is one of the most prominent features of quantum mechanics. While the tunneling process in one-dimensional integrable systems is well understood, its quantitative prediction for systems with a mixed phase space is a long-standing open challenge. In such systems regions of regular and chaotic dynamics coexist in phase space, which are classically separated but quantum mechanically coupled by the process of dynamical tunneling. We derive a prediction of dynamical tunneling rates which describe the decay of states localized inside the regular region towards the so-called chaotic sea. This approach uses a fictitious integrable system which mimics the dynamics inside the regular domain and extends it into the chaotic region. Excellent agreement with numerical data is found for kicked systems, billiards, and optical microcavities, if nonlinear resonances are negligible. Semiclassically, however, such nonlinear resonance chains dominate the tunneling process. Hence, we combine our approach with an improved resonance-assisted tunneling theory and derive a unified prediction which is valid from the quantum to the semiclassical regime. We obtain results which show a drastically improved accuracy of several orders of magnitude compared to previous studies. (orig.)

  6. Strictly-regular number system and data structures

    DEFF Research Database (Denmark)

    Elmasry, Amr Ahmed Abd Elmoneim; Jensen, Claus; Katajainen, Jyrki

    2010-01-01

    We introduce a new number system that we call the strictly-regular system, which efficiently supports the operations: digit-increment, digit-decrement, cut, concatenate, and add. Compared to other number systems, the strictly-regular system has distinguishable properties. It is superior to the re...

  7. Analysis of regularized Navier-Stokes equations, 2

    Science.gov (United States)

    Ou, Yuh-Roung; Sritharan, S. S.

    1989-01-01

    A practically important regularization of the Navier-Stokes equations was analyzed. As a continuation of the previous work, the structure of the attractors characterizing the solutins was studied. Local as well as global invariant manifolds were found. Regularity properties of these manifolds are analyzed.

  8. Regularities, Natural Patterns and Laws of Nature

    Directory of Open Access Journals (Sweden)

    Stathis Psillos

    2014-02-01

    Full Text Available  The goal of this paper is to sketch an empiricist metaphysics of laws of nature. The key idea is that there are regularities without regularity-enforcers. Differently put, there are natural laws without law-makers of a distinct metaphysical kind. This sketch will rely on the concept of a natural pattern and more significantly on the existence of a network of natural patterns in nature. The relation between a regularity and a pattern will be analysed in terms of mereology.  Here is the road map. In section 2, I will briefly discuss the relation between empiricism and metaphysics, aiming to show that an empiricist metaphysics is possible. In section 3, I will offer arguments against stronger metaphysical views of laws. Then, in section 4 I will motivate nomic objectivism. In section 5, I will address the question ‘what is a regularity?’ and will develop a novel answer to it, based on the notion of a natural pattern. In section 6, I will raise the question: ‘what is a law of nature?’, the answer to which will be: a law of nature is a regularity that is characterised by the unity of a natural pattern.

  9. Consistent Partial Least Squares Path Modeling via Regularization.

    Science.gov (United States)

    Jung, Sunho; Park, JaeHong

    2018-01-01

    Partial least squares (PLS) path modeling is a component-based structural equation modeling that has been adopted in social and psychological research due to its data-analytic capability and flexibility. A recent methodological advance is consistent PLS (PLSc), designed to produce consistent estimates of path coefficients in structural models involving common factors. In practice, however, PLSc may frequently encounter multicollinearity in part because it takes a strategy of estimating path coefficients based on consistent correlations among independent latent variables. PLSc has yet no remedy for this multicollinearity problem, which can cause loss of statistical power and accuracy in parameter estimation. Thus, a ridge type of regularization is incorporated into PLSc, creating a new technique called regularized PLSc. A comprehensive simulation study is conducted to evaluate the performance of regularized PLSc as compared to its non-regularized counterpart in terms of power and accuracy. The results show that our regularized PLSc is recommended for use when serious multicollinearity is present.

  10. Consistent Partial Least Squares Path Modeling via Regularization

    Directory of Open Access Journals (Sweden)

    Sunho Jung

    2018-02-01

    Full Text Available Partial least squares (PLS path modeling is a component-based structural equation modeling that has been adopted in social and psychological research due to its data-analytic capability and flexibility. A recent methodological advance is consistent PLS (PLSc, designed to produce consistent estimates of path coefficients in structural models involving common factors. In practice, however, PLSc may frequently encounter multicollinearity in part because it takes a strategy of estimating path coefficients based on consistent correlations among independent latent variables. PLSc has yet no remedy for this multicollinearity problem, which can cause loss of statistical power and accuracy in parameter estimation. Thus, a ridge type of regularization is incorporated into PLSc, creating a new technique called regularized PLSc. A comprehensive simulation study is conducted to evaluate the performance of regularized PLSc as compared to its non-regularized counterpart in terms of power and accuracy. The results show that our regularized PLSc is recommended for use when serious multicollinearity is present.

  11. On Some General Regularities of Formation of the Planetary Systems

    Directory of Open Access Journals (Sweden)

    Belyakov A. V.

    2014-01-01

    Full Text Available J.Wheeler’s geometrodynamic concept has been used, in which space continuum is considered as a topologically non-unitary coherent surface admitting the existence of transitions of the input-output kind between distant regions of the space in an additional dimension. This model assumes the existence of closed structures (micro- and macro- contours formed due to the balance between main interactions: gravitational, electric, magnetic, and inertial forces. It is such macrocontours that have been demonstrated to form — independently of their material basis — the essential structure of objects at various levels of organization of matter. On the basis of this concept in this paper basic regularities acting during formation planetary systems have been obtained. The existence of two sharply different types of planetary systems has been determined. The dependencies linking the masses of the planets, the diameters of the planets, the orbital radii of the planet, and the mass of the central body have been deduced. The possibility of formation of Earth-like planets near brown dwarfs has been grounded. The minimum mass of the planet, which may arise in the planetary system, has been defined.

  12. Regular Benzodiazepine and Z-Substance Use and Risk of Dementia: An Analysis of German Claims Data.

    Science.gov (United States)

    Gomm, Willy; von Holt, Klaus; Thomé, Friederike; Broich, Karl; Maier, Wolfgang; Weckbecker, Klaus; Fink, Anne; Doblhammer, Gabriele; Haenisch, Britta

    2016-09-06

    While acute detrimental effects of benzodiazepine (BDZ), and BDZ and related z-substance (BDZR) use on cognition and memory are known, the association of BDZR use and risk of dementia in the elderly is controversially discussed. Previous studies on cohort or claims data mostly show an increased risk for dementia with the use of BDZs or BDZRs. For Germany, analyses on large population-based data sets are missing. To evaluate the association between regular BDZR use and incident any dementia in a large German claims data set. Using longitudinal German public health insurance data from 2004 to 2011 we analyzed the association between regular BDZR use (versus no BDZR use) and incident dementia in a case-control design. We examined patient samples aged≥60 years that were free of dementia at baseline. To address potential protopathic bias we introduced a lag time between BDZR prescription and dementia diagnosis. Odds ratios were calculated applying conditional logistic regression, adjusted for potential confounding factors such as comorbidities and polypharmacy. The regular use of BDZRs was associated with a significant increased risk of incident dementia for patients aged≥60 years (adjusted odds ratio [OR] 1.21, 95% confidence interval [CI] 1.13-1.29). The association was slightly stronger for long-acting substances than for short-acting ones. A trend for increased risk for dementia with higher exposure was observed. The restricted use of BDZRs may contribute to dementia prevention in the elderly.

  13. Bilinear Regularized Locality Preserving Learning on Riemannian Graph for Motor Imagery BCI.

    Science.gov (United States)

    Xie, Xiaofeng; Yu, Zhu Liang; Gu, Zhenghui; Zhang, Jun; Cen, Ling; Li, Yuanqing

    2018-03-01

    In off-line training of motor imagery-based brain-computer interfaces (BCIs), to enhance the generalization performance of the learned classifier, the local information contained in test data could be used to improve the performance of motor imagery as well. Further considering that the covariance matrices of electroencephalogram (EEG) signal lie on Riemannian manifold, in this paper, we construct a Riemannian graph to incorporate the information of training and test data into processing. The adjacency and weight in Riemannian graph are determined by the geodesic distance of Riemannian manifold. Then, a new graph embedding algorithm, called bilinear regularized locality preserving (BRLP), is derived upon the Riemannian graph for addressing the problems of high dimensionality frequently arising in BCIs. With a proposed regularization term encoding prior information of EEG channels, the BRLP could obtain more robust performance. Finally, an efficient classification algorithm based on extreme learning machine is proposed to perform on the tangent space of learned embedding. Experimental evaluations on the BCI competition and in-house data sets reveal that the proposed algorithms could obtain significantly higher performance than many competition algorithms after using same filter process.

  14. Regularization of the Boundary-Saddle-Node Bifurcation

    Directory of Open Access Journals (Sweden)

    Xia Liu

    2018-01-01

    Full Text Available In this paper we treat a particular class of planar Filippov systems which consist of two smooth systems that are separated by a discontinuity boundary. In such systems one vector field undergoes a saddle-node bifurcation while the other vector field is transversal to the boundary. The boundary-saddle-node (BSN bifurcation occurs at a critical value when the saddle-node point is located on the discontinuity boundary. We derive a local topological normal form for the BSN bifurcation and study its local dynamics by applying the classical Filippov’s convex method and a novel regularization approach. In fact, by the regularization approach a given Filippov system is approximated by a piecewise-smooth continuous system. Moreover, the regularization process produces a singular perturbation problem where the original discontinuous set becomes a center manifold. Thus, the regularization enables us to make use of the established theories for continuous systems and slow-fast systems to study the local behavior around the BSN bifurcation.

  15. Low-Complexity Regularization Algorithms for Image Deblurring

    KAUST Repository

    Alanazi, Abdulrahman

    2016-11-01

    Image restoration problems deal with images in which information has been degraded by blur or noise. In practice, the blur is usually caused by atmospheric turbulence, motion, camera shake, and several other mechanical or physical processes. In this study, we present two regularization algorithms for the image deblurring problem. We first present a new method based on solving a regularized least-squares (RLS) problem. This method is proposed to find a near-optimal value of the regularization parameter in the RLS problems. Experimental results on the non-blind image deblurring problem are presented. In all experiments, comparisons are made with three benchmark methods. The results demonstrate that the proposed method clearly outperforms the other methods in terms of both the output PSNR and structural similarity, as well as the visual quality of the deblurred images. To reduce the complexity of the proposed algorithm, we propose a technique based on the bootstrap method to estimate the regularization parameter in low and high-resolution images. Numerical results show that the proposed technique can effectively reduce the computational complexity of the proposed algorithms. In addition, for some cases where the point spread function (PSF) is separable, we propose using a Kronecker product so as to reduce the computations. Furthermore, in the case where the image is smooth, it is always desirable to replace the regularization term in the RLS problems by a total variation term. Therefore, we propose a novel method for adaptively selecting the regularization parameter in a so-called square root regularized total variation (SRTV). Experimental results demonstrate that our proposed method outperforms the other benchmark methods when applied to smooth images in terms of PSNR, SSIM and the restored image quality. In this thesis, we focus on the non-blind image deblurring problem, where the blur kernel is assumed to be known. However, we developed algorithms that also work

  16. Improvements in GRACE Gravity Fields Using Regularization

    Science.gov (United States)

    Save, H.; Bettadpur, S.; Tapley, B. D.

    2008-12-01

    The unconstrained global gravity field models derived from GRACE are susceptible to systematic errors that show up as broad "stripes" aligned in a North-South direction on the global maps of mass flux. These errors are believed to be a consequence of both systematic and random errors in the data that are amplified by the nature of the gravity field inverse problem. These errors impede scientific exploitation of the GRACE data products, and limit the realizable spatial resolution of the GRACE global gravity fields in certain regions. We use regularization techniques to reduce these "stripe" errors in the gravity field products. The regularization criteria are designed such that there is no attenuation of the signal and that the solutions fit the observations as well as an unconstrained solution. We have used a computationally inexpensive method, normally referred to as "L-ribbon", to find the regularization parameter. This paper discusses the characteristics and statistics of a 5-year time-series of regularized gravity field solutions. The solutions show markedly reduced stripes, are of uniformly good quality over time, and leave little or no systematic observation residuals, which is a frequent consequence of signal suppression from regularization. Up to degree 14, the signal in regularized solution shows correlation greater than 0.8 with the un-regularized CSR Release-04 solutions. Signals from large-amplitude and small-spatial extent events - such as the Great Sumatra Andaman Earthquake of 2004 - are visible in the global solutions without using special post-facto error reduction techniques employed previously in the literature. Hydrological signals as small as 5 cm water-layer equivalent in the small river basins, like Indus and Nile for example, are clearly evident, in contrast to noisy estimates from RL04. The residual variability over the oceans relative to a seasonal fit is small except at higher latitudes, and is evident without the need for de-striping or

  17. Loop-space quantum formulation of free electromagnetism

    International Nuclear Information System (INIS)

    Di Bartolo, C.; Nori, F.; Gambini, R.; Trias, A.

    1983-01-01

    A procedure for direct quantization of free electromagnetism in the loop-space is proposed. Explicit solutions for the loop-dependent vacuum and the Wilson loop-average are given. It is shown that elementary lines of magnetic field appear as extremals in the vacuum state as a result of the regularization procedure

  18. The effects of interval- vs. continuous exercise on excess post-exercise oxygen consumption and substrate oxidation rates in subjects with type 2 diabetes

    DEFF Research Database (Denmark)

    Karstoft, Kristian; Wallis, Gareth A.; Pedersen, Bente K.

    2016-01-01

    Background For unknown reasons, interval training often reduces body weight more than energy-expenditure matched continuous training. We compared the acute effects of time-duration and oxygen-consumption matched interval- vs. continuous exercise on excess post-exercise oxygen consumption (EPOC...... (MMTT, 450 kcal) was consumed by the subjects 45 min after completion of the intervention with blood samples taken regularly. Results Exercise interventions were successfully matched for total oxygen consumption (CW = 1641 ± 133 mL/min; IW = 1634 ± 126 mL/min, P > 0.05). EPOC was higher after IW (8......, free fatty acids and glycerol concentrations, and glycerol kinetics were increased comparably during and after IW and CW compared to CON. Conclusions Interval exercise results in greater EPOC than oxygen-consumption matched continuous exercise during a post-exercise MMTT in subjects with T2D, whereas...

  19. Deterministic automata for extended regular expressions

    Directory of Open Access Journals (Sweden)

    Syzdykov Mirzakhmet

    2017-12-01

    Full Text Available In this work we present the algorithms to produce deterministic finite automaton (DFA for extended operators in regular expressions like intersection, subtraction and complement. The method like “overriding” of the source NFA(NFA not defined with subset construction rules is used. The past work described only the algorithm for AND-operator (or intersection of regular languages; in this paper the construction for the MINUS-operator (and complement is shown.

  20. Regularities of intermediate adsorption complex relaxation

    International Nuclear Information System (INIS)

    Manukova, L.A.

    1982-01-01

    The experimental data, characterizing the regularities of intermediate adsorption complex relaxation in the polycrystalline Mo-N 2 system at 77 K are given. The method of molecular beam has been used in the investigation. The analytical expressions of change regularity in the relaxation process of full and specific rates - of transition from intermediate state into ''non-reversible'', of desorption into the gas phase and accumUlation of the particles in the intermediate state are obtained

  1. Systematic implementation of implicit regularization for multi-loop Feynman Diagrams

    International Nuclear Information System (INIS)

    Cherchiglia, Adriano Lana; Sampaio, Marcos; Nemes, Maria Carolina

    2011-01-01

    Full text: Implicit Regularization (IR) is a candidate to become an invariant framework in momentum space to perform Feynman diagram calculations to arbitrary loop order. The essence of the method is to write the divergences in terms of loop integrals in one internal momentum which do not need to be explicitly evaluated. Moreover it acts in the physical dimension of the theory and gauge invariance is controlled by regularization dependent surface terms which when set to zero define a constrained version of IR (CIR) and deliver gauge invariant amplitudes automatically. Therefore it is in principle applicable to all physical relevant quantum field theories, supersymmetric gauge theories included. A non trivial question is whether we can generalize this program to arbitrary loop order in consonance with locality, unitarity and Lorentz invariance, especially when overlapping divergences occur. In this work we present a systematic implementation of our method that automatically displays the terms to be subtracted by Bogoliubov's recursion formula. Therefore, we achieve a twofold objective: we show that the IR program respects unitarity, locality and Lorentz invariance and we show that our method is consistent since we are able to display the divergent content of a multi-loop amplitude in a well defined set of basic divergent integrals in one internal momentum. We present several examples (from 1-loop to n-loops) using scalar φ 6 3 theory in order to help the reader understand and visualize the essence of the IR program. The choice of a scalar theory does not reduce the generality of the method presented since all other physical theories can be treated within the same strategy after space time and internal algebra are performed. Another result of this contribution is to show that if the surface terms are not set to zero they will contaminate the renormalization group coefficients. Thus, we are forced to adopt CIR which is equivalent to demand momentum routing invariance

  2. Space - A unique environment for process modeling R&D

    Science.gov (United States)

    Overfelt, Tony

    1991-01-01

    Process modeling, the application of advanced computational techniques to simulate real processes as they occur in regular use, e.g., welding, casting and semiconductor crystal growth, is discussed. Using the low-gravity environment of space will accelerate the technical validation of the procedures and enable extremely accurate determinations of the many necessary thermophysical properties. Attention is given to NASA's centers for the commercial development of space; joint ventures of universities, industries, and goverment agencies to study the unique attributes of space that offer potential for applied R&D and eventual commercial exploitation.

  3. Sparse structure regularized ranking

    KAUST Repository

    Wang, Jim Jing-Yan; Sun, Yijun; Gao, Xin

    2014-01-01

    Learning ranking scores is critical for the multimedia database retrieval problem. In this paper, we propose a novel ranking score learning algorithm by exploring the sparse structure and using it to regularize ranking scores. To explore the sparse

  4. Phaseless tomographic inverse scattering in Banach spaces

    International Nuclear Information System (INIS)

    Estatico, C.; Fedeli, A.; Pastorino, M.; Randazzo, A.; Tavanti, E.

    2016-01-01

    In conventional microwave imaging, a hidden dielectric object under test is illuminated by microwave incident waves and the field it scatters is measured in magnitude and phase in order to retrieve the dielectric properties by solving the related non-homogenous Helmholtz equation or its Lippmann-Schwinger integral formulation. Since the measurement of the phase of electromagnetic waves can be still considered expensive in real applications, in this paper only the magnitude of the scattering wave fields is measured in order to allow a reduction of the cost of the measurement apparatus. In this respect, we firstly analyse the properties of the phaseless scattering nonlinear forward modelling operator in its integral form and we provide an analytical expression for computing its Fréchet derivative. Then, we propose an inexact Newton method to solve the associated nonlinear inverse problems, where any linearized step is solved by a L p Banach space iterative regularization method which acts on the dual space L p* . Indeed, it is well known that regularization in special Banach spaces, such us L p with 1 < p < 2, allows to promote sparsity and to reduce Gibbs phenomena and over-smoothness. Preliminary results concerning numerically computed field data are shown. (paper)

  5. Identifying factors associated with regular physical activity in leisure time among Canadian adolescents.

    Science.gov (United States)

    Godin, Gaston; Anderson, Donna; Lambert, Léo-Daniel; Desharnais, Raymond

    2005-01-01

    The purpose of this study was to identify the factors explaining regular physical activity among Canadian adolescents. A cohort study conducted over a period of 2 years. A French-language high school located near Québec City. A cohort of 740 students (352 girls; 388 boys) aged 13.3 +/- 1.0 years at baseline. Psychosocial, life context, profile, and sociodemographic variables were assessed at baseline and 1 and 2 years after baseline. Exercising almost every day during leisure time at each measurement time was the dependent variable. The Generalized Estimating Equations (GEE) analysis indicated that exercising almost every day was significantly associated with a high intention to exercise (odds ratio [OR]: 8.33, confidence interval [CI] 95%: 5.26, 13.18), being satisfied with the activity practiced (OR: 2.07, CI 95%: 1.27, 3.38), perceived descriptive norm (OR: 1.82, CI 95%: 1.41, 2.35), being a boy (OR: 1.83, CI 95%: 1.37, 2.46), practicing "competitive" activities (OR: 1.80, CI 95%: 1.37, 2.36), eating a healthy breakfast (OR: 1.68, CI 95%: 1.09, 2.60), and normative beliefs (OR: 1.48, CI 95%: 1.14, 1.90). Specific GEE analysis for gender indicated slight but significant differences. This study provides evidence for the need to design interventions that are gender specific and that focus on increasing intention to exercise regularly.

  6. Generalized Confidence Intervals and Fiducial Intervals for Some Epidemiological Measures

    Directory of Open Access Journals (Sweden)

    Ionut Bebu

    2016-06-01

    Full Text Available For binary outcome data from epidemiological studies, this article investigates the interval estimation of several measures of interest in the absence or presence of categorical covariates. When covariates are present, the logistic regression model as well as the log-binomial model are investigated. The measures considered include the common odds ratio (OR from several studies, the number needed to treat (NNT, and the prevalence ratio. For each parameter, confidence intervals are constructed using the concepts of generalized pivotal quantities and fiducial quantities. Numerical results show that the confidence intervals so obtained exhibit satisfactory performance in terms of maintaining the coverage probabilities even when the sample sizes are not large. An appealing feature of the proposed solutions is that they are not based on maximization of the likelihood, and hence are free from convergence issues associated with the numerical calculation of the maximum likelihood estimators, especially in the context of the log-binomial model. The results are illustrated with a number of examples. The overall conclusion is that the proposed methodologies based on generalized pivotal quantities and fiducial quantities provide an accurate and unified approach for the interval estimation of the various epidemiological measures in the context of binary outcome data with or without covariates.

  7. Sparse structure regularized ranking

    KAUST Repository

    Wang, Jim Jing-Yan

    2014-04-17

    Learning ranking scores is critical for the multimedia database retrieval problem. In this paper, we propose a novel ranking score learning algorithm by exploring the sparse structure and using it to regularize ranking scores. To explore the sparse structure, we assume that each multimedia object could be represented as a sparse linear combination of all other objects, and combination coefficients are regarded as a similarity measure between objects and used to regularize their ranking scores. Moreover, we propose to learn the sparse combination coefficients and the ranking scores simultaneously. A unified objective function is constructed with regard to both the combination coefficients and the ranking scores, and is optimized by an iterative algorithm. Experiments on two multimedia database retrieval data sets demonstrate the significant improvements of the propose algorithm over state-of-the-art ranking score learning algorithms.

  8. 20 CFR 226.35 - Deductions from regular annuity rate.

    Science.gov (United States)

    2010-04-01

    ... 20 Employees' Benefits 1 2010-04-01 2010-04-01 false Deductions from regular annuity rate. 226.35... COMPUTING EMPLOYEE, SPOUSE, AND DIVORCED SPOUSE ANNUITIES Computing a Spouse or Divorced Spouse Annuity § 226.35 Deductions from regular annuity rate. The regular annuity rate of the spouse and divorced...

  9. Downscaling Satellite Precipitation with Emphasis on Extremes: A Variational 1-Norm Regularization in the Derivative Domain

    Science.gov (United States)

    Foufoula-Georgiou, E.; Ebtehaj, A. M.; Zhang, S. Q.; Hou, A. Y.

    2013-01-01

    The increasing availability of precipitation observations from space, e.g., from the Tropical Rainfall Measuring Mission (TRMM) and the forthcoming Global Precipitation Measuring (GPM) Mission, has fueled renewed interest in developing frameworks for downscaling and multi-sensor data fusion that can handle large data sets in computationally efficient ways while optimally reproducing desired properties of the underlying rainfall fields. Of special interest is the reproduction of extreme precipitation intensities and gradients, as these are directly relevant to hazard prediction. In this paper, we present a new formalism for downscaling satellite precipitation observations, which explicitly allows for the preservation of some key geometrical and statistical properties of spatial precipitation. These include sharp intensity gradients (due to high-intensity regions embedded within lower-intensity areas), coherent spatial structures (due to regions of slowly varying rainfall),and thicker-than-Gaussian tails of precipitation gradients and intensities. Specifically, we pose the downscaling problem as a discrete inverse problem and solve it via a regularized variational approach (variational downscaling) where the regularization term is selected to impose the desired smoothness in the solution while allowing for some steep gradients(called 1-norm or total variation regularization). We demonstrate the duality between this geometrically inspired solution and its Bayesian statistical interpretation, which is equivalent to assuming a Laplace prior distribution for the precipitation intensities in the derivative (wavelet) space. When the observation operator is not known, we discuss the effect of its misspecification and explore a previously proposed dictionary-based sparse inverse downscaling methodology to indirectly learn the observation operator from a database of coincidental high- and low-resolution observations. The proposed method and ideas are illustrated in case

  10. General Relativity without paradigm of space-time covariance, and resolution of the problem of time

    Science.gov (United States)

    Soo, Chopin; Yu, Hoi-Lai

    2014-01-01

    The framework of a theory of gravity from the quantum to the classical regime is presented. The paradigm shift from full space-time covariance to spatial diffeomorphism invariance, together with clean decomposition of the canonical structure, yield transparent physical dynamics and a resolution of the problem of time. The deep divide between quantum mechanics and conventional canonical formulations of quantum gravity is overcome with a Schrödinger equation for quantum geometrodynamics that describes evolution in intrinsic time. Unitary time development with gauge-invariant temporal ordering is also viable. All Kuchar observables become physical; and classical space-time, with direct correlation between its proper times and intrinsic time intervals, emerges from constructive interference. The framework not only yields a physical Hamiltonian for Einstein's theory, but also prompts natural extensions and improvements towards a well behaved quantum theory of gravity. It is a consistent canonical scheme to discuss Horava-Lifshitz theories with intrinsic time evolution, and of the many possible alternatives that respect 3-covariance (rather than the more restrictive 4-covariance of Einstein's theory), Horava's "detailed balance" form of the Hamiltonian constraint is essentially pinned down by this framework. Issues in quantum gravity that depend on radiative corrections and the rigorous definition and regularization of the Hamiltonian operator are not addressed in this work.

  11. Regularization theory for ill-posed problems selected topics

    CERN Document Server

    Lu, Shuai

    2013-01-01

    Thismonograph is a valuable contribution to thehighly topical and extremly productive field ofregularisationmethods for inverse and ill-posed problems. The author is an internationally outstanding and acceptedmathematicianin this field. In his book he offers a well-balanced mixtureof basic and innovative aspects.He demonstrates new,differentiatedviewpoints, and important examples for applications. The bookdemontrates thecurrent developments inthe field of regularization theory,such as multiparameter regularization and regularization in learning theory. The book is written for graduate and PhDs

  12. Finite Metric Spaces of Strictly Negative Type

    DEFF Research Database (Denmark)

    Hjorth, Poul; Lisonek, P.; Markvorsen, Steen

    1998-01-01

    of Euclidean spaces. We prove that, if the distance matrix is both hypermetric and regular, then it is of strictly negative type. We show that the strictly negative type finite subspaces of spheres are precisely those which do not contain two pairs of antipodal points. In connection with an open problem raised...

  13. 20 CFR 226.34 - Divorced spouse regular annuity rate.

    Science.gov (United States)

    2010-04-01

    ... 20 Employees' Benefits 1 2010-04-01 2010-04-01 false Divorced spouse regular annuity rate. 226.34... COMPUTING EMPLOYEE, SPOUSE, AND DIVORCED SPOUSE ANNUITIES Computing a Spouse or Divorced Spouse Annuity § 226.34 Divorced spouse regular annuity rate. The regular annuity rate of a divorced spouse is equal to...

  14. Chimeric mitochondrial peptides from contiguous regular and swinger RNA.

    Science.gov (United States)

    Seligmann, Hervé

    2016-01-01

    Previous mass spectrometry analyses described human mitochondrial peptides entirely translated from swinger RNAs, RNAs where polymerization systematically exchanged nucleotides. Exchanges follow one among 23 bijective transformation rules, nine symmetric exchanges (X ↔ Y, e.g. A ↔ C) and fourteen asymmetric exchanges (X → Y → Z → X, e.g. A → C → G → A), multiplying by 24 DNA's protein coding potential. Abrupt switches from regular to swinger polymerization produce chimeric RNAs. Here, human mitochondrial proteomic analyses assuming abrupt switches between regular and swinger transcriptions, detect chimeric peptides, encoded by part regular, part swinger RNA. Contiguous regular- and swinger-encoded residues within single peptides are stronger evidence for translation of swinger RNA than previously detected, entirely swinger-encoded peptides: regular parts are positive controls matched with contiguous swinger parts, increasing confidence in results. Chimeric peptides are 200 × rarer than swinger peptides (3/100,000 versus 6/1000). Among 186 peptides with > 8 residues for each regular and swinger parts, regular parts of eleven chimeric peptides correspond to six among the thirteen recognized, mitochondrial protein-coding genes. Chimeric peptides matching partly regular proteins are rarer and less expressed than chimeric peptides matching non-coding sequences, suggesting targeted degradation of misfolded proteins. Present results strengthen hypotheses that the short mitogenome encodes far more proteins than hitherto assumed. Entirely swinger-encoded proteins could exist.

  15. Dental plaque pH variation with regular soft drink, diet soft drink and high energy drink: an in vivo study.

    Science.gov (United States)

    Jawale, Bhushan Arun; Bendgude, Vikas; Mahuli, Amit V; Dave, Bhavana; Kulkarni, Harshal; Mittal, Simpy

    2012-03-01

    A high incidence of dental caries and dental erosion associated with frequent consumption of soft drinks has been reported. The purpose of this study was to evaluate the pH response of dental plaque to a regular, diet and high energy drink. Twenty subjects were recruited for this study. All subjects were between the ages of 20 and 25 and had at least four restored tooth surfaces present. The subjects were asked to refrain from brushing for 48 hours prior to the study. At baseline, plaque pH was measured from four separate locations using harvesting method. Subjects were asked to swish with 15 ml of the respective soft drink for 1 minute. Plaque pH was measured at the four designated tooth sites at 5, 10 and 20 minutes intervals. Subjects then repeated the experiment using the other two soft drinks. pH was minimum for regular soft drink (2.65 ± 0.026) followed by high energy drink (3.39 ± 0.026) and diet soft drink (3.78 ± 0.006). The maximum drop in plaque pH was seen with regular soft drink followed by high energy drink and diet soft drink. Regular soft drink possesses a greater acid challenge potential on enamel than diet and high energy soft drinks. However, in this clinical trial, the pH associated with either soft drink did not reach the critical pH which is expected for enamel demineralization and dissolution.

  16. On the minimizers of calculus of variations problems in Hilbert spaces

    KAUST Repository

    Gomes, Diogo A.

    2014-01-19

    The objective of this paper is to discuss existence, uniqueness and regularity issues of minimizers of one dimensional calculus of variations problem in Hilbert spaces. © 2014 Springer-Verlag Berlin Heidelberg.

  17. On the minimizers of calculus of variations problems in Hilbert spaces

    KAUST Repository

    Gomes, Diogo A.; Nurbekyan, Levon

    2014-01-01

    The objective of this paper is to discuss existence, uniqueness and regularity issues of minimizers of one dimensional calculus of variations problem in Hilbert spaces. © 2014 Springer-Verlag Berlin Heidelberg.

  18. Implicit Regularization for Reconstructing 3D Building Rooftop Models Using Airborne LiDAR Data

    Directory of Open Access Journals (Sweden)

    Jaewook Jung

    2017-03-01

    Full Text Available With rapid urbanization, highly accurate and semantically rich virtualization of building assets in 3D become more critical for supporting various applications, including urban planning, emergency response and location-based services. Many research efforts have been conducted to automatically reconstruct building models at city-scale from remotely sensed data. However, developing a fully-automated photogrammetric computer vision system enabling the massive generation of highly accurate building models still remains a challenging task. One the most challenging task for 3D building model reconstruction is to regularize the noises introduced in the boundary of building object retrieved from a raw data with lack of knowledge on its true shape. This paper proposes a data-driven modeling approach to reconstruct 3D rooftop models at city-scale from airborne laser scanning (ALS data. The focus of the proposed method is to implicitly derive the shape regularity of 3D building rooftops from given noisy information of building boundary in a progressive manner. This study covers a full chain of 3D building modeling from low level processing to realistic 3D building rooftop modeling. In the element clustering step, building-labeled point clouds are clustered into homogeneous groups by applying height similarity and plane similarity. Based on segmented clusters, linear modeling cues including outer boundaries, intersection lines, and step lines are extracted. Topology elements among the modeling cues are recovered by the Binary Space Partitioning (BSP technique. The regularity of the building rooftop model is achieved by an implicit regularization process in the framework of Minimum Description Length (MDL combined with Hypothesize and Test (HAT. The parameters governing the MDL optimization are automatically estimated based on Min-Max optimization and Entropy-based weighting method. The performance of the proposed method is tested over the International

  19. Implicit Regularization for Reconstructing 3D Building Rooftop Models Using Airborne LiDAR Data.

    Science.gov (United States)

    Jung, Jaewook; Jwa, Yoonseok; Sohn, Gunho

    2017-03-19

    With rapid urbanization, highly accurate and semantically rich virtualization of building assets in 3D become more critical for supporting various applications, including urban planning, emergency response and location-based services. Many research efforts have been conducted to automatically reconstruct building models at city-scale from remotely sensed data. However, developing a fully-automated photogrammetric computer vision system enabling the massive generation of highly accurate building models still remains a challenging task. One the most challenging task for 3D building model reconstruction is to regularize the noises introduced in the boundary of building object retrieved from a raw data with lack of knowledge on its true shape. This paper proposes a data-driven modeling approach to reconstruct 3D rooftop models at city-scale from airborne laser scanning (ALS) data. The focus of the proposed method is to implicitly derive the shape regularity of 3D building rooftops from given noisy information of building boundary in a progressive manner. This study covers a full chain of 3D building modeling from low level processing to realistic 3D building rooftop modeling. In the element clustering step, building-labeled point clouds are clustered into homogeneous groups by applying height similarity and plane similarity. Based on segmented clusters, linear modeling cues including outer boundaries, intersection lines, and step lines are extracted. Topology elements among the modeling cues are recovered by the Binary Space Partitioning (BSP) technique. The regularity of the building rooftop model is achieved by an implicit regularization process in the framework of Minimum Description Length (MDL) combined with Hypothesize and Test (HAT). The parameters governing the MDL optimization are automatically estimated based on Min-Max optimization and Entropy-based weighting method. The performance of the proposed method is tested over the International Society for

  20. Reliable computation of roots in analytical waveguide modeling using an interval-Newton approach and algorithmic differentiation.

    Science.gov (United States)

    Bause, Fabian; Walther, Andrea; Rautenberg, Jens; Henning, Bernd

    2013-12-01

    For the modeling and simulation of wave propagation in geometrically simple waveguides such as plates or rods, one may employ the analytical global matrix method. That is, a certain (global) matrix depending on the two parameters wavenumber and frequency is built. Subsequently, one must calculate all parameter pairs within the domain of interest where the global matrix becomes singular. For this purpose, one could compute all roots of the determinant of the global matrix when the two parameters vary in the given intervals. This requirement to calculate all roots is actually the method's most concerning restriction. Previous approaches are based on so-called mode-tracers, which use the physical phenomenon that solutions, i.e., roots of the determinant of the global matrix, appear in a certain pattern, the waveguide modes, to limit the root-finding algorithm's search space with respect to consecutive solutions. In some cases, these reductions of the search space yield only an incomplete set of solutions, because some roots may be missed as a result of uncertain predictions. Therefore, we propose replacement of the mode-tracer approach with a suitable version of an interval- Newton method. To apply this interval-based method, we extended the interval and derivative computation provided by a numerical computing environment such that corresponding information is also available for Bessel functions used in circular models of acoustic waveguides. We present numerical results for two different scenarios. First, a polymeric cylindrical waveguide is simulated, and second, we show simulation results of a one-sided fluid-loaded plate. For both scenarios, we compare results obtained with the proposed interval-Newton algorithm and commercial software.

  1. Dimensional regularization and analytical continuation at finite temperature

    International Nuclear Information System (INIS)

    Chen Xiangjun; Liu Lianshou

    1998-01-01

    The relationship between dimensional regularization and analytical continuation of infrared divergent integrals at finite temperature is discussed and a method of regularization of infrared divergent integrals and infrared divergent sums is given

  2. Current situation of fish farming in Togo

    African Journals Online (AJOL)

    practiced in ponds and water reservoirs and the semi intensive system (12.09%) implemented in tanks and ponds .... primarily intended to be a source of drinking water ... big fish at regular intervals. .... has the advantage of requiring little space.

  3. On Bayesian treatment of systematic uncertainties in confidence interval calculation

    CERN Document Server

    Tegenfeldt, Fredrik

    2005-01-01

    In high energy physics, a widely used method to treat systematic uncertainties in confidence interval calculations is based on combining a frequentist construction of confidence belts with a Bayesian treatment of systematic uncertainties. In this note we present a study of the coverage of this method for the standard Likelihood Ratio (aka Feldman & Cousins) construction for a Poisson process with known background and Gaussian or log-Normal distributed uncertainties in the background or signal efficiency. For uncertainties in the signal efficiency of upto 40 % we find over-coverage on the level of 2 to 4 % depending on the size of uncertainties and the region in signal space. Uncertainties in the background generally have smaller effect on the coverage. A considerable smoothing of the coverage curves is observed. A software package is presented which allows fast calculation of the confidence intervals for a variety of assumptions on shape and size of systematic uncertainties for different nuisance paramete...

  4. Regular and conformal regular cores for static and rotating solutions

    Energy Technology Data Exchange (ETDEWEB)

    Azreg-Aïnou, Mustapha

    2014-03-07

    Using a new metric for generating rotating solutions, we derive in a general fashion the solution of an imperfect fluid and that of its conformal homolog. We discuss the conditions that the stress–energy tensors and invariant scalars be regular. On classical physical grounds, it is stressed that conformal fluids used as cores for static or rotating solutions are exempt from any malicious behavior in that they are finite and defined everywhere.

  5. Regular and conformal regular cores for static and rotating solutions

    International Nuclear Information System (INIS)

    Azreg-Aïnou, Mustapha

    2014-01-01

    Using a new metric for generating rotating solutions, we derive in a general fashion the solution of an imperfect fluid and that of its conformal homolog. We discuss the conditions that the stress–energy tensors and invariant scalars be regular. On classical physical grounds, it is stressed that conformal fluids used as cores for static or rotating solutions are exempt from any malicious behavior in that they are finite and defined everywhere.

  6. Interval stability for complex systems

    Science.gov (United States)

    Klinshov, Vladimir V.; Kirillov, Sergey; Kurths, Jürgen; Nekorkin, Vladimir I.

    2018-04-01

    Stability of dynamical systems against strong perturbations is an important problem of nonlinear dynamics relevant to many applications in various areas. Here, we develop a novel concept of interval stability, referring to the behavior of the perturbed system during a finite time interval. Based on this concept, we suggest new measures of stability, namely interval basin stability (IBS) and interval stability threshold (IST). IBS characterizes the likelihood that the perturbed system returns to the stable regime (attractor) in a given time. IST provides the minimal magnitude of the perturbation capable to disrupt the stable regime for a given interval of time. The suggested measures provide important information about the system susceptibility to external perturbations which may be useful for practical applications. Moreover, from a theoretical viewpoint the interval stability measures are shown to bridge the gap between linear and asymptotic stability. We also suggest numerical algorithms for quantification of the interval stability characteristics and demonstrate their potential for several dynamical systems of various nature, such as power grids and neural networks.

  7. Low-rank matrix approximation with manifold regularization.

    Science.gov (United States)

    Zhang, Zhenyue; Zhao, Keke

    2013-07-01

    This paper proposes a new model of low-rank matrix factorization that incorporates manifold regularization to the matrix factorization. Superior to the graph-regularized nonnegative matrix factorization, this new regularization model has globally optimal and closed-form solutions. A direct algorithm (for data with small number of points) and an alternate iterative algorithm with inexact inner iteration (for large scale data) are proposed to solve the new model. A convergence analysis establishes the global convergence of the iterative algorithm. The efficiency and precision of the algorithm are demonstrated numerically through applications to six real-world datasets on clustering and classification. Performance comparison with existing algorithms shows the effectiveness of the proposed method for low-rank factorization in general.

  8. Aircraft Configuration and Flight Crew Compliance with Procedures While Conducting Flight Deck Based Interval Management (FIM) Operations

    Science.gov (United States)

    Shay, Rick; Swieringa, Kurt A.; Baxley, Brian T.

    2012-01-01

    Flight deck based Interval Management (FIM) applications using ADS-B are being developed to improve both the safety and capacity of the National Airspace System (NAS). FIM is expected to improve the safety and efficiency of the NAS by giving pilots the technology and procedures to precisely achieve an interval behind the preceding aircraft by a specific point. Concurrently but independently, Optimized Profile Descents (OPD) are being developed to help reduce fuel consumption and noise, however, the range of speeds available when flying an OPD results in a decrease in the delivery precision of aircraft to the runway. This requires the addition of a spacing buffer between aircraft, reducing system throughput. FIM addresses this problem by providing pilots with speed guidance to achieve a precise interval behind another aircraft, even while flying optimized descents. The Interval Management with Spacing to Parallel Dependent Runways (IMSPiDR) human-in-the-loop experiment employed 24 commercial pilots to explore the use of FIM equipment to conduct spacing operations behind two aircraft arriving to parallel runways, while flying an OPD during high-density operations. This paper describes the impact of variations in pilot operations; in particular configuring the aircraft, their compliance with FIM operating procedures, and their response to changes of the FIM speed. An example of the displayed FIM speeds used incorrectly by a pilot is also discussed. Finally, this paper examines the relationship between achieving airline operational goals for individual aircraft and the need for ATC to deliver aircraft to the runway with greater precision. The results show that aircraft can fly an OPD and conduct FIM operations to dependent parallel runways, enabling operational goals to be achieved efficiently while maintaining system throughput.

  9. Regular-fat dairy and human health

    DEFF Research Database (Denmark)

    Astrup, Arne; Bradley, Beth H Rice; Brenna, J Thomas

    2016-01-01

    In recent history, some dietary recommendations have treated dairy fat as an unnecessary source of calories and saturated fat in the human diet. These assumptions, however, have recently been brought into question by current research on regular fat dairy products and human health. In an effort to......, cheese and yogurt, can be important components of an overall healthy dietary pattern. Systematic examination of the effects of dietary patterns that include regular-fat milk, cheese and yogurt on human health is warranted....

  10. Bounded Perturbation Regularization for Linear Least Squares Estimation

    KAUST Repository

    Ballal, Tarig; Suliman, Mohamed Abdalla Elhag; Al-Naffouri, Tareq Y.

    2017-01-01

    This paper addresses the problem of selecting the regularization parameter for linear least-squares estimation. We propose a new technique called bounded perturbation regularization (BPR). In the proposed BPR method, a perturbation with a bounded

  11. Clustering, randomness, and regularity in cloud fields: 2. Cumulus cloud fields

    Science.gov (United States)

    Zhu, T.; Lee, J.; Weger, R. C.; Welch, R. M.

    1992-12-01

    During the last decade a major controversy has been brewing concerning the proper characterization of cumulus convection. The prevailing view has been that cumulus clouds form in clusters, in which cloud spacing is closer than that found for the overall cloud field and which maintains its identity over many cloud lifetimes. This "mutual protection hypothesis" of Randall and Huffman (1980) has been challenged by the "inhibition hypothesis" of Ramirez et al. (1990) which strongly suggests that the spatial distribution of cumuli must tend toward a regular distribution. A dilemma has resulted because observations have been reported to support both hypotheses. The present work reports a detailed analysis of cumulus cloud field spatial distributions based upon Landsat, Advanced Very High Resolution Radiometer, and Skylab data. Both nearest-neighbor and point-to-cloud cumulative distribution function statistics are investigated. The results show unequivocally that when both large and small clouds are included in the cloud field distribution, the cloud field always has a strong clustering signal. The strength of clustering is largest at cloud diameters of about 200-300 m, diminishing with increasing cloud diameter. In many cases, clusters of small clouds are found which are not closely associated with large clouds. As the small clouds are eliminated from consideration, the cloud field typically tends towards regularity. Thus it would appear that the "inhibition hypothesis" of Ramirez and Bras (1990) has been verified for the large clouds. However, these results are based upon the analysis of point processes. A more exact analysis also is made which takes into account the cloud size distributions. Since distinct clouds are by definition nonoverlapping, cloud size effects place a restriction upon the possible locations of clouds in the cloud field. The net effect of this analysis is that the large clouds appear to be randomly distributed, with only weak tendencies towards

  12. Recognition Memory for Novel Stimuli: The Structural Regularity Hypothesis

    Science.gov (United States)

    Cleary, Anne M.; Morris, Alison L.; Langley, Moses M.

    2007-01-01

    Early studies of human memory suggest that adherence to a known structural regularity (e.g., orthographic regularity) benefits memory for an otherwise novel stimulus (e.g., G. A. Miller, 1958). However, a more recent study suggests that structural regularity can lead to an increase in false-positive responses on recognition memory tests (B. W. A.…

  13. Statistical intervals a guide for practitioners

    CERN Document Server

    Hahn, Gerald J

    2011-01-01

    Presents a detailed exposition of statistical intervals and emphasizes applications in industry. The discussion differentiates at an elementary level among different kinds of statistical intervals and gives instruction with numerous examples and simple math on how to construct such intervals from sample data. This includes confidence intervals to contain a population percentile, confidence intervals on probability of meeting specified threshold value, and prediction intervals to include observation in a future sample. Also has an appendix containing computer subroutines for nonparametric stati

  14. Statistics of return intervals between long heartbeat intervals and their usability for online prediction of disorders

    International Nuclear Information System (INIS)

    Bogachev, Mikhail I; Bunde, Armin; Kireenkov, Igor S; Nifontov, Eugene M

    2009-01-01

    We study the statistics of return intervals between large heartbeat intervals (above a certain threshold Q) in 24 h records obtained from healthy subjects. We find that both the linear and the nonlinear long-term memory inherent in the heartbeat intervals lead to power-laws in the probability density function P Q (r) of the return intervals. As a consequence, the probability W Q (t; Δt) that at least one large heartbeat interval will occur within the next Δt heartbeat intervals, with an increasing elapsed number of intervals t after the last large heartbeat interval, follows a power-law. Based on these results, we suggest a method of obtaining a priori information about the occurrence of the next large heartbeat interval, and thus to predict it. We show explicitly that the proposed method, which exploits long-term memory, is superior to the conventional precursory pattern recognition technique, which focuses solely on short-term memory. We believe that our results can be straightforwardly extended to obtain more reliable predictions in other physiological signals like blood pressure, as well as in other complex records exhibiting multifractal behaviour, e.g. turbulent flow, precipitation, river flows and network traffic.

  15. Regularization Techniques for Linear Least-Squares Problems

    KAUST Repository

    Suliman, Mohamed

    2016-04-01

    Linear estimation is a fundamental branch of signal processing that deals with estimating the values of parameters from a corrupted measured data. Throughout the years, several optimization criteria have been used to achieve this task. The most astonishing attempt among theses is the linear least-squares. Although this criterion enjoyed a wide popularity in many areas due to its attractive properties, it appeared to suffer from some shortcomings. Alternative optimization criteria, as a result, have been proposed. These new criteria allowed, in one way or another, the incorporation of further prior information to the desired problem. Among theses alternative criteria is the regularized least-squares (RLS). In this thesis, we propose two new algorithms to find the regularization parameter for linear least-squares problems. In the constrained perturbation regularization algorithm (COPRA) for random matrices and COPRA for linear discrete ill-posed problems, an artificial perturbation matrix with a bounded norm is forced into the model matrix. This perturbation is introduced to enhance the singular value structure of the matrix. As a result, the new modified model is expected to provide a better stabilize substantial solution when used to estimate the original signal through minimizing the worst-case residual error function. Unlike many other regularization algorithms that go in search of minimizing the estimated data error, the two new proposed algorithms are developed mainly to select the artifcial perturbation bound and the regularization parameter in a way that approximately minimizes the mean-squared error (MSE) between the original signal and its estimate under various conditions. The first proposed COPRA method is developed mainly to estimate the regularization parameter when the measurement matrix is complex Gaussian, with centered unit variance (standard), and independent and identically distributed (i.i.d.) entries. Furthermore, the second proposed COPRA

  16. Regularized Regression and Density Estimation based on Optimal Transport

    KAUST Repository

    Burger, M.

    2012-03-11

    The aim of this paper is to investigate a novel nonparametric approach for estimating and smoothing density functions as well as probability densities from discrete samples based on a variational regularization method with the Wasserstein metric as a data fidelity. The approach allows a unified treatment of discrete and continuous probability measures and is hence attractive for various tasks. In particular, the variational model for special regularization functionals yields a natural method for estimating densities and for preserving edges in the case of total variation regularization. In order to compute solutions of the variational problems, a regularized optimal transport problem needs to be solved, for which we discuss several formulations and provide a detailed analysis. Moreover, we compute special self-similar solutions for standard regularization functionals and we discuss several computational approaches and results. © 2012 The Author(s).

  17. The Analysis of Two-Way Functional Data Using Two-Way Regularized Singular Value Decompositions

    KAUST Repository

    Huang, Jianhua Z.

    2009-12-01

    Two-way functional data consist of a data matrix whose row and column domains are both structured, for example, temporally or spatially, as when the data are time series collected at different locations in space. We extend one-way functional principal component analysis (PCA) to two-way functional data by introducing regularization of both left and right singular vectors in the singular value decomposition (SVD) of the data matrix. We focus on a penalization approach and solve the nontrivial problem of constructing proper two-way penalties from oneway regression penalties. We introduce conditional cross-validated smoothing parameter selection whereby left-singular vectors are cross- validated conditional on right-singular vectors, and vice versa. The concept can be realized as part of an alternating optimization algorithm. In addition to the penalization approach, we briefly consider two-way regularization with basis expansion. The proposed methods are illustrated with one simulated and two real data examples. Supplemental materials available online show that several "natural" approaches to penalized SVDs are flawed and explain why so. © 2009 American Statistical Association.

  18. Architecture of interstitial nodal spaces in the rodent renal inner medulla.

    Science.gov (United States)

    Gilbert, Rebecca L; Pannabecker, Thomas L

    2013-09-01

    Every collecting duct (CD) of the rat inner medulla is uniformly surrounded by about four abutting ascending vasa recta (AVR) running parallel to it. One or two ascending thin limbs (ATLs) lie between and parallel to each abutting AVR pair, opposite the CD. These structures form boundaries of axially running interstitial compartments. Viewed in transverse sections, these compartments appear as four interstitial nodal spaces (INSs) positioned symmetrically around each CD. The axially running compartments are segmented by interstitial cells spaced at regular intervals. The pairing of ATLs and CDs bounded by an abundant supply of AVR carrying reabsorbed water, NaCl, and urea make a strong argument that the mixing of NaCl and urea within the INSs and countercurrent flows play a critical role in generating the inner medullary osmotic gradient. The results of this study fully support that hypothesis. We quantified interactions of all structures comprising INSs along the corticopapillary axis for two rodent species, the Munich-Wistar rat and the kangaroo rat. The results showed remarkable similarities in the configurations of INSs, suggesting that the structural arrangement of INSs is a highly conserved architecture that plays a fundamental role in renal function. The number density of INSs along the corticopapillary axis directly correlated with a loop population that declines exponentially with distance below the outer medullary-inner medullary boundary. The axial configurations were consistent with discrete association between near-bend loop segments and INSs and with upper loop segments lying distant from INSs.

  19. Energy functions for regularization algorithms

    Science.gov (United States)

    Delingette, H.; Hebert, M.; Ikeuchi, K.

    1991-01-01

    Regularization techniques are widely used for inverse problem solving in computer vision such as surface reconstruction, edge detection, or optical flow estimation. Energy functions used for regularization algorithms measure how smooth a curve or surface is, and to render acceptable solutions these energies must verify certain properties such as invariance with Euclidean transformations or invariance with parameterization. The notion of smoothness energy is extended here to the notion of a differential stabilizer, and it is shown that to void the systematic underestimation of undercurvature for planar curve fitting, it is necessary that circles be the curves of maximum smoothness. A set of stabilizers is proposed that meet this condition as well as invariance with rotation and parameterization.

  20. Three regularities of recognition memory: the role of bias.

    Science.gov (United States)

    Hilford, Andrew; Maloney, Laurence T; Glanzer, Murray; Kim, Kisok

    2015-12-01

    A basic assumption of Signal Detection Theory is that decisions are made on the basis of likelihood ratios. In a preceding paper, Glanzer, Hilford, and Maloney (Psychonomic Bulletin & Review, 16, 431-455, 2009) showed that the likelihood ratio assumption implies that three regularities will occur in recognition memory: (1) the Mirror Effect, (2) the Variance Effect, (3) the normalized Receiver Operating Characteristic (z-ROC) Length Effect. The paper offered formal proofs and computational demonstrations that decisions based on likelihood ratios produce the three regularities. A survey of data based on group ROCs from 36 studies validated the likelihood ratio assumption by showing that its three implied regularities are ubiquitous. The study noted, however, that bias, another basic factor in Signal Detection Theory, can obscure the Mirror Effect. In this paper we examine how bias affects the regularities at the theoretical level. The theoretical analysis shows: (1) how bias obscures the Mirror Effect, not the other two regularities, and (2) four ways to counter that obscuring. We then report the results of five experiments that support the theoretical analysis. The analyses and the experimental results also demonstrate: (1) that the three regularities govern individual, as well as group, performance, (2) alternative explanations of the regularities are ruled out, and (3) that Signal Detection Theory, correctly applied, gives a simple and unified explanation of recognition memory data.

  1. Hydrogen atom in the phase-space formulation of quantum mechanics

    International Nuclear Information System (INIS)

    Gracia-Bondia, J.M.

    1984-01-01

    Using a coordinate transformation which regularizes the classical Kepler problem, we show that the hydrogen-atom case may be analytically solved via the phase-space formulation of nonrelativistic quantum mechanics. The problem is essentially reduced to that of a four-dimensional oscillator whose treatment in the phase-space formulation is developed. Furthermore, the method allows us to calculate the Green's function for the H atom in a surprisingly simple way

  2. Two-sorted Point-Interval Temporal Logics

    DEFF Research Database (Denmark)

    Balbiani, Philippe; Goranko, Valentin; Sciavicco, Guido

    2011-01-01

    There are two natural and well-studied approaches to temporal ontology and reasoning: point-based and interval-based. Usually, interval-based temporal reasoning deals with points as particular, duration-less intervals. Here we develop explicitly two-sorted point-interval temporal logical framework...... whereby time instants (points) and time periods (intervals) are considered on a par, and the perspective can shift between them within the formal discourse. We focus on fragments involving only modal operators that correspond to the inter-sort relations between points and intervals. We analyze...

  3. Method of transferring regular shaped vessel into cell

    International Nuclear Information System (INIS)

    Murai, Tsunehiko.

    1997-01-01

    The present invention concerns a method of transferring regular shaped vessels from a non-contaminated area to a contaminated cell. A passage hole for allowing the regular shaped vessels to pass in the longitudinal direction is formed to a partitioning wall at the bottom of the contaminated cell. A plurality of regular shaped vessel are stacked in multiple stages in a vertical direction from the non-contaminated area present below the passage hole, allowed to pass while being urged and transferred successively into the contaminated cell. As a result, since they are transferred while substantially closing the passage hole by the regular shaped vessels, radiation rays or contaminated materials are prevented from discharging from the contaminated cell to the non-contaminated area. Since there is no requirement to open/close an isolation door frequently, the workability upon transfer can be improved remarkably. In addition, the sealing member for sealing the gap between the regular shaped vessel passing through the passage hole and the partitioning wall of the bottom is disposed to the passage hole, the contaminated materials in the contaminated cells can be prevented from discharging from the gap to the non-contaminated area. (N.H.)

  4. High-intensity interval training: Modulating interval duration in overweight/obese men.

    Science.gov (United States)

    Smith-Ryan, Abbie E; Melvin, Malia N; Wingfield, Hailee L

    2015-05-01

    High-intensity interval training (HIIT) is a time-efficient strategy shown to induce various cardiovascular and metabolic adaptations. Little is known about the optimal tolerable combination of intensity and volume necessary for adaptations, especially in clinical populations. In a randomized controlled pilot design, we evaluated the effects of two types of interval training protocols, varying in intensity and interval duration, on clinical outcomes in overweight/obese men. Twenty-five men [body mass index (BMI) > 25 kg · m(2)] completed baseline body composition measures: fat mass (FM), lean mass (LM) and percent body fat (%BF) and fasting blood glucose, lipids and insulin (IN). A graded exercise cycling test was completed for peak oxygen consumption (VO2peak) and power output (PO). Participants were randomly assigned to high-intensity short interval (1MIN-HIIT), high-intensity interval (2MIN-HIIT) or control groups. 1MIN-HIIT and 2MIN-HIIT completed 3 weeks of cycling interval training, 3 days/week, consisting of either 10 × 1 min bouts at 90% PO with 1 min rests (1MIN-HIIT) or 5 × 2 min bouts with 1 min rests at undulating intensities (80%-100%) (2MIN-HIIT). There were no significant training effects on FM (Δ1.06 ± 1.25 kg) or %BF (Δ1.13% ± 1.88%), compared to CON. Increases in LM were not significant but increased by 1.7 kg and 2.1 kg for 1MIN and 2MIN-HIIT groups, respectively. Increases in VO2peak were also not significant for 1MIN (3.4 ml·kg(-1) · min(-1)) or 2MIN groups (2.7 ml · kg(-1) · min(-1)). IN sensitivity (HOMA-IR) improved for both training groups (Δ-2.78 ± 3.48 units; p < 0.05) compared to CON. HIIT may be an effective short-term strategy to improve cardiorespiratory fitness and IN sensitivity in overweight males.

  5. Automatic Constraint Detection for 2D Layout Regularization.

    Science.gov (United States)

    Jiang, Haiyong; Nan, Liangliang; Yan, Dong-Ming; Dong, Weiming; Zhang, Xiaopeng; Wonka, Peter

    2016-08-01

    In this paper, we address the problem of constraint detection for layout regularization. The layout we consider is a set of two-dimensional elements where each element is represented by its bounding box. Layout regularization is important in digitizing plans or images, such as floor plans and facade images, and in the improvement of user-created contents, such as architectural drawings and slide layouts. To regularize a layout, we aim to improve the input by detecting and subsequently enforcing alignment, size, and distance constraints between layout elements. Similar to previous work, we formulate layout regularization as a quadratic programming problem. In addition, we propose a novel optimization algorithm that automatically detects constraints. We evaluate the proposed framework using a variety of input layouts from different applications. Our results demonstrate that our method has superior performance to the state of the art.

  6. Automatic Constraint Detection for 2D Layout Regularization

    KAUST Repository

    Jiang, Haiyong

    2015-09-18

    In this paper, we address the problem of constraint detection for layout regularization. As layout we consider a set of two-dimensional elements where each element is represented by its bounding box. Layout regularization is important for digitizing plans or images, such as floor plans and facade images, and for the improvement of user created contents, such as architectural drawings and slide layouts. To regularize a layout, we aim to improve the input by detecting and subsequently enforcing alignment, size, and distance constraints between layout elements. Similar to previous work, we formulate the layout regularization as a quadratic programming problem. In addition, we propose a novel optimization algorithm to automatically detect constraints. In our results, we evaluate the proposed framework on a variety of input layouts from different applications, which demonstrates our method has superior performance to the state of the art.

  7. Online Manifold Regularization by Dual Ascending Procedure

    Directory of Open Access Journals (Sweden)

    Boliang Sun

    2013-01-01

    Full Text Available We propose a novel online manifold regularization framework based on the notion of duality in constrained optimization. The Fenchel conjugate of hinge functions is a key to transfer manifold regularization from offline to online in this paper. Our algorithms are derived by gradient ascent in the dual function. For practical purpose, we propose two buffering strategies and two sparse approximations to reduce the computational complexity. Detailed experiments verify the utility of our approaches. An important conclusion is that our online MR algorithms can handle the settings where the target hypothesis is not fixed but drifts with the sequence of examples. We also recap and draw connections to earlier works. This paper paves a way to the design and analysis of online manifold regularization algorithms.

  8. Parameter identification in ODE models with oscillatory dynamics: a Fourier regularization approach

    Science.gov (United States)

    Chiara D'Autilia, Maria; Sgura, Ivonne; Bozzini, Benedetto

    2017-12-01

    In this paper we consider a parameter identification problem (PIP) for data oscillating in time, that can be described in terms of the dynamics of some ordinary differential equation (ODE) model, resulting in an optimization problem constrained by the ODEs. In problems with this type of data structure, simple application of the direct method of control theory (discretize-then-optimize) yields a least-squares cost function exhibiting multiple ‘low’ minima. Since in this situation any optimization algorithm is liable to fail in the approximation of a good solution, here we propose a Fourier regularization approach that is able to identify an iso-frequency manifold {{ S}} of codimension-one in the parameter space \

  9. Alternation of regular and chaotic dynamics in a simple two-degree-of-freedom system with nonlinear inertial coupling.

    Science.gov (United States)

    Sigalov, G; Gendelman, O V; AL-Shudeifat, M A; Manevitch, L I; Vakakis, A F; Bergman, L A

    2012-03-01

    We show that nonlinear inertial coupling between a linear oscillator and an eccentric rotator can lead to very interesting interchanges between regular and chaotic dynamical behavior. Indeed, we show that this model demonstrates rather unusual behavior from the viewpoint of nonlinear dynamics. Specifically, at a discrete set of values of the total energy, the Hamiltonian system exhibits non-conventional nonlinear normal modes, whose shape is determined by phase locking of rotatory and oscillatory motions of the rotator at integer ratios of characteristic frequencies. Considering the weakly damped system, resonance capture of the dynamics into the vicinity of these modes brings about regular motion of the system. For energy levels far from these discrete values, the motion of the system is chaotic. Thus, the succession of resonance captures and escapes by a discrete set of the normal modes causes a sequence of transitions between regular and chaotic behavior, provided that the damping is sufficiently small. We begin from the Hamiltonian system and present a series of Poincaré sections manifesting the complex structure of the phase space of the considered system with inertial nonlinear coupling. Then an approximate analytical description is presented for the non-conventional nonlinear normal modes. We confirm the analytical results by numerical simulation and demonstrate the alternate transitions between regular and chaotic dynamics mentioned above. The origin of the chaotic behavior is also discussed.

  10. Limit Formulae and Jump Relations of Potential Theory in Sobolev Spaces

    OpenAIRE

    Raskop, Thomas; Grothaus, Martin

    2009-01-01

    In this article we combine the modern theory of Sobolev spaces with the classical theory of limit formulae and jump relations of potential theory. Also other authors proved the convergence in Lebesgue spaces for integrable functions. The achievement of this paper is the L2 convergence for the weak derivatives of higher orders. Also the layer functions F are elements of Sobolev spaces and a two dimensional suitable smooth submanifold in R3, called regular Cm-surface. We are considering the pot...

  11. Regularized principal covariates regression and its application to finding coupled patterns in climate fields

    Science.gov (United States)

    Fischer, M. J.

    2014-02-01

    There are many different methods for investigating the coupling between two climate fields, which are all based on the multivariate regression model. Each different method of solving the multivariate model has its own attractive characteristics, but often the suitability of a particular method for a particular problem is not clear. Continuum regression methods search the solution space between the conventional methods and thus can find regression model subspaces that mix the attractive characteristics of the end-member subspaces. Principal covariates regression is a continuum regression method that is easily applied to climate fields and makes use of two end-members: principal components regression and redundancy analysis. In this study, principal covariates regression is extended to additionally span a third end-member (partial least squares or maximum covariance analysis). The new method, regularized principal covariates regression, has several attractive features including the following: it easily applies to problems in which the response field has missing values or is temporally sparse, it explores a wide range of model spaces, and it seeks a model subspace that will, for a set number of components, have a predictive skill that is the same or better than conventional regression methods. The new method is illustrated by applying it to the problem of predicting the southern Australian winter rainfall anomaly field using the regional atmospheric pressure anomaly field. Regularized principal covariates regression identifies four major coupled patterns in these two fields. The two leading patterns, which explain over half the variance in the rainfall field, are related to the subtropical ridge and features of the zonally asymmetric circulation.

  12. Lack of spacing effects during piano learning.

    Directory of Open Access Journals (Sweden)

    Melody Wiseheart

    Full Text Available Spacing effects during retention of verbal information are easily obtained, and the effect size is large. Relatively little evidence exists on whether motor skill retention benefits from distributed practice, with even less evidence on complex motor skills. We taught a 17-note musical sequence on a piano to individuals without prior formal training. There were five lags between learning episodes: 0-, 1-, 5-, 10-, and 15-min. After a 5-min retention interval, participants' performance was measured using three criteria: accuracy of note playing, consistency in pressure applied to the keys, and consistency in timing. No spacing effect was found, suggesting that the effect may not always be demonstrable for complex motor skills or non-verbal abilities (timing and motor skills. Additionally, we taught short phrases from five songs, using the same set of lags and retention interval, and did not find any spacing effect for accuracy of song reproduction. Our findings indicate that although the spacing effect is one of the most robust phenomena in the memory literature (as demonstrated by verbal learning studies, the effect may vary when considered in the novel realm of complex motor skills such as piano performance.

  13. Lack of spacing effects during piano learning.

    Science.gov (United States)

    Wiseheart, Melody; D'Souza, Annalise A; Chae, Jacey

    2017-01-01

    Spacing effects during retention of verbal information are easily obtained, and the effect size is large. Relatively little evidence exists on whether motor skill retention benefits from distributed practice, with even less evidence on complex motor skills. We taught a 17-note musical sequence on a piano to individuals without prior formal training. There were five lags between learning episodes: 0-, 1-, 5-, 10-, and 15-min. After a 5-min retention interval, participants' performance was measured using three criteria: accuracy of note playing, consistency in pressure applied to the keys, and consistency in timing. No spacing effect was found, suggesting that the effect may not always be demonstrable for complex motor skills or non-verbal abilities (timing and motor skills). Additionally, we taught short phrases from five songs, using the same set of lags and retention interval, and did not find any spacing effect for accuracy of song reproduction. Our findings indicate that although the spacing effect is one of the most robust phenomena in the memory literature (as demonstrated by verbal learning studies), the effect may vary when considered in the novel realm of complex motor skills such as piano performance.

  14. Compressing the hidden variable space of a qubit

    International Nuclear Information System (INIS)

    Montina, Alberto

    2011-01-01

    In previously exhibited hidden variable models of quantum state preparation and measurement, the number of continuous hidden variables describing the actual state of single realizations is never smaller than the quantum state manifold dimension. We introduce a simple model for a qubit whose hidden variable space is one-dimensional, i.e., smaller than the two-dimensional Bloch sphere. The hidden variable probability distributions associated with quantum states satisfy reasonable criteria of regularity. Possible generalizations of this shrinking to an N-dimensional Hilbert space are discussed.

  15. Regularities of magnetic field penetration into half-space in type-II superconductors

    International Nuclear Information System (INIS)

    Medvedev, Yu.V.; Krasnyuk, I.B.

    2003-01-01

    The equations, modeling the distributions of the magnetic field induction and current density in the half-space with an account of the exponential volt-ampere characteristics, are obtained. The velocity of the magnetization front propagation by the assigned average rate of the change by the time of the external magnetic field at the sample boundary is determined. The integral condition for the electric resistance, nonlinearly dependent on the magnetic field, by accomplishing whereof the magnetic flux penetrates into the sample with the finite velocity is indicated. The analytical representation of the equation with the exponential boundary mode, which models the change in the magnetic field at the area boundary, is pointed out [ru

  16. Some Characterizations of Convex Interval Games

    NARCIS (Netherlands)

    Brânzei, R.; Tijs, S.H.; Alparslan-Gok, S.Z.

    2008-01-01

    This paper focuses on new characterizations of convex interval games using the notions of exactness and superadditivity. We also relate big boss interval games with concave interval games and obtain characterizations of big boss interval games in terms of exactness and subadditivity.

  17. A Galerkin Finite Element Method for Numerical Solutions of the Modified Regularized Long Wave Equation

    Directory of Open Access Journals (Sweden)

    Liquan Mei

    2014-01-01

    Full Text Available A Galerkin method for a modified regularized long wave equation is studied using finite elements in space, the Crank-Nicolson scheme, and the Runge-Kutta scheme in time. In addition, an extrapolation technique is used to transform a nonlinear system into a linear system in order to improve the time accuracy of this method. A Fourier stability analysis for the method is shown to be marginally stable. Three invariants of motion are investigated. Numerical experiments are presented to check the theoretical study of this method.

  18. Downscaling Satellite Precipitation with Emphasis on Extremes: A Variational ℓ1-Norm Regularization in the Derivative Domain

    Science.gov (United States)

    Foufoula-Georgiou, E.; Ebtehaj, A. M.; Zhang, S. Q.; Hou, A. Y.

    2014-05-01

    The increasing availability of precipitation observations from space, e.g., from the Tropical Rainfall Measuring Mission (TRMM) and the forthcoming Global Precipitation Measuring (GPM) Mission, has fueled renewed interest in developing frameworks for downscaling and multi-sensor data fusion that can handle large data sets in computationally efficient ways while optimally reproducing desired properties of the underlying rainfall fields. Of special interest is the reproduction of extreme precipitation intensities and gradients, as these are directly relevant to hazard prediction. In this paper, we present a new formalism for downscaling satellite precipitation observations, which explicitly allows for the preservation of some key geometrical and statistical properties of spatial precipitation. These include sharp intensity gradients (due to high-intensity regions embedded within lower-intensity areas), coherent spatial structures (due to regions of slowly varying rainfall), and thicker-than-Gaussian tails of precipitation gradients and intensities. Specifically, we pose the downscaling problem as a discrete inverse problem and solve it via a regularized variational approach (variational downscaling) where the regularization term is selected to impose the desired smoothness in the solution while allowing for some steep gradients (called ℓ1-norm or total variation regularization). We demonstrate the duality between this geometrically inspired solution and its Bayesian statistical interpretation, which is equivalent to assuming a Laplace prior distribution for the precipitation intensities in the derivative (wavelet) space. When the observation operator is not known, we discuss the effect of its misspecification and explore a previously proposed dictionary-based sparse inverse downscaling methodology to indirectly learn the observation operator from a data base of coincidental high- and low-resolution observations. The proposed method and ideas are illustrated in case

  19. Classical and quantum investigations of four-dimensional maps with a mixed phase space

    International Nuclear Information System (INIS)

    Richter, Martin

    2012-01-01

    Systems with more than two degrees of freedom are of fundamental importance for the understanding of problems ranging from celestial mechanics to molecules. Due to the dimensionality the classical phase-space structure of such systems is more difficult to understand than for systems with two or fewer degrees of freedom. This thesis aims for a better insight into the classical as well as the quantum mechanics of 4D mappings representing driven systems with two degrees of freedom. In order to analyze such systems, we introduce 3D sections through the 4D phase space which reveal the regular and chaotic structures. We introduce these concepts by means of three example mappings of increasing complexity. After a classical analysis the systems are investigated quantum mechanically. We focus especially on two important aspects: First, we address quantum mechanical consequences of the classical Arnold web and demonstrate how quantum mechanics can resolve this web in the semiclassical limit. Second, we investigate the quantum mechanical tunneling couplings between regular and chaotic regions in phase space. We determine regular-to-chaotic tunneling rates numerically and extend the fictitious integrable system approach to higher dimensions for their prediction. Finally, we study resonance-assisted tunneling in 4D maps.

  20. Gene selection for microarray data classification via subspace learning and manifold regularization.

    Science.gov (United States)

    Tang, Chang; Cao, Lijuan; Zheng, Xiao; Wang, Minhui

    2017-12-19

    With the rapid development of DNA microarray technology, large amount of genomic data has been generated. Classification of these microarray data is a challenge task since gene expression data are often with thousands of genes but a small number of samples. In this paper, an effective gene selection method is proposed to select the best subset of genes for microarray data with the irrelevant and redundant genes removed. Compared with original data, the selected gene subset can benefit the classification task. We formulate the gene selection task as a manifold regularized subspace learning problem. In detail, a projection matrix is used to project the original high dimensional microarray data into a lower dimensional subspace, with the constraint that the original genes can be well represented by the selected genes. Meanwhile, the local manifold structure of original data is preserved by a Laplacian graph regularization term on the low-dimensional data space. The projection matrix can serve as an importance indicator of different genes. An iterative update algorithm is developed for solving the problem. Experimental results on six publicly available microarray datasets and one clinical dataset demonstrate that the proposed method performs better when compared with other state-of-the-art methods in terms of microarray data classification. Graphical Abstract The graphical abstract of this work.

  1. Multivariate interval-censored survival data

    DEFF Research Database (Denmark)

    Hougaard, Philip

    2014-01-01

    Interval censoring means that an event time is only known to lie in an interval (L,R], with L the last examination time before the event, and R the first after. In the univariate case, parametric models are easily fitted, whereas for non-parametric models, the mass is placed on some intervals, de...

  2. Clustered iterative stochastic ensemble method for multi-modal calibration of subsurface flow models

    KAUST Repository

    Elsheikh, Ahmed H.; Wheeler, Mary Fanett; Hoteit, Ibrahim

    2013-01-01

    estimation. ISEM is augmented with a clustering step based on k-means algorithm to form sub-ensembles. These sub-ensembles are used to explore different parts of the search space. Clusters are updated at regular intervals of the algorithm to allow merging

  3. A General Model of the Atmospheric Scattering in the Wavelength Interval 300 - 1100nm

    Directory of Open Access Journals (Sweden)

    K. Dimitrov

    2009-12-01

    Full Text Available We have presented and developed new theoretic-empirical models of the extinction coefficients of the molecular scattering in the lower, close to the ground troposphere. We have included the indicatrices of backscattering. The models have been presented using general analytical functions valid for the whole wavelength interval 300-1100 nm and for the whole interval of visibility from 0.1 km up to 50 km. The results have been compared in quantity with the model and experimental data of other authors. The modeling of troposphere scattering is necessary for the analysis and design of all optoelectronic free space systems: atmospheric optical communication systems, location systems for atmospheric research (LIDAR, optical radiometric systems.

  4. Predicting fecal coliform using the interval-to-interval approach and SWAT in the Miyun watershed, China.

    Science.gov (United States)

    Bai, Jianwen; Shen, Zhenyao; Yan, Tiezhu; Qiu, Jiali; Li, Yangyang

    2017-06-01

    Pathogens in manure can cause waterborne-disease outbreaks, serious illness, and even death in humans. Therefore, information about the transformation and transport of bacteria is crucial for determining their source. In this study, the Soil and Water Assessment Tool (SWAT) was applied to simulate fecal coliform bacteria load in the Miyun Reservoir watershed, China. The data for the fecal coliform were obtained at three sampling sites, Chenying (CY), Gubeikou (GBK), and Xiahui (XH). The calibration processes of the fecal coliform were conducted using the CY and GBK sites, and validation was conducted at the XH site. An interval-to-interval approach was designed and incorporated into the processes of fecal coliform calibration and validation. The 95% confidence interval of the predicted values and the 95% confidence interval of measured values were considered during calibration and validation in the interval-to-interval approach. Compared with the traditional point-to-point comparison, this method can improve simulation accuracy. The results indicated that the simulation of fecal coliform using the interval-to-interval approach was reasonable for the watershed. This method could provide a new research direction for future model calibration and validation studies.

  5. Information-theoretic semi-supervised metric learning via entropy regularization.

    Science.gov (United States)

    Niu, Gang; Dai, Bo; Yamada, Makoto; Sugiyama, Masashi

    2014-08-01

    We propose a general information-theoretic approach to semi-supervised metric learning called SERAPH (SEmi-supervised metRic leArning Paradigm with Hypersparsity) that does not rely on the manifold assumption. Given the probability parameterized by a Mahalanobis distance, we maximize its entropy on labeled data and minimize its entropy on unlabeled data following entropy regularization. For metric learning, entropy regularization improves manifold regularization by considering the dissimilarity information of unlabeled data in the unsupervised part, and hence it allows the supervised and unsupervised parts to be integrated in a natural and meaningful way. Moreover, we regularize SERAPH by trace-norm regularization to encourage low-dimensional projections associated with the distance metric. The nonconvex optimization problem of SERAPH could be solved efficiently and stably by either a gradient projection algorithm or an EM-like iterative algorithm whose M-step is convex. Experiments demonstrate that SERAPH compares favorably with many well-known metric learning methods, and the learned Mahalanobis distance possesses high discriminability even under noisy environments.

  6. A Novel Approach to Calculation of Reproducing Kernel on Infinite Interval and Applications to Boundary Value Problems

    Directory of Open Access Journals (Sweden)

    Jing Niu

    2013-01-01

    reproducing kernel on infinite interval is obtained concisely in polynomial form for the first time. Furthermore, as a particular effective application of this method, we give an explicit representation formula for calculation of reproducing kernel in reproducing kernel space with boundary value conditions.

  7. Fluctuations of quantum fields via zeta function regularization

    International Nuclear Information System (INIS)

    Cognola, Guido; Zerbini, Sergio; Elizalde, Emilio

    2002-01-01

    Explicit expressions for the expectation values and the variances of some observables, which are bilinear quantities in the quantum fields on a D-dimensional manifold, are derived making use of zeta function regularization. It is found that the variance, related to the second functional variation of the effective action, requires a further regularization and that the relative regularized variance turns out to be 2/N, where N is the number of the fields, thus being independent of the dimension D. Some illustrating examples are worked through. The issue of the stress tensor is also briefly addressed

  8. X-ray computed tomography using curvelet sparse regularization.

    Science.gov (United States)

    Wieczorek, Matthias; Frikel, Jürgen; Vogel, Jakob; Eggl, Elena; Kopp, Felix; Noël, Peter B; Pfeiffer, Franz; Demaret, Laurent; Lasser, Tobias

    2015-04-01

    Reconstruction of x-ray computed tomography (CT) data remains a mathematically challenging problem in medical imaging. Complementing the standard analytical reconstruction methods, sparse regularization is growing in importance, as it allows inclusion of prior knowledge. The paper presents a method for sparse regularization based on the curvelet frame for the application to iterative reconstruction in x-ray computed tomography. In this work, the authors present an iterative reconstruction approach based on the alternating direction method of multipliers using curvelet sparse regularization. Evaluation of the method is performed on a specifically crafted numerical phantom dataset to highlight the method's strengths. Additional evaluation is performed on two real datasets from commercial scanners with different noise characteristics, a clinical bone sample acquired in a micro-CT and a human abdomen scanned in a diagnostic CT. The results clearly illustrate that curvelet sparse regularization has characteristic strengths. In particular, it improves the restoration and resolution of highly directional, high contrast features with smooth contrast variations. The authors also compare this approach to the popular technique of total variation and to traditional filtered backprojection. The authors conclude that curvelet sparse regularization is able to improve reconstruction quality by reducing noise while preserving highly directional features.

  9. Inverse spectral results for Schroedinger operators on the unit interval with partial information given on the potentials

    International Nuclear Information System (INIS)

    Amour, L.; Raoux, T.; Faupin, J.

    2009-01-01

    We pursue the analysis of the Schroedinger operator on the unit interval in inverse spectral theory initiated in the work of Amour and Raoux [''Inverse spectral results for Schroedinger operators on the unit interval with potentials in Lp spaces,'' Inverse Probl. 23, 2367 (2007)]. While the potentials in the work of Amour and Raoux belong to L 1 with their difference in L p (1≤p k,1 spaces having their difference in W k,p , where 1≤p≤+∞, k(set-membership sign)(0,1,2). It is proved that two potentials in W k,1 ([0,1]) being equal on [a,1] are also equal on [0,1] if their difference belongs to W k,p ([0,a]) and if the number of their common eigenvalues is sufficiently high. Naturally, this number decreases as the parameter a decreases and as the parameters k and p increase

  10. Stability of the Regular Hayward Thin-Shell Wormholes

    Directory of Open Access Journals (Sweden)

    M. Sharif

    2016-01-01

    Full Text Available The aim of this paper is to construct regular Hayward thin-shell wormholes and analyze their stability. We adopt Israel formalism to calculate surface stresses of the shell and check the null and weak energy conditions for the constructed wormholes. It is found that the stress-energy tensor components violate the null and weak energy conditions leading to the presence of exotic matter at the throat. We analyze the attractive and repulsive characteristics of wormholes corresponding to ar>0 and ar<0, respectively. We also explore stability conditions for the existence of traversable thin-shell wormholes with arbitrarily small amount of fluid describing cosmic expansion. We find that the space-time has nonphysical regions which give rise to event horizon for 0

  11. Second order elastic metrics on the shape space of curves

    DEFF Research Database (Denmark)

    Bauer, Martin; Bruveris, Martins; Harms, Philipp

    2015-01-01

    Second order Sobolev metrics on the space of regular unparametrized planar curves have several desirable completeness properties not present in lower order metrics, but numerics are still largely missing. In this paper, we present algorithms to numerically solve the initial and boundary value......, due to its generality, it could be applied to more general spaces of mapping. We demonstrate the effectiveness of our approach by analyzing a collection of shapes representing physical objects....

  12. Cognitive Aspects of Regularity Exhibit When Neighborhood Disappears

    Science.gov (United States)

    Chen, Sau-Chin; Hu, Jon-Fan

    2015-01-01

    Although regularity refers to the compatibility between pronunciation of character and sound of phonetic component, it has been suggested as being part of consistency, which is defined by neighborhood characteristics. Two experiments demonstrate how regularity effect is amplified or reduced by neighborhood characteristics and reveals the…

  13. Rote Rehearsal and Spacing Effects in the Free Recall of Pure and Mixed Lists

    Science.gov (United States)

    Verkoeijen, Peter P. J. L.; Delaney, Peter F.

    2008-01-01

    The "spacing effect" is the commonly observed phenomenon that memory for spaced repetitions is better than memory for massed repetitions. To further investigate the role of rehearsal in spacing effects, three experiments were conducted. With pure lists we found spacing effects in free recall when spacing intervals were relatively long (Experiments…

  14. Imaging of the internal structure of comet 67P/Churyumov-Gerasimenko from radiotomography CONSERT Data (Rosetta Mission) through a full 3D regularized inversion of the Helmholtz equations on functional spaces

    Science.gov (United States)

    Barriot, Jean-Pierre; Serafini, Jonathan; Sichoix, Lydie; Benna, Mehdi; Kofman, Wlodek; Herique, Alain

    We investigate the inverse problem of imaging the internal structure of comet 67P/ Churyumov-Gerasimenko from radiotomography CONSERT data by using a coupled regularized inversion of the Helmholtz equations. A first set of Helmholtz equations, written w.r.t a basis of 3D Hankel functions describes the wave propagation outside the comet at large distances, a second set of Helmholtz equations, written w.r.t. a basis of 3D Zernike functions describes the wave propagation throughout the comet with avariable permittivity. Both sets are connected by continuity equations over a sphere that surrounds the comet. This approach, derived from GPS water vapor tomography of the atmosphere,will permit a full 3D inversion of the internal structure of the comet, contrary to traditional approaches that use a discretization of space at a fraction of the radiowave wavelength.

  15. Conformally flat spaces and solutions to Yang-Mills equations

    International Nuclear Information System (INIS)

    Chaohao, G.

    1980-01-01

    Using the conformal invariance of Yang-Mills equations in four-dimensional manifolds, it is proved that in a simply connected space of negative constant curvature Yang-Mills equations admit solutions with any real number as their Pontryagin number. It is also shown that the space S 3 x S 1 which is the regular counterpart of the meron solution is one example of a class of solutions to Yang-Mills equations on compact manifolds that are neither self-dual nor anti-self-dual

  16. An adaptive regularization parameter choice strategy for multispectral bioluminescence tomography

    Energy Technology Data Exchange (ETDEWEB)

    Feng Jinchao; Qin Chenghu; Jia Kebin; Han Dong; Liu Kai; Zhu Shouping; Yang Xin; Tian Jie [Medical Image Processing Group, Institute of Automation, Chinese Academy of Sciences, P. O. Box 2728, Beijing 100190 (China); College of Electronic Information and Control Engineering, Beijing University of Technology, Beijing 100124 (China); Medical Image Processing Group, Institute of Automation, Chinese Academy of Sciences, P. O. Box 2728, Beijing 100190 (China); Medical Image Processing Group, Institute of Automation, Chinese Academy of Sciences, P. O. Box 2728, Beijing 100190 (China) and School of Life Sciences and Technology, Xidian University, Xi' an 710071 (China)

    2011-11-15

    Purpose: Bioluminescence tomography (BLT) provides an effective tool for monitoring physiological and pathological activities in vivo. However, the measured data in bioluminescence imaging are corrupted by noise. Therefore, regularization methods are commonly used to find a regularized solution. Nevertheless, for the quality of the reconstructed bioluminescent source obtained by regularization methods, the choice of the regularization parameters is crucial. To date, the selection of regularization parameters remains challenging. With regards to the above problems, the authors proposed a BLT reconstruction algorithm with an adaptive parameter choice rule. Methods: The proposed reconstruction algorithm uses a diffusion equation for modeling the bioluminescent photon transport. The diffusion equation is solved with a finite element method. Computed tomography (CT) images provide anatomical information regarding the geometry of the small animal and its internal organs. To reduce the ill-posedness of BLT, spectral information and the optimal permissible source region are employed. Then, the relationship between the unknown source distribution and multiview and multispectral boundary measurements is established based on the finite element method and the optimal permissible source region. Since the measured data are noisy, the BLT reconstruction is formulated as l{sub 2} data fidelity and a general regularization term. When choosing the regularization parameters for BLT, an efficient model function approach is proposed, which does not require knowledge of the noise level. This approach only requests the computation of the residual and regularized solution norm. With this knowledge, we construct the model function to approximate the objective function, and the regularization parameter is updated iteratively. Results: First, the micro-CT based mouse phantom was used for simulation verification. Simulation experiments were used to illustrate why multispectral data were used

  17. Scaling Non-Regular Shared-Memory Codes by Reusing Custom Loop Schedules

    Directory of Open Access Journals (Sweden)

    Dimitrios S. Nikolopoulos

    2003-01-01

    Full Text Available In this paper we explore the idea of customizing and reusing loop schedules to improve the scalability of non-regular numerical codes in shared-memory architectures with non-uniform memory access latency. The main objective is to implicitly setup affinity links between threads and data, by devising loop schedules that achieve balanced work distribution within irregular data spaces and reusing them as much as possible along the execution of the program for better memory access locality. This transformation provides a great deal of flexibility in optimizing locality, without compromising the simplicity of the shared-memory programming paradigm. In particular, the programmer does not need to explicitly distribute data between processors. The paper presents practical examples from real applications and experiments showing the efficiency of the approach.

  18. Low social rhythm regularity predicts first onset of bipolar spectrum disorders among at-risk individuals with reward hypersensitivity.

    Science.gov (United States)

    Alloy, Lauren B; Boland, Elaine M; Ng, Tommy H; Whitehouse, Wayne G; Abramson, Lyn Y

    2015-11-01

    The social zeitgeber model (Ehlers, Frank, & Kupfer, 1988) suggests that irregular daily schedules or social rhythms provide vulnerability to bipolar spectrum disorders. This study tested whether social rhythm regularity prospectively predicted first lifetime onset of bipolar spectrum disorders in adolescents already at risk for bipolar disorder based on exhibiting reward hypersensitivity. Adolescents (ages 14-19 years) previously screened to have high (n = 138) or moderate (n = 95) reward sensitivity, but no lifetime history of bipolar spectrum disorder, completed measures of depressive and manic symptoms, family history of bipolar disorder, and the Social Rhythm Metric. They were followed prospectively with semistructured diagnostic interviews every 6 months for an average of 31.7 (SD = 20.1) months. Hierarchical logistic regression indicated that low social rhythm regularity at baseline predicted greater likelihood of first onset of bipolar spectrum disorder over follow-up among high-reward-sensitivity adolescents but not moderate-reward-sensitivity adolescents, controlling for follow-up time, gender, age, family history of bipolar disorder, and initial manic and depressive symptoms (β = -.150, Wald = 4.365, p = .037, odds ratio = .861, 95% confidence interval [.748, .991]). Consistent with the social zeitgeber theory, low social rhythm regularity provides vulnerability to first onset of bipolar spectrum disorder among at-risk adolescents. It may be possible to identify adolescents at risk for developing a bipolar spectrum disorder based on exhibiting both reward hypersensitivity and social rhythm irregularity before onset occurs. (c) 2015 APA, all rights reserved).

  19. Optimal Tikhonov Regularization in Finite-Frequency Tomography

    Science.gov (United States)

    Fang, Y.; Yao, Z.; Zhou, Y.

    2017-12-01

    The last decade has witnessed a progressive transition in seismic tomography from ray theory to finite-frequency theory which overcomes the resolution limit of the high-frequency approximation in ray theory. In addition to approximations in wave propagation physics, a main difference between ray-theoretical tomography and finite-frequency tomography is the sparseness of the associated sensitivity matrix. It is well known that seismic tomographic problems are ill-posed and regularizations such as damping and smoothing are often applied to analyze the tradeoff between data misfit and model uncertainty. The regularizations depend on the structure of the matrix as well as noise level of the data. Cross-validation has been used to constrain data uncertainties in body-wave finite-frequency inversions when measurements at multiple frequencies are available to invert for a common structure. In this study, we explore an optimal Tikhonov regularization in surface-wave phase-velocity tomography based on minimization of an empirical Bayes risk function using theoretical training datasets. We exploit the structure of the sensitivity matrix in the framework of singular value decomposition (SVD) which also allows for the calculation of complete resolution matrix. We compare the optimal Tikhonov regularization in finite-frequency tomography with traditional tradeo-off analysis using surface wave dispersion measurements from global as well as regional studies.

  20. Los ámbitos de intervención en la educación social

    Directory of Open Access Journals (Sweden)

    Sindo FROUFE QUINTAS

    2009-11-01

    Full Text Available RESUMEN: Dentro de la Educación Social como objeto científico de la Pedagogía Social, uno de los temas más sugerentes y transcendentales es la señalización de sus ámbitos de intervención. Asunto no aclarado de un modo definitivo, como tampoco lo está la interpretación del mismo concepto «educación social». Nuestro estudio tiene como objetivo básico la propuesta de varios ámbitos o espacios de intervención. Cuatro serían los más importantes: educación social especializada, educación de las personas adultas, animación y tiempo libre, y formación laboral u ocupacional. Nuestra aportación más importante se centra en la descripción de los distintos ámbitos de intervención de la Educación Social, así como la de aquellos espacios en los cuales puede actuar el educador social como figura profesional. Este tipo de intervención siempre tendrá una finalidad educativa.ABSTRACT: Within Social Education as the scientific aim of Social Pedagogy, one of the richest and most important topics is the signposting of its spheres of intenvention. However, this matter has not been clasified in a definitive way as neither has interpretation, of the concept of «social education». The basic aim of this research is to propose several spheres or spaces of intervention. Four of them are the most important: specialized social education, adult education, animation and free-time, and occupational training. Our most important contribution is the description of the different spheres of intervention in social education, as well as the description of the spaces in which the social educator as professional figure can operate. This kind of intervention will always have an educational purpose.